[ 443.180543] env[62066]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'linux_bridge' {{(pid=62066) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 443.180987] env[62066]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'noop' {{(pid=62066) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 443.180987] env[62066]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'ovs' {{(pid=62066) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 443.181326] env[62066]: INFO os_vif [-] Loaded VIF plugins: linux_bridge, noop, ovs [ 443.277666] env[62066]: DEBUG oslo_concurrency.processutils [-] Running cmd (subprocess): grep -F node.session.scan /sbin/iscsiadm {{(pid=62066) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:390}} [ 443.287375] env[62066]: DEBUG oslo_concurrency.processutils [-] CMD "grep -F node.session.scan /sbin/iscsiadm" returned: 0 in 0.010s {{(pid=62066) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:428}} [ 443.891901] env[62066]: INFO nova.virt.driver [None req-32f01db5-e0d0-4550-9108-5458523bbfd0 None None] Loading compute driver 'vmwareapi.VMwareVCDriver' [ 443.963060] env[62066]: DEBUG oslo_concurrency.lockutils [-] Acquiring lock "oslo_vmware_api_lock" by "oslo_vmware.api.VMwareAPISession._create_session" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 443.963280] env[62066]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" acquired by "oslo_vmware.api.VMwareAPISession._create_session" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 443.963344] env[62066]: DEBUG oslo_vmware.service [-] Creating suds client with soap_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk' and wsdl_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk/vimService.wsdl' {{(pid=62066) __init__ /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:242}} [ 447.052887] env[62066]: DEBUG oslo_vmware.service [-] Invoking ServiceInstance.RetrieveServiceContent with opID=oslo.vmware-4b7edd41-da06-4b81-83f3-aa28fbe67d96 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 447.068812] env[62066]: DEBUG oslo_vmware.api [-] Logging into host: vc1.osci.c.eu-de-1.cloud.sap. {{(pid=62066) _create_session /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:242}} [ 447.069017] env[62066]: DEBUG oslo_vmware.service [-] Invoking SessionManager.Login with opID=oslo.vmware-8029e71c-64c8-4909-abe6-19228467e270 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 447.100556] env[62066]: INFO oslo_vmware.api [-] Successfully established new session; session ID is 1b2b0. [ 447.100737] env[62066]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" "released" by "oslo_vmware.api.VMwareAPISession._create_session" :: held 3.137s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 447.101190] env[62066]: INFO nova.virt.vmwareapi.driver [None req-32f01db5-e0d0-4550-9108-5458523bbfd0 None None] VMware vCenter version: 7.0.3 [ 447.104782] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5ab7c57-6aeb-488d-a179-92b296149c24 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 447.125879] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffa6ccdd-96ad-44fa-af1c-bd55ebc08fa4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 447.131730] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46a71b52-3ade-462a-9c4c-92ff5b02d911 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 447.138140] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d24d041-f26d-412d-bdf8-0bfa053bc306 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 447.151019] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd870137-64a2-407e-ae7f-8eb2d9141f4b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 447.156798] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30f80ad4-8a50-4f2e-89b5-4a13bf714674 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 447.186481] env[62066]: DEBUG oslo_vmware.service [-] Invoking ExtensionManager.FindExtension with opID=oslo.vmware-22cb238b-4b35-4464-bcd5-48f36e95fca0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 447.191088] env[62066]: DEBUG nova.virt.vmwareapi.driver [None req-32f01db5-e0d0-4550-9108-5458523bbfd0 None None] Extension org.openstack.compute already exists. {{(pid=62066) _register_openstack_extension /opt/stack/nova/nova/virt/vmwareapi/driver.py:226}} [ 447.193737] env[62066]: INFO nova.compute.provider_config [None req-32f01db5-e0d0-4550-9108-5458523bbfd0 None None] No provider configs found in /etc/nova/provider_config/. If files are present, ensure the Nova process has access. [ 447.697056] env[62066]: DEBUG nova.context [None req-32f01db5-e0d0-4550-9108-5458523bbfd0 None None] Found 2 cells: 00000000-0000-0000-0000-000000000000(cell0),4a22167b-449d-41dc-8310-bc32b60e65cf(cell1) {{(pid=62066) load_cells /opt/stack/nova/nova/context.py:464}} [ 447.699104] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 447.699327] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 447.700029] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 447.700465] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] Acquiring lock "4a22167b-449d-41dc-8310-bc32b60e65cf" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 447.700655] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] Lock "4a22167b-449d-41dc-8310-bc32b60e65cf" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 447.701662] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] Lock "4a22167b-449d-41dc-8310-bc32b60e65cf" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 447.721418] env[62066]: INFO dbcounter [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] Registered counter for database nova_cell0 [ 447.729429] env[62066]: INFO dbcounter [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] Registered counter for database nova_cell1 [ 447.732907] env[62066]: DEBUG oslo_db.sqlalchemy.engines [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=62066) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:342}} [ 447.733520] env[62066]: DEBUG oslo_db.sqlalchemy.engines [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=62066) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:342}} [ 447.738261] env[62066]: ERROR nova.db.main.api [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 447.738261] env[62066]: result = function(*args, **kwargs) [ 447.738261] env[62066]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 447.738261] env[62066]: return func(*args, **kwargs) [ 447.738261] env[62066]: File "/opt/stack/nova/nova/context.py", line 422, in gather_result [ 447.738261] env[62066]: result = fn(*args, **kwargs) [ 447.738261] env[62066]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 447.738261] env[62066]: return f(*args, **kwargs) [ 447.738261] env[62066]: File "/opt/stack/nova/nova/objects/service.py", line 554, in _db_service_get_minimum_version [ 447.738261] env[62066]: return db.service_get_minimum_version(context, binaries) [ 447.738261] env[62066]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 447.738261] env[62066]: _check_db_access() [ 447.738261] env[62066]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 447.738261] env[62066]: stacktrace = ''.join(traceback.format_stack()) [ 447.738261] env[62066]: [ 447.739065] env[62066]: ERROR nova.db.main.api [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 447.739065] env[62066]: result = function(*args, **kwargs) [ 447.739065] env[62066]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 447.739065] env[62066]: return func(*args, **kwargs) [ 447.739065] env[62066]: File "/opt/stack/nova/nova/context.py", line 422, in gather_result [ 447.739065] env[62066]: result = fn(*args, **kwargs) [ 447.739065] env[62066]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 447.739065] env[62066]: return f(*args, **kwargs) [ 447.739065] env[62066]: File "/opt/stack/nova/nova/objects/service.py", line 554, in _db_service_get_minimum_version [ 447.739065] env[62066]: return db.service_get_minimum_version(context, binaries) [ 447.739065] env[62066]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 447.739065] env[62066]: _check_db_access() [ 447.739065] env[62066]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 447.739065] env[62066]: stacktrace = ''.join(traceback.format_stack()) [ 447.739065] env[62066]: [ 447.739647] env[62066]: WARNING nova.objects.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] Failed to get minimum service version for cell 00000000-0000-0000-0000-000000000000 [ 447.739647] env[62066]: WARNING nova.objects.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] Failed to get minimum service version for cell 4a22167b-449d-41dc-8310-bc32b60e65cf [ 447.739981] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] Acquiring lock "singleton_lock" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 447.740154] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] Acquired lock "singleton_lock" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 447.740395] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] Releasing lock "singleton_lock" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 447.740722] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] Full set of CONF: {{(pid=62066) _wait_for_exit_or_signal /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/service.py:363}} [ 447.740872] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] ******************************************************************************** {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2806}} [ 447.740998] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] Configuration options gathered from: {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2807}} [ 447.741147] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] command line args: ['--config-file', '/etc/nova/nova.conf', '--config-file', '/etc/nova/nova-cpu-common.conf', '--config-file', '/etc/nova/nova-cpu-1.conf'] {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2808}} [ 447.741337] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] config files: ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2809}} [ 447.741463] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] ================================================================================ {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2811}} [ 447.741668] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] allow_resize_to_same_host = True {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.741834] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] arq_binding_timeout = 300 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.741963] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] backdoor_port = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.742100] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] backdoor_socket = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.742266] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] block_device_allocate_retries = 60 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.742430] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] block_device_allocate_retries_interval = 3 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.742597] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] cert = self.pem {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.742762] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] compute_driver = vmwareapi.VMwareVCDriver {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.742932] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] compute_monitors = [] {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.743122] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] config_dir = [] {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.743298] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] config_drive_format = iso9660 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.743434] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] config_file = ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.743599] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] config_source = [] {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.743786] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] console_host = devstack {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.743958] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] control_exchange = nova {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.744129] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] cpu_allocation_ratio = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.744287] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] daemon = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.744453] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] debug = True {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.744613] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] default_access_ip_network_name = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.744775] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] default_availability_zone = nova {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.744930] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] default_ephemeral_format = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.745095] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] default_green_pool_size = 1000 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.745332] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] default_log_levels = ['amqp=WARN', 'amqplib=WARN', 'boto=WARN', 'qpid=WARN', 'sqlalchemy=WARN', 'suds=INFO', 'oslo.messaging=INFO', 'oslo_messaging=INFO', 'iso8601=WARN', 'requests.packages.urllib3.connectionpool=WARN', 'urllib3.connectionpool=WARN', 'websocket=WARN', 'requests.packages.urllib3.util.retry=WARN', 'urllib3.util.retry=WARN', 'keystonemiddleware=WARN', 'routes.middleware=WARN', 'stevedore=WARN', 'taskflow=WARN', 'keystoneauth=WARN', 'oslo.cache=INFO', 'oslo_policy=INFO', 'dogpile.core.dogpile=INFO', 'glanceclient=WARN', 'oslo.privsep.daemon=INFO'] {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.745523] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] default_schedule_zone = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.745691] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] disk_allocation_ratio = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.745851] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] enable_new_services = True {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.746037] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] enabled_apis = ['osapi_compute'] {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.746207] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] enabled_ssl_apis = [] {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.746369] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] flat_injected = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.746553] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] force_config_drive = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.746718] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] force_raw_images = True {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.746889] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] graceful_shutdown_timeout = 5 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.747065] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] heal_instance_info_cache_interval = 60 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.747281] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] host = cpu-1 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.747457] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] initial_cpu_allocation_ratio = 4.0 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.747624] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] initial_disk_allocation_ratio = 1.0 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.747786] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] initial_ram_allocation_ratio = 1.0 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.747999] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] injected_network_template = /opt/stack/nova/nova/virt/interfaces.template {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.748178] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] instance_build_timeout = 0 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.748338] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] instance_delete_interval = 300 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.748504] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] instance_format = [instance: %(uuid)s] {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.748673] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] instance_name_template = instance-%08x {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.748835] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] instance_usage_audit = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.749015] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] instance_usage_audit_period = month {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.749178] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] instance_uuid_format = [instance: %(uuid)s] {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.749343] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] instances_path = /opt/stack/data/nova/instances {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.749529] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] internal_service_availability_zone = internal {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.749698] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] key = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.749860] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] live_migration_retry_count = 30 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.750037] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] log_color = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.750203] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] log_config_append = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.750369] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] log_date_format = %Y-%m-%d %H:%M:%S {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.750533] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] log_dir = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.750690] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] log_file = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.750818] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] log_options = True {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.750978] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] log_rotate_interval = 1 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.751160] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] log_rotate_interval_type = days {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.751327] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] log_rotation_type = none {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.751476] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] logging_context_format_string = %(color)s%(levelname)s %(name)s [%(global_request_id)s %(request_id)s %(project_name)s %(user_name)s%(color)s] %(instance)s%(color)s%(message)s {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.751610] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] logging_debug_format_suffix = {{(pid=%(process)d) %(funcName)s %(pathname)s:%(lineno)d}} {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.751821] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] logging_default_format_string = %(color)s%(levelname)s %(name)s [-%(color)s] %(instance)s%(color)s%(message)s {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.752032] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] logging_exception_prefix = ERROR %(name)s %(instance)s {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.752170] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] logging_user_identity_format = %(user)s %(project)s %(domain)s %(system_scope)s %(user_domain)s %(project_domain)s {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.752342] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] long_rpc_timeout = 1800 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.752503] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] max_concurrent_builds = 10 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.752664] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] max_concurrent_live_migrations = 1 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.752822] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] max_concurrent_snapshots = 5 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.752980] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] max_local_block_devices = 3 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.753151] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] max_logfile_count = 30 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.753309] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] max_logfile_size_mb = 200 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.753468] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] maximum_instance_delete_attempts = 5 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.753637] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] metadata_listen = 0.0.0.0 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.753806] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] metadata_listen_port = 8775 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.753974] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] metadata_workers = 2 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.754147] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] migrate_max_retries = -1 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.754311] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] mkisofs_cmd = genisoimage {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.754516] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] my_block_storage_ip = 10.180.1.21 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.754652] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] my_ip = 10.180.1.21 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.754834] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] network_allocate_retries = 0 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.755035] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] non_inheritable_image_properties = ['cache_in_nova', 'bittorrent'] {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.755211] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] osapi_compute_listen = 0.0.0.0 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.755375] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] osapi_compute_listen_port = 8774 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.755565] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] osapi_compute_unique_server_name_scope = {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.755737] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] osapi_compute_workers = 2 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.755902] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] password_length = 12 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.756073] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] periodic_enable = True {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.756237] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] periodic_fuzzy_delay = 60 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.756406] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] pointer_model = usbtablet {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.756569] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] preallocate_images = none {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.756726] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] publish_errors = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.756854] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] pybasedir = /opt/stack/nova {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.757013] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] ram_allocation_ratio = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.757176] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] rate_limit_burst = 0 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.757340] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] rate_limit_except_level = CRITICAL {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.757497] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] rate_limit_interval = 0 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.757654] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] reboot_timeout = 0 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.757826] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] reclaim_instance_interval = 0 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.757995] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] record = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.758176] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] reimage_timeout_per_gb = 60 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.758341] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] report_interval = 120 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.758500] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] rescue_timeout = 0 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.758658] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] reserved_host_cpus = 0 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.758816] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] reserved_host_disk_mb = 0 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.758970] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] reserved_host_memory_mb = 512 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.759140] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] reserved_huge_pages = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.759298] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] resize_confirm_window = 0 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.759455] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] resize_fs_using_block_device = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.759612] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] resume_guests_state_on_host_boot = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.759780] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] rootwrap_config = /etc/nova/rootwrap.conf {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.759940] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] rpc_response_timeout = 60 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.760108] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] run_external_periodic_tasks = True {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.760279] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] running_deleted_instance_action = reap {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.760438] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] running_deleted_instance_poll_interval = 1800 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.760595] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] running_deleted_instance_timeout = 0 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.760751] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] scheduler_instance_sync_interval = 120 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.760988] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] service_down_time = 720 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.761197] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] servicegroup_driver = db {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.761359] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] shell_completion = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.761521] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] shelved_offload_time = 0 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.761681] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] shelved_poll_interval = 3600 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.761847] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] shutdown_timeout = 0 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.762014] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] source_is_ipv6 = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.762178] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] ssl_only = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.762418] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] state_path = /opt/stack/data/n-cpu-1 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.762588] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] sync_power_state_interval = 600 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.762745] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] sync_power_state_pool_size = 1000 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.762910] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] syslog_log_facility = LOG_USER {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.763077] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] tempdir = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.763238] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] timeout_nbd = 10 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.763406] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] transport_url = **** {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.763568] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] update_resources_interval = 0 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.763728] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] use_cow_images = True {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.763906] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] use_eventlog = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.764090] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] use_journal = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.764255] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] use_json = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.764413] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] use_rootwrap_daemon = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.764573] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] use_stderr = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.764729] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] use_syslog = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.764881] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] vcpu_pin_set = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.765057] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] vif_plugging_is_fatal = True {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.765227] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] vif_plugging_timeout = 300 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.765391] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] virt_mkfs = [] {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.765582] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] volume_usage_poll_interval = 0 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.765747] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] watch_log_file = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.765915] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] web = /usr/share/spice-html5 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 447.766107] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] os_brick.lock_path = /opt/stack/data/n-cpu-1 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.766278] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] os_brick.wait_mpath_device_attempts = 4 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.766465] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] os_brick.wait_mpath_device_interval = 1 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.766652] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] oslo_concurrency.disable_process_locking = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.767182] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] oslo_concurrency.lock_path = /opt/stack/data/n-cpu-1 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.767375] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] oslo_messaging_metrics.metrics_buffer_size = 1000 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.767549] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] oslo_messaging_metrics.metrics_enabled = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.767727] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] oslo_messaging_metrics.metrics_process_name = {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.767897] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] oslo_messaging_metrics.metrics_socket_file = /var/tmp/metrics_collector.sock {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.768074] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] oslo_messaging_metrics.metrics_thread_stop_timeout = 10 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.768261] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] api.auth_strategy = keystone {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.768433] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] api.compute_link_prefix = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.768613] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] api.config_drive_skip_versions = 1.0 2007-01-19 2007-03-01 2007-08-29 2007-10-10 2007-12-15 2008-02-01 2008-09-01 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.768790] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] api.dhcp_domain = novalocal {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.768959] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] api.enable_instance_password = True {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.769140] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] api.glance_link_prefix = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.769307] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] api.instance_list_cells_batch_fixed_size = 100 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.769498] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] api.instance_list_cells_batch_strategy = distributed {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.769680] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] api.instance_list_per_project_cells = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.769847] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] api.list_records_by_skipping_down_cells = True {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.770015] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] api.local_metadata_per_cell = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.770193] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] api.max_limit = 1000 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.770362] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] api.metadata_cache_expiration = 15 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.770537] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] api.neutron_default_tenant_id = default {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.770713] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] api.response_validation = warn {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.770883] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] api.use_neutron_default_nets = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.771063] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] api.vendordata_dynamic_connect_timeout = 5 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.771232] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] api.vendordata_dynamic_failure_fatal = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.771401] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] api.vendordata_dynamic_read_timeout = 5 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.771607] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] api.vendordata_dynamic_ssl_certfile = {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.771857] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] api.vendordata_dynamic_targets = [] {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.772051] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] api.vendordata_jsonfile_path = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.772245] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] api.vendordata_providers = ['StaticJSON'] {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.772443] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] cache.backend = dogpile.cache.memcached {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.772639] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] cache.backend_argument = **** {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.772818] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] cache.config_prefix = cache.oslo {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.772990] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] cache.dead_timeout = 60.0 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.773171] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] cache.debug_cache_backend = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.773335] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] cache.enable_retry_client = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.773500] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] cache.enable_socket_keepalive = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.773674] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] cache.enabled = True {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.773840] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] cache.enforce_fips_mode = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.774106] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] cache.expiration_time = 600 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.774297] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] cache.hashclient_retry_attempts = 2 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.774467] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] cache.hashclient_retry_delay = 1.0 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.774636] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] cache.memcache_dead_retry = 300 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.774794] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] cache.memcache_password = **** {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.774958] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] cache.memcache_pool_connection_get_timeout = 10 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.775133] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] cache.memcache_pool_flush_on_reconnect = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.775299] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] cache.memcache_pool_maxsize = 10 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.775483] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] cache.memcache_pool_unused_timeout = 60 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.775680] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] cache.memcache_sasl_enabled = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.775868] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] cache.memcache_servers = ['localhost:11211'] {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.776047] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] cache.memcache_socket_timeout = 1.0 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.776213] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] cache.memcache_username = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.776382] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] cache.proxies = [] {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.776548] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] cache.redis_db = 0 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.776713] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] cache.redis_password = **** {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.776884] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] cache.redis_sentinel_service_name = mymaster {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.777087] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] cache.redis_sentinels = ['localhost:26379'] {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.777264] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] cache.redis_server = localhost:6379 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.777432] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] cache.redis_socket_timeout = 1.0 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.777594] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] cache.redis_username = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.777756] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] cache.retry_attempts = 2 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.777920] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] cache.retry_delay = 0.0 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.778093] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] cache.socket_keepalive_count = 1 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.778258] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] cache.socket_keepalive_idle = 1 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.778419] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] cache.socket_keepalive_interval = 1 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.778599] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] cache.tls_allowed_ciphers = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.778769] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] cache.tls_cafile = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.778928] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] cache.tls_certfile = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.779100] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] cache.tls_enabled = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.779262] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] cache.tls_keyfile = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.779434] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] cinder.auth_section = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.779614] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] cinder.auth_type = password {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.779778] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] cinder.cafile = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.779955] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] cinder.catalog_info = volumev3::publicURL {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.780131] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] cinder.certfile = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.780295] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] cinder.collect_timing = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.780455] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] cinder.cross_az_attach = True {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.781082] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] cinder.debug = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.781257] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] cinder.endpoint_template = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.781423] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] cinder.http_retries = 3 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.781602] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] cinder.insecure = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.781774] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] cinder.keyfile = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.781947] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] cinder.os_region_name = RegionOne {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.782127] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] cinder.split_loggers = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.782288] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] cinder.timeout = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.782458] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] compute.consecutive_build_service_disable_threshold = 10 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.782617] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] compute.cpu_dedicated_set = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.782775] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] compute.cpu_shared_set = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.782942] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] compute.image_type_exclude_list = [] {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.783116] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] compute.live_migration_wait_for_vif_plug = True {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.783279] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] compute.max_concurrent_disk_ops = 0 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.783458] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] compute.max_disk_devices_to_attach = -1 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.783652] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] compute.packing_host_numa_cells_allocation_strategy = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.783828] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] compute.provider_config_location = /etc/nova/provider_config/ {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.784065] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] compute.resource_provider_association_refresh = 300 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.784247] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] compute.sharing_providers_max_uuids_per_request = 200 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.784415] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] compute.shutdown_retry_interval = 10 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.784603] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] compute.vmdk_allowed_types = ['streamOptimized', 'monolithicSparse'] {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.784784] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] conductor.workers = 2 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.784964] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] console.allowed_origins = [] {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.785143] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] console.ssl_ciphers = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.785316] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] console.ssl_minimum_version = default {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.785515] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] consoleauth.enforce_session_timeout = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.785703] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] consoleauth.token_ttl = 600 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.785877] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] cyborg.cafile = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.786041] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] cyborg.certfile = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.786209] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] cyborg.collect_timing = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.786370] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] cyborg.connect_retries = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.786559] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] cyborg.connect_retry_delay = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.786727] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] cyborg.endpoint_override = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.786889] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] cyborg.insecure = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.787059] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] cyborg.keyfile = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.787221] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] cyborg.max_version = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.787377] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] cyborg.min_version = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.787535] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] cyborg.region_name = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.787697] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] cyborg.retriable_status_codes = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.787854] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] cyborg.service_name = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.788030] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] cyborg.service_type = accelerator {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.788196] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] cyborg.split_loggers = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.788355] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] cyborg.status_code_retries = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.788514] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] cyborg.status_code_retry_delay = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.788674] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] cyborg.timeout = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.788856] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] cyborg.valid_interfaces = ['internal', 'public'] {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.789024] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] cyborg.version = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.789211] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] database.backend = sqlalchemy {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.789383] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] database.connection = **** {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.789571] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] database.connection_debug = 0 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.789753] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] database.connection_parameters = {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.789916] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] database.connection_recycle_time = 3600 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.790088] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] database.connection_trace = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.790254] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] database.db_inc_retry_interval = True {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.790416] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] database.db_max_retries = 20 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.790579] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] database.db_max_retry_interval = 10 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.790739] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] database.db_retry_interval = 1 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.790899] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] database.max_overflow = 50 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.791070] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] database.max_pool_size = 5 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.791234] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] database.max_retries = 10 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.791405] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] database.mysql_sql_mode = TRADITIONAL {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.791566] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] database.mysql_wsrep_sync_wait = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.791725] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] database.pool_timeout = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.791885] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] database.retry_interval = 10 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.792052] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] database.slave_connection = **** {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.792216] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] database.sqlite_synchronous = True {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.792376] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] database.use_db_reconnect = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.792608] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] api_database.backend = sqlalchemy {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.792803] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] api_database.connection = **** {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.792975] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] api_database.connection_debug = 0 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.793160] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] api_database.connection_parameters = {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.793326] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] api_database.connection_recycle_time = 3600 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.793508] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] api_database.connection_trace = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.793684] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] api_database.db_inc_retry_interval = True {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.793848] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] api_database.db_max_retries = 20 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.794016] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] api_database.db_max_retry_interval = 10 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.794179] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] api_database.db_retry_interval = 1 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.794339] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] api_database.max_overflow = 50 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.794500] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] api_database.max_pool_size = 5 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.794660] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] api_database.max_retries = 10 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.794831] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] api_database.mysql_sql_mode = TRADITIONAL {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.795006] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] api_database.mysql_wsrep_sync_wait = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.795173] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] api_database.pool_timeout = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.795335] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] api_database.retry_interval = 10 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.795537] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] api_database.slave_connection = **** {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.795736] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] api_database.sqlite_synchronous = True {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.795918] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] devices.enabled_mdev_types = [] {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.796114] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] ephemeral_storage_encryption.cipher = aes-xts-plain64 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.796288] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] ephemeral_storage_encryption.default_format = luks {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.796453] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] ephemeral_storage_encryption.enabled = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.796617] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] ephemeral_storage_encryption.key_size = 512 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.796789] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] glance.api_servers = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.796952] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] glance.cafile = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.797125] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] glance.certfile = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.797290] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] glance.collect_timing = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.797450] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] glance.connect_retries = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.797610] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] glance.connect_retry_delay = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.797771] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] glance.debug = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.797938] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] glance.default_trusted_certificate_ids = [] {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.798110] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] glance.enable_certificate_validation = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.798273] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] glance.enable_rbd_download = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.798433] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] glance.endpoint_override = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.798619] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] glance.insecure = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.798802] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] glance.keyfile = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.798965] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] glance.max_version = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.799136] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] glance.min_version = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.799297] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] glance.num_retries = 3 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.799465] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] glance.rbd_ceph_conf = {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.799628] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] glance.rbd_connect_timeout = 5 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.799796] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] glance.rbd_pool = {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.799963] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] glance.rbd_user = {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.800136] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] glance.region_name = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.800296] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] glance.retriable_status_codes = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.800453] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] glance.service_name = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.800623] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] glance.service_type = image {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.800783] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] glance.split_loggers = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.800948] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] glance.status_code_retries = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.801124] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] glance.status_code_retry_delay = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.801284] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] glance.timeout = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.801462] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] glance.valid_interfaces = ['internal', 'public'] {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.801657] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] glance.verify_glance_signatures = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.801850] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] glance.version = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.802033] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] guestfs.debug = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.802204] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] mks.enabled = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.802557] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] mks.mksproxy_base_url = http://127.0.0.1:6090/ {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.802751] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] image_cache.manager_interval = 2400 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.802922] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] image_cache.precache_concurrency = 1 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.803103] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] image_cache.remove_unused_base_images = True {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.803275] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] image_cache.remove_unused_original_minimum_age_seconds = 86400 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.803462] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] image_cache.remove_unused_resized_minimum_age_seconds = 3600 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.803657] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] image_cache.subdirectory_name = _base {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.803838] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] ironic.api_max_retries = 60 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.804023] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] ironic.api_retry_interval = 2 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.804181] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] ironic.auth_section = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.804344] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] ironic.auth_type = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.804504] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] ironic.cafile = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.804686] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] ironic.certfile = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.804867] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] ironic.collect_timing = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.805043] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] ironic.conductor_group = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.805209] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] ironic.connect_retries = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.805371] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] ironic.connect_retry_delay = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.805557] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] ironic.endpoint_override = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.805727] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] ironic.insecure = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.805888] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] ironic.keyfile = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.806058] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] ironic.max_version = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.806221] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] ironic.min_version = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.806392] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] ironic.peer_list = [] {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.806552] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] ironic.region_name = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.806712] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] ironic.retriable_status_codes = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.806876] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] ironic.serial_console_state_timeout = 10 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.807044] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] ironic.service_name = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.807222] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] ironic.service_type = baremetal {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.807382] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] ironic.shard = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.807555] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] ironic.split_loggers = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.807757] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] ironic.status_code_retries = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.807930] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] ironic.status_code_retry_delay = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.808103] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] ironic.timeout = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.808292] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] ironic.valid_interfaces = ['internal', 'public'] {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.808456] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] ironic.version = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.808642] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] key_manager.backend = nova.keymgr.conf_key_mgr.ConfKeyManager {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.808816] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] key_manager.fixed_key = **** {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.809007] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] barbican.auth_endpoint = http://localhost/identity/v3 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.809180] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] barbican.barbican_api_version = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.809341] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] barbican.barbican_endpoint = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.809513] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] barbican.barbican_endpoint_type = public {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.809673] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] barbican.barbican_region_name = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.809831] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] barbican.cafile = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.809990] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] barbican.certfile = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.810166] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] barbican.collect_timing = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.810326] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] barbican.insecure = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.810485] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] barbican.keyfile = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.810649] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] barbican.number_of_retries = 60 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.810833] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] barbican.retry_delay = 1 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.811008] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] barbican.send_service_user_token = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.811181] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] barbican.split_loggers = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.811340] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] barbican.timeout = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.811503] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] barbican.verify_ssl = True {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.811665] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] barbican.verify_ssl_path = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.811832] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] barbican_service_user.auth_section = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.811995] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] barbican_service_user.auth_type = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.812168] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] barbican_service_user.cafile = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.812328] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] barbican_service_user.certfile = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.812529] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] barbican_service_user.collect_timing = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.812723] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] barbican_service_user.insecure = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.812887] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] barbican_service_user.keyfile = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.813063] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] barbican_service_user.split_loggers = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.813228] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] barbican_service_user.timeout = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.813398] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] vault.approle_role_id = **** {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.813583] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] vault.approle_secret_id = **** {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.813759] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] vault.kv_mountpoint = secret {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.813922] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] vault.kv_path = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.814096] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] vault.kv_version = 2 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.814260] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] vault.namespace = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.814430] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] vault.root_token_id = **** {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.814578] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] vault.ssl_ca_crt_file = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.814748] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] vault.timeout = 60.0 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.814911] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] vault.use_ssl = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.815099] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] vault.vault_url = http://127.0.0.1:8200 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.815277] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] keystone.auth_section = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.815469] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] keystone.auth_type = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.815644] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] keystone.cafile = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.815809] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] keystone.certfile = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.815972] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] keystone.collect_timing = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.816146] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] keystone.connect_retries = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.816305] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] keystone.connect_retry_delay = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.816467] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] keystone.endpoint_override = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.816632] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] keystone.insecure = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.816790] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] keystone.keyfile = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.816946] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] keystone.max_version = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.817117] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] keystone.min_version = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.817275] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] keystone.region_name = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.817434] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] keystone.retriable_status_codes = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.817594] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] keystone.service_name = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.817763] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] keystone.service_type = identity {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.817925] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] keystone.split_loggers = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.818092] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] keystone.status_code_retries = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.818252] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] keystone.status_code_retry_delay = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.818407] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] keystone.timeout = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.818610] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] keystone.valid_interfaces = ['internal', 'public'] {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.818785] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] keystone.version = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.818982] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] libvirt.connection_uri = {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.819158] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] libvirt.cpu_mode = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.819325] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] libvirt.cpu_model_extra_flags = [] {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.819496] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] libvirt.cpu_models = [] {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.819698] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] libvirt.cpu_power_governor_high = performance {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.819883] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] libvirt.cpu_power_governor_low = powersave {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.820059] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] libvirt.cpu_power_management = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.820235] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] libvirt.cpu_power_management_strategy = cpu_state {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.820402] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] libvirt.device_detach_attempts = 8 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.820566] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] libvirt.device_detach_timeout = 20 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.820734] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] libvirt.disk_cachemodes = [] {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.820893] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] libvirt.disk_prefix = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.821068] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] libvirt.enabled_perf_events = [] {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.821235] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] libvirt.file_backed_memory = 0 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.821401] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] libvirt.gid_maps = [] {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.821581] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] libvirt.hw_disk_discard = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.821755] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] libvirt.hw_machine_type = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.821931] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] libvirt.images_rbd_ceph_conf = {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.822112] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] libvirt.images_rbd_glance_copy_poll_interval = 15 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.822277] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] libvirt.images_rbd_glance_copy_timeout = 600 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.822445] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] libvirt.images_rbd_glance_store_name = {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.822617] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] libvirt.images_rbd_pool = rbd {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.822785] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] libvirt.images_type = default {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.822945] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] libvirt.images_volume_group = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.823120] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] libvirt.inject_key = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.823284] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] libvirt.inject_partition = -2 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.823467] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] libvirt.inject_password = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.823654] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] libvirt.iscsi_iface = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.823824] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] libvirt.iser_use_multipath = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.823989] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] libvirt.live_migration_bandwidth = 0 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.824166] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] libvirt.live_migration_completion_timeout = 800 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.824329] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] libvirt.live_migration_downtime = 500 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.824489] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] libvirt.live_migration_downtime_delay = 75 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.824675] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] libvirt.live_migration_downtime_steps = 10 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.824844] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] libvirt.live_migration_inbound_addr = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.825015] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] libvirt.live_migration_permit_auto_converge = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.825187] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] libvirt.live_migration_permit_post_copy = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.825350] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] libvirt.live_migration_scheme = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.825548] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] libvirt.live_migration_timeout_action = abort {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.825720] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] libvirt.live_migration_tunnelled = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.825880] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] libvirt.live_migration_uri = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.826055] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] libvirt.live_migration_with_native_tls = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.826219] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] libvirt.max_queues = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.826381] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] libvirt.mem_stats_period_seconds = 10 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.826614] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] libvirt.migration_inbound_addr = 10.180.1.21 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.826782] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] libvirt.nfs_mount_options = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.827124] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] libvirt.nfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.827261] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] libvirt.num_aoe_discover_tries = 3 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.827430] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] libvirt.num_iser_scan_tries = 5 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.827610] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] libvirt.num_memory_encrypted_guests = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.827794] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] libvirt.num_nvme_discover_tries = 5 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.827961] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] libvirt.num_pcie_ports = 0 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.828150] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] libvirt.num_volume_scan_tries = 5 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.828319] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] libvirt.pmem_namespaces = [] {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.828481] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] libvirt.quobyte_client_cfg = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.828766] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] libvirt.quobyte_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.828938] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] libvirt.rbd_connect_timeout = 5 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.829117] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] libvirt.rbd_destroy_volume_retries = 12 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.829284] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] libvirt.rbd_destroy_volume_retry_interval = 5 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.829447] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] libvirt.rbd_secret_uuid = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.829610] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] libvirt.rbd_user = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.829773] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] libvirt.realtime_scheduler_priority = 1 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.829954] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] libvirt.remote_filesystem_transport = ssh {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.830111] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] libvirt.rescue_image_id = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.830273] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] libvirt.rescue_kernel_id = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.830433] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] libvirt.rescue_ramdisk_id = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.830609] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] libvirt.rng_dev_path = /dev/urandom {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.830787] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] libvirt.rx_queue_size = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.830958] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] libvirt.smbfs_mount_options = {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.831242] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] libvirt.smbfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.831416] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] libvirt.snapshot_compression = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.831608] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] libvirt.snapshot_image_format = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.831878] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] libvirt.snapshots_directory = /opt/stack/data/nova/instances/snapshots {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.832068] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] libvirt.sparse_logical_volumes = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.832240] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] libvirt.swtpm_enabled = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.832413] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] libvirt.swtpm_group = tss {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.832584] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] libvirt.swtpm_user = tss {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.832754] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] libvirt.sysinfo_serial = unique {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.832912] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] libvirt.tb_cache_size = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.833082] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] libvirt.tx_queue_size = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.833252] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] libvirt.uid_maps = [] {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.833421] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] libvirt.use_virtio_for_bridges = True {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.833592] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] libvirt.virt_type = kvm {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.833792] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] libvirt.volume_clear = zero {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.833962] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] libvirt.volume_clear_size = 0 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.834142] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] libvirt.volume_use_multipath = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.834304] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] libvirt.vzstorage_cache_path = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.834474] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] libvirt.vzstorage_log_path = /var/log/vstorage/%(cluster_name)s/nova.log.gz {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.834646] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] libvirt.vzstorage_mount_group = qemu {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.834813] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] libvirt.vzstorage_mount_opts = [] {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.834981] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] libvirt.vzstorage_mount_perms = 0770 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.835273] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] libvirt.vzstorage_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.835483] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] libvirt.vzstorage_mount_user = stack {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.835649] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] libvirt.wait_soft_reboot_seconds = 120 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.835828] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] neutron.auth_section = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.836010] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] neutron.auth_type = password {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.836182] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] neutron.cafile = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.836346] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] neutron.certfile = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.836539] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] neutron.collect_timing = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.836713] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] neutron.connect_retries = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.836878] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] neutron.connect_retry_delay = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.837083] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] neutron.default_floating_pool = public {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.837254] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] neutron.endpoint_override = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.837421] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] neutron.extension_sync_interval = 600 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.837586] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] neutron.http_retries = 3 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.837750] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] neutron.insecure = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.837908] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] neutron.keyfile = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.838077] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] neutron.max_version = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.838252] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] neutron.metadata_proxy_shared_secret = **** {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.838414] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] neutron.min_version = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.838583] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] neutron.ovs_bridge = br-int {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.838749] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] neutron.physnets = [] {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.838919] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] neutron.region_name = RegionOne {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.839089] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] neutron.retriable_status_codes = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.839262] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] neutron.service_metadata_proxy = True {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.839426] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] neutron.service_name = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.839624] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] neutron.service_type = network {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.839792] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] neutron.split_loggers = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.839954] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] neutron.status_code_retries = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.840127] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] neutron.status_code_retry_delay = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.840288] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] neutron.timeout = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.840468] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] neutron.valid_interfaces = ['internal', 'public'] {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.840633] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] neutron.version = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.840805] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] notifications.bdms_in_notifications = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.840981] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] notifications.default_level = INFO {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.841170] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] notifications.notification_format = unversioned {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.841335] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] notifications.notify_on_state_change = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.841512] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] notifications.versioned_notifications_topics = ['versioned_notifications'] {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.841689] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] pci.alias = [] {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.841858] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] pci.device_spec = [] {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.842030] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] pci.report_in_placement = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.842208] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] placement.auth_section = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.842380] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] placement.auth_type = password {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.842564] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] placement.auth_url = http://10.180.1.21/identity {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.842738] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] placement.cafile = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.842900] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] placement.certfile = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.843077] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] placement.collect_timing = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.843240] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] placement.connect_retries = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.843401] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] placement.connect_retry_delay = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.843599] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] placement.default_domain_id = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.843767] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] placement.default_domain_name = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.843933] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] placement.domain_id = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.844105] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] placement.domain_name = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.844267] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] placement.endpoint_override = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.844431] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] placement.insecure = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.844592] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] placement.keyfile = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.844748] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] placement.max_version = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.844905] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] placement.min_version = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.845081] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] placement.password = **** {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.845243] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] placement.project_domain_id = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.845410] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] placement.project_domain_name = Default {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.845628] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] placement.project_id = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.845813] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] placement.project_name = service {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.845985] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] placement.region_name = RegionOne {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.846161] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] placement.retriable_status_codes = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.846324] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] placement.service_name = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.846493] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] placement.service_type = placement {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.846661] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] placement.split_loggers = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.846819] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] placement.status_code_retries = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.846979] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] placement.status_code_retry_delay = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.847151] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] placement.system_scope = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.847311] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] placement.timeout = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.847470] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] placement.trust_id = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.847627] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] placement.user_domain_id = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.847793] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] placement.user_domain_name = Default {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.847949] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] placement.user_id = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.848132] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] placement.username = nova {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.848317] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] placement.valid_interfaces = ['internal', 'public'] {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.848480] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] placement.version = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.848690] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] quota.cores = 20 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.848860] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] quota.count_usage_from_placement = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.849045] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] quota.driver = nova.quota.DbQuotaDriver {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.849225] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] quota.injected_file_content_bytes = 10240 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.849392] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] quota.injected_file_path_length = 255 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.849556] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] quota.injected_files = 5 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.849722] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] quota.instances = 10 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.849888] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] quota.key_pairs = 100 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.850067] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] quota.metadata_items = 128 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.850238] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] quota.ram = 51200 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.850403] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] quota.recheck_quota = True {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.850570] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] quota.server_group_members = 10 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.850736] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] quota.server_groups = 10 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.850910] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] scheduler.discover_hosts_in_cells_interval = -1 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.851084] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] scheduler.enable_isolated_aggregate_filtering = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.851249] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] scheduler.image_metadata_prefilter = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.851412] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] scheduler.limit_tenants_to_placement_aggregate = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.851591] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] scheduler.max_attempts = 3 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.851767] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] scheduler.max_placement_results = 1000 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.851934] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] scheduler.placement_aggregate_required_for_tenants = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.852108] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] scheduler.query_placement_for_image_type_support = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.852272] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] scheduler.query_placement_for_routed_network_aggregates = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.852445] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] scheduler.workers = 2 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.852627] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] filter_scheduler.aggregate_image_properties_isolation_namespace = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.852788] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] filter_scheduler.aggregate_image_properties_isolation_separator = . {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.852966] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] filter_scheduler.available_filters = ['nova.scheduler.filters.all_filters'] {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.853146] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] filter_scheduler.build_failure_weight_multiplier = 1000000.0 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.853315] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] filter_scheduler.cpu_weight_multiplier = 1.0 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.853510] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] filter_scheduler.cross_cell_move_weight_multiplier = 1000000.0 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.853697] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] filter_scheduler.disk_weight_multiplier = 1.0 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.853892] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] filter_scheduler.enabled_filters = ['ComputeFilter', 'ComputeCapabilitiesFilter', 'ImagePropertiesFilter', 'ServerGroupAntiAffinityFilter', 'ServerGroupAffinityFilter', 'SameHostFilter', 'DifferentHostFilter'] {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.854072] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] filter_scheduler.host_subset_size = 1 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.854240] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] filter_scheduler.hypervisor_version_weight_multiplier = 1.0 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.854400] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] filter_scheduler.image_properties_default_architecture = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.854566] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] filter_scheduler.io_ops_weight_multiplier = -1.0 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.854756] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] filter_scheduler.isolated_hosts = [] {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.854924] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] filter_scheduler.isolated_images = [] {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.855098] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] filter_scheduler.max_instances_per_host = 50 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.855263] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] filter_scheduler.max_io_ops_per_host = 8 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.855446] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] filter_scheduler.num_instances_weight_multiplier = 0.0 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.855624] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] filter_scheduler.pci_in_placement = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.855789] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] filter_scheduler.pci_weight_multiplier = 1.0 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.855950] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] filter_scheduler.ram_weight_multiplier = 1.0 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.856128] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] filter_scheduler.restrict_isolated_hosts_to_isolated_images = True {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.856292] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] filter_scheduler.shuffle_best_same_weighed_hosts = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.856454] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] filter_scheduler.soft_affinity_weight_multiplier = 1.0 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.856617] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] filter_scheduler.soft_anti_affinity_weight_multiplier = 1.0 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.856779] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] filter_scheduler.track_instance_changes = True {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.856956] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] filter_scheduler.weight_classes = ['nova.scheduler.weights.all_weighers'] {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.857140] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] metrics.required = True {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.857307] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] metrics.weight_multiplier = 1.0 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.857473] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] metrics.weight_of_unavailable = -10000.0 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.857660] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] metrics.weight_setting = [] {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.857990] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] serial_console.base_url = ws://127.0.0.1:6083/ {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.858182] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] serial_console.enabled = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.858360] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] serial_console.port_range = 10000:20000 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.858535] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] serial_console.proxyclient_address = 127.0.0.1 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.858705] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] serial_console.serialproxy_host = 0.0.0.0 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.858874] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] serial_console.serialproxy_port = 6083 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.859051] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] service_user.auth_section = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.859229] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] service_user.auth_type = password {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.859392] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] service_user.cafile = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.859554] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] service_user.certfile = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.859718] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] service_user.collect_timing = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.859878] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] service_user.insecure = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.860044] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] service_user.keyfile = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.860218] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] service_user.send_service_user_token = True {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.860382] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] service_user.split_loggers = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.860558] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] service_user.timeout = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.860772] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] spice.agent_enabled = True {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.860996] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] spice.enabled = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.861337] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] spice.html5proxy_base_url = http://127.0.0.1:6082/spice_auto.html {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.861530] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] spice.html5proxy_host = 0.0.0.0 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.861705] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] spice.html5proxy_port = 6082 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.861868] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] spice.image_compression = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.862037] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] spice.jpeg_compression = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.862202] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] spice.playback_compression = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.862364] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] spice.require_secure = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.862534] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] spice.server_listen = 127.0.0.1 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.862704] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] spice.server_proxyclient_address = 127.0.0.1 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.862863] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] spice.streaming_mode = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.863032] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] spice.zlib_compression = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.863203] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] upgrade_levels.baseapi = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.863375] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] upgrade_levels.compute = auto {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.863568] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] upgrade_levels.conductor = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.863794] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] upgrade_levels.scheduler = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.863996] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] vendordata_dynamic_auth.auth_section = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.864180] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] vendordata_dynamic_auth.auth_type = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.864345] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] vendordata_dynamic_auth.cafile = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.864506] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] vendordata_dynamic_auth.certfile = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.864673] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] vendordata_dynamic_auth.collect_timing = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.864834] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] vendordata_dynamic_auth.insecure = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.864993] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] vendordata_dynamic_auth.keyfile = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.865170] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] vendordata_dynamic_auth.split_loggers = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.865329] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] vendordata_dynamic_auth.timeout = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.865523] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] vmware.api_retry_count = 10 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.865694] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] vmware.ca_file = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.865869] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] vmware.cache_prefix = devstack-image-cache {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.866047] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] vmware.cluster_name = testcl1 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.866219] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] vmware.connection_pool_size = 10 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.866383] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] vmware.console_delay_seconds = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.866589] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] vmware.datastore_regex = ^datastore.* {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.866802] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] vmware.host_ip = vc1.osci.c.eu-de-1.cloud.sap {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.866979] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] vmware.host_password = **** {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.867163] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] vmware.host_port = 443 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.867334] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] vmware.host_username = administrator@vsphere.local {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.867504] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] vmware.insecure = True {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.867668] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] vmware.integration_bridge = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.867832] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] vmware.maximum_objects = 100 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.867992] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] vmware.pbm_default_policy = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.868169] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] vmware.pbm_enabled = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.868331] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] vmware.pbm_wsdl_location = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.868500] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] vmware.serial_log_dir = /opt/vmware/vspc {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.868662] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] vmware.serial_port_proxy_uri = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.868822] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] vmware.serial_port_service_uri = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.868988] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] vmware.task_poll_interval = 0.5 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.869173] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] vmware.use_linked_clone = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.869344] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] vmware.vnc_keymap = en-us {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.869528] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] vmware.vnc_port = 5900 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.869706] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] vmware.vnc_port_total = 10000 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.869897] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] vnc.auth_schemes = ['none'] {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.870086] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] vnc.enabled = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.870384] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] vnc.novncproxy_base_url = http://127.0.0.1:6080/vnc_auto.html {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.870570] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] vnc.novncproxy_host = 0.0.0.0 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.870743] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] vnc.novncproxy_port = 6080 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.870921] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] vnc.server_listen = 127.0.0.1 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.871107] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] vnc.server_proxyclient_address = 127.0.0.1 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.871271] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] vnc.vencrypt_ca_certs = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.871433] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] vnc.vencrypt_client_cert = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.871620] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] vnc.vencrypt_client_key = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.871850] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] workarounds.disable_compute_service_check_for_ffu = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.872026] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] workarounds.disable_deep_image_inspection = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.872194] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] workarounds.disable_fallback_pcpu_query = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.872358] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] workarounds.disable_group_policy_check_upcall = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.872522] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] workarounds.disable_libvirt_livesnapshot = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.872687] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] workarounds.disable_rootwrap = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.872848] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] workarounds.enable_numa_live_migration = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.873016] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] workarounds.enable_qemu_monitor_announce_self = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.873185] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] workarounds.ensure_libvirt_rbd_instance_dir_cleanup = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.873348] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] workarounds.handle_virt_lifecycle_events = True {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.873509] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] workarounds.libvirt_disable_apic = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.873672] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] workarounds.never_download_image_if_on_rbd = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.873835] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] workarounds.qemu_monitor_announce_self_count = 3 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.873995] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] workarounds.qemu_monitor_announce_self_interval = 1 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.874177] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] workarounds.reserve_disk_resource_for_image_cache = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.874342] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] workarounds.skip_cpu_compare_at_startup = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.874506] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] workarounds.skip_cpu_compare_on_dest = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.874694] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] workarounds.skip_hypervisor_version_check_on_lm = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.874868] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] workarounds.skip_reserve_in_use_ironic_nodes = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.875040] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] workarounds.unified_limits_count_pcpu_as_vcpu = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.875215] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] workarounds.wait_for_vif_plugged_event_during_hard_reboot = [] {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.875403] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] wsgi.api_paste_config = /etc/nova/api-paste.ini {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.875605] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] wsgi.client_socket_timeout = 900 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.875779] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] wsgi.default_pool_size = 1000 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.875970] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] wsgi.keep_alive = True {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.876186] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] wsgi.max_header_line = 16384 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.876363] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] wsgi.secure_proxy_ssl_header = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.876521] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] wsgi.ssl_ca_file = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.876685] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] wsgi.ssl_cert_file = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.876845] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] wsgi.ssl_key_file = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.877015] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] wsgi.tcp_keepidle = 600 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.877198] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] wsgi.wsgi_log_format = %(client_ip)s "%(request_line)s" status: %(status_code)s len: %(body_length)s time: %(wall_seconds).7f {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.877365] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] zvm.ca_file = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.877525] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] zvm.cloud_connector_url = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.877842] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] zvm.image_tmp_path = /opt/stack/data/n-cpu-1/images {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.878030] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] zvm.reachable_timeout = 300 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.878218] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] oslo_policy.enforce_new_defaults = True {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.878610] env[62066]: WARNING oslo_config.cfg [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] Deprecated: Option "enforce_scope" from group "oslo_policy" is deprecated for removal (This configuration was added temporarily to facilitate a smooth transition to the new RBAC. OpenStack will always enforce scope checks. This configuration option is deprecated and will be removed in the 2025.2 cycle.). Its value may be silently ignored in the future. [ 447.878796] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] oslo_policy.enforce_scope = True {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.878974] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] oslo_policy.policy_default_rule = default {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.879169] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] oslo_policy.policy_dirs = ['policy.d'] {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.879346] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] oslo_policy.policy_file = policy.yaml {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.879518] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] oslo_policy.remote_content_type = application/x-www-form-urlencoded {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.879682] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] oslo_policy.remote_ssl_ca_crt_file = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.879842] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] oslo_policy.remote_ssl_client_crt_file = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.879999] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] oslo_policy.remote_ssl_client_key_file = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.880174] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] oslo_policy.remote_ssl_verify_server_crt = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.880342] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] oslo_versionedobjects.fatal_exception_format_errors = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.880517] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] oslo_middleware.http_basic_auth_user_file = /etc/htpasswd {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.880714] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] profiler.connection_string = messaging:// {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.880896] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] profiler.enabled = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.881080] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] profiler.es_doc_type = notification {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.881248] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] profiler.es_scroll_size = 10000 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.881418] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] profiler.es_scroll_time = 2m {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.881583] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] profiler.filter_error_trace = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.881752] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] profiler.hmac_keys = **** {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.881920] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] profiler.sentinel_service_name = mymaster {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.882098] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] profiler.socket_timeout = 0.1 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.882264] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] profiler.trace_requests = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.882428] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] profiler.trace_sqlalchemy = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.882613] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] profiler_jaeger.process_tags = {} {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.882775] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] profiler_jaeger.service_name_prefix = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.882939] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] profiler_otlp.service_name_prefix = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.883116] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] remote_debug.host = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.883279] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] remote_debug.port = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.883462] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] oslo_messaging_rabbit.amqp_auto_delete = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.883626] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] oslo_messaging_rabbit.amqp_durable_queues = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.883812] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] oslo_messaging_rabbit.conn_pool_min_size = 2 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.883980] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] oslo_messaging_rabbit.conn_pool_ttl = 1200 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.884158] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] oslo_messaging_rabbit.direct_mandatory_flag = True {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.884318] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] oslo_messaging_rabbit.enable_cancel_on_failover = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.884478] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] oslo_messaging_rabbit.heartbeat_in_pthread = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.884647] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] oslo_messaging_rabbit.heartbeat_rate = 3 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.884808] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] oslo_messaging_rabbit.heartbeat_timeout_threshold = 60 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.884975] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] oslo_messaging_rabbit.hostname = devstack {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.885147] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] oslo_messaging_rabbit.kombu_compression = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.885319] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] oslo_messaging_rabbit.kombu_failover_strategy = round-robin {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.885513] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] oslo_messaging_rabbit.kombu_missing_consumer_retry_timeout = 60 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.885707] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] oslo_messaging_rabbit.kombu_reconnect_delay = 1.0 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.885892] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] oslo_messaging_rabbit.processname = nova-compute {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.886071] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] oslo_messaging_rabbit.rabbit_ha_queues = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.886238] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] oslo_messaging_rabbit.rabbit_interval_max = 30 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.886420] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] oslo_messaging_rabbit.rabbit_login_method = AMQPLAIN {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.886606] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] oslo_messaging_rabbit.rabbit_qos_prefetch_count = 0 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.886769] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] oslo_messaging_rabbit.rabbit_quorum_delivery_limit = 0 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.886937] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_bytes = 0 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.887112] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_length = 0 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.887279] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] oslo_messaging_rabbit.rabbit_quorum_queue = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.887446] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] oslo_messaging_rabbit.rabbit_retry_backoff = 2 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.887610] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] oslo_messaging_rabbit.rabbit_retry_interval = 1 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.887792] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] oslo_messaging_rabbit.rabbit_stream_fanout = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.887970] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] oslo_messaging_rabbit.rabbit_transient_queues_ttl = 1800 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.888147] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] oslo_messaging_rabbit.rabbit_transient_quorum_queue = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.888317] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] oslo_messaging_rabbit.rpc_conn_pool_size = 30 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.888485] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] oslo_messaging_rabbit.ssl = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.888660] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] oslo_messaging_rabbit.ssl_ca_file = {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.888832] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] oslo_messaging_rabbit.ssl_cert_file = {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.888994] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] oslo_messaging_rabbit.ssl_enforce_fips_mode = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.889180] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] oslo_messaging_rabbit.ssl_key_file = {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.889352] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] oslo_messaging_rabbit.ssl_version = {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.889541] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] oslo_messaging_rabbit.use_queue_manager = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.889743] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] oslo_messaging_notifications.driver = ['messagingv2'] {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.889912] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] oslo_messaging_notifications.retry = -1 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.890109] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] oslo_messaging_notifications.topics = ['notifications'] {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.890290] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] oslo_messaging_notifications.transport_url = **** {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.890464] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] oslo_limit.auth_section = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.890636] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] oslo_limit.auth_type = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.890794] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] oslo_limit.cafile = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.890951] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] oslo_limit.certfile = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.891126] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] oslo_limit.collect_timing = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.891288] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] oslo_limit.connect_retries = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.891446] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] oslo_limit.connect_retry_delay = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.891604] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] oslo_limit.endpoint_id = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.891762] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] oslo_limit.endpoint_override = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.891920] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] oslo_limit.insecure = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.892093] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] oslo_limit.keyfile = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.892255] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] oslo_limit.max_version = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.892411] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] oslo_limit.min_version = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.892590] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] oslo_limit.region_name = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.892762] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] oslo_limit.retriable_status_codes = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.892924] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] oslo_limit.service_name = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.893099] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] oslo_limit.service_type = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.893266] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] oslo_limit.split_loggers = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.893427] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] oslo_limit.status_code_retries = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.893589] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] oslo_limit.status_code_retry_delay = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.893745] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] oslo_limit.timeout = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.893904] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] oslo_limit.valid_interfaces = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.894071] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] oslo_limit.version = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.894241] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] oslo_reports.file_event_handler = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.894406] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] oslo_reports.file_event_handler_interval = 1 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.894566] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] oslo_reports.log_dir = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.894797] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] vif_plug_linux_bridge_privileged.capabilities = [12] {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.894975] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] vif_plug_linux_bridge_privileged.group = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.895152] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] vif_plug_linux_bridge_privileged.helper_command = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.895319] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] vif_plug_linux_bridge_privileged.logger_name = oslo_privsep.daemon {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.895510] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] vif_plug_linux_bridge_privileged.thread_pool_size = 8 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.895682] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] vif_plug_linux_bridge_privileged.user = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.895853] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] vif_plug_ovs_privileged.capabilities = [12, 1] {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.896030] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] vif_plug_ovs_privileged.group = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.896195] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] vif_plug_ovs_privileged.helper_command = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.896362] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] vif_plug_ovs_privileged.logger_name = oslo_privsep.daemon {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.896525] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] vif_plug_ovs_privileged.thread_pool_size = 8 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.896686] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] vif_plug_ovs_privileged.user = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.896856] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] os_vif_linux_bridge.flat_interface = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.897043] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] os_vif_linux_bridge.forward_bridge_interface = ['all'] {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.897224] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] os_vif_linux_bridge.iptables_bottom_regex = {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.897397] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] os_vif_linux_bridge.iptables_drop_action = DROP {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.897568] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] os_vif_linux_bridge.iptables_top_regex = {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.897785] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] os_vif_linux_bridge.network_device_mtu = 1500 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.897971] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] os_vif_linux_bridge.use_ipv6 = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.898153] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] os_vif_linux_bridge.vlan_interface = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.898334] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] os_vif_ovs.default_qos_type = linux-noop {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.898507] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] os_vif_ovs.isolate_vif = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.898676] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] os_vif_ovs.network_device_mtu = 1500 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.898844] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] os_vif_ovs.ovs_vsctl_timeout = 120 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.899021] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] os_vif_ovs.ovsdb_connection = tcp:127.0.0.1:6640 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.899198] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] os_vif_ovs.ovsdb_interface = native {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.899362] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] os_vif_ovs.per_port_bridge = False {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.899535] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] privsep_osbrick.capabilities = [21] {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.899693] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] privsep_osbrick.group = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.899851] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] privsep_osbrick.helper_command = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.900023] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] privsep_osbrick.logger_name = os_brick.privileged {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.900192] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] privsep_osbrick.thread_pool_size = 8 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.900351] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] privsep_osbrick.user = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.900525] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] nova_sys_admin.capabilities = [0, 1, 2, 3, 12, 21] {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.900707] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] nova_sys_admin.group = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.900882] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] nova_sys_admin.helper_command = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.901069] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] nova_sys_admin.logger_name = oslo_privsep.daemon {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.901236] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] nova_sys_admin.thread_pool_size = 8 {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.901396] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] nova_sys_admin.user = None {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 447.901528] env[62066]: DEBUG oslo_service.service [None req-c39624b9-d7f4-473c-9f18-c04c51e086d4 None None] ******************************************************************************** {{(pid=62066) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2830}} [ 447.902017] env[62066]: INFO nova.service [-] Starting compute node (version 0.0.1) [ 448.405351] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-d24a9ef4-af4d-4545-99c0-8c354fecafca None None] Getting list of instances from cluster (obj){ [ 448.405351] env[62066]: value = "domain-c8" [ 448.405351] env[62066]: _type = "ClusterComputeResource" [ 448.405351] env[62066]: } {{(pid=62066) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 448.406566] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08f2c956-edcd-4d4e-ab0d-e6c91cb833bb {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 448.415265] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-d24a9ef4-af4d-4545-99c0-8c354fecafca None None] Got total of 0 instances {{(pid=62066) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 448.415827] env[62066]: WARNING nova.virt.vmwareapi.driver [None req-d24a9ef4-af4d-4545-99c0-8c354fecafca None None] The vmwareapi driver is not tested by the OpenStack project nor does it have clear maintainer(s) and thus its quality can not be ensured. It should be considered experimental and may be removed in a future release. If you are using the driver in production please let us know via the openstack-discuss mailing list. [ 448.416310] env[62066]: INFO nova.virt.node [None req-d24a9ef4-af4d-4545-99c0-8c354fecafca None None] Generated node identity 8e4f7194-1498-4f08-8723-ab7260524bcb [ 448.416548] env[62066]: INFO nova.virt.node [None req-d24a9ef4-af4d-4545-99c0-8c354fecafca None None] Wrote node identity 8e4f7194-1498-4f08-8723-ab7260524bcb to /opt/stack/data/n-cpu-1/compute_id [ 448.919015] env[62066]: WARNING nova.compute.manager [None req-d24a9ef4-af4d-4545-99c0-8c354fecafca None None] Compute nodes ['8e4f7194-1498-4f08-8723-ab7260524bcb'] for host cpu-1 were not found in the database. If this is the first time this service is starting on this host, then you can ignore this warning. [ 449.925034] env[62066]: INFO nova.compute.manager [None req-d24a9ef4-af4d-4545-99c0-8c354fecafca None None] Looking for unclaimed instances stuck in BUILDING status for nodes managed by this host [ 450.930612] env[62066]: WARNING nova.compute.manager [None req-d24a9ef4-af4d-4545-99c0-8c354fecafca None None] No compute node record found for host cpu-1. If this is the first time this service is starting on this host, then you can ignore this warning.: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host cpu-1 could not be found. [ 450.931092] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d24a9ef4-af4d-4545-99c0-8c354fecafca None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 450.931092] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d24a9ef4-af4d-4545-99c0-8c354fecafca None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 450.931258] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d24a9ef4-af4d-4545-99c0-8c354fecafca None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 450.931411] env[62066]: DEBUG nova.compute.resource_tracker [None req-d24a9ef4-af4d-4545-99c0-8c354fecafca None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62066) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 450.932360] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0d865e5-38ae-46dd-bd5e-8ad622927085 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 450.941052] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82e1a23a-705e-4fbe-8597-2823d83f7e4c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 450.954187] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a84c1ee-9015-44cf-9b9e-7c0a92e30d19 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 450.960373] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8dbc0787-99f4-4eed-ab21-c3b17b6dc633 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 450.988835] env[62066]: DEBUG nova.compute.resource_tracker [None req-d24a9ef4-af4d-4545-99c0-8c354fecafca None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181476MB free_disk=154GB free_vcpus=48 pci_devices=None {{(pid=62066) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 450.988996] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d24a9ef4-af4d-4545-99c0-8c354fecafca None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 450.989167] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d24a9ef4-af4d-4545-99c0-8c354fecafca None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 451.491671] env[62066]: WARNING nova.compute.resource_tracker [None req-d24a9ef4-af4d-4545-99c0-8c354fecafca None None] No compute node record for cpu-1:8e4f7194-1498-4f08-8723-ab7260524bcb: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host 8e4f7194-1498-4f08-8723-ab7260524bcb could not be found. [ 451.995792] env[62066]: INFO nova.compute.resource_tracker [None req-d24a9ef4-af4d-4545-99c0-8c354fecafca None None] Compute node record created for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 with uuid: 8e4f7194-1498-4f08-8723-ab7260524bcb [ 453.503421] env[62066]: DEBUG nova.compute.resource_tracker [None req-d24a9ef4-af4d-4545-99c0-8c354fecafca None None] Total usable vcpus: 48, total allocated vcpus: 0 {{(pid=62066) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 453.503859] env[62066]: DEBUG nova.compute.resource_tracker [None req-d24a9ef4-af4d-4545-99c0-8c354fecafca None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=512MB phys_disk=200GB used_disk=0GB total_vcpus=48 used_vcpus=0 pci_stats=[] {{(pid=62066) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 453.651221] env[62066]: INFO nova.scheduler.client.report [None req-d24a9ef4-af4d-4545-99c0-8c354fecafca None None] [req-dadaf68f-4408-49b3-b29b-9a4ec884ec80] Created resource provider record via placement API for resource provider with UUID 8e4f7194-1498-4f08-8723-ab7260524bcb and name domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28. [ 453.667426] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c91d0257-9ec7-4aa8-87ff-97dad5234538 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 453.674753] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7275c469-ef6d-4b76-85c1-62ab60639f37 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 453.703783] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fcdc580-4157-47d4-9be3-900f94042b07 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 453.711167] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8caf7966-59a0-4044-b667-86bd9eb2de7c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 453.724093] env[62066]: DEBUG nova.compute.provider_tree [None req-d24a9ef4-af4d-4545-99c0-8c354fecafca None None] Updating inventory in ProviderTree for provider 8e4f7194-1498-4f08-8723-ab7260524bcb with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 454.266241] env[62066]: DEBUG nova.scheduler.client.report [None req-d24a9ef4-af4d-4545-99c0-8c354fecafca None None] Updated inventory for provider 8e4f7194-1498-4f08-8723-ab7260524bcb with generation 0 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 454.266490] env[62066]: DEBUG nova.compute.provider_tree [None req-d24a9ef4-af4d-4545-99c0-8c354fecafca None None] Updating resource provider 8e4f7194-1498-4f08-8723-ab7260524bcb generation from 0 to 1 during operation: update_inventory {{(pid=62066) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 454.266658] env[62066]: DEBUG nova.compute.provider_tree [None req-d24a9ef4-af4d-4545-99c0-8c354fecafca None None] Updating inventory in ProviderTree for provider 8e4f7194-1498-4f08-8723-ab7260524bcb with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 454.315915] env[62066]: DEBUG nova.compute.provider_tree [None req-d24a9ef4-af4d-4545-99c0-8c354fecafca None None] Updating resource provider 8e4f7194-1498-4f08-8723-ab7260524bcb generation from 1 to 2 during operation: update_traits {{(pid=62066) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 454.820875] env[62066]: DEBUG nova.compute.resource_tracker [None req-d24a9ef4-af4d-4545-99c0-8c354fecafca None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62066) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 454.821140] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d24a9ef4-af4d-4545-99c0-8c354fecafca None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.832s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 454.821286] env[62066]: DEBUG nova.service [None req-d24a9ef4-af4d-4545-99c0-8c354fecafca None None] Creating RPC server for service compute {{(pid=62066) start /opt/stack/nova/nova/service.py:186}} [ 454.836593] env[62066]: DEBUG nova.service [None req-d24a9ef4-af4d-4545-99c0-8c354fecafca None None] Join ServiceGroup membership for this service compute {{(pid=62066) start /opt/stack/nova/nova/service.py:203}} [ 454.836813] env[62066]: DEBUG nova.servicegroup.drivers.db [None req-d24a9ef4-af4d-4545-99c0-8c354fecafca None None] DB_Driver: join new ServiceGroup member cpu-1 to the compute group, service = {{(pid=62066) join /opt/stack/nova/nova/servicegroup/drivers/db.py:44}} [ 468.838929] env[62066]: DEBUG oslo_service.periodic_task [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Running periodic task ComputeManager._sync_power_states {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 469.342282] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Getting list of instances from cluster (obj){ [ 469.342282] env[62066]: value = "domain-c8" [ 469.342282] env[62066]: _type = "ClusterComputeResource" [ 469.342282] env[62066]: } {{(pid=62066) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 469.343390] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcbc4927-c6de-4c30-a02e-b55f3dd21ca1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 469.352490] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Got total of 0 instances {{(pid=62066) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 469.352717] env[62066]: DEBUG oslo_service.periodic_task [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Running periodic task ComputeManager._cleanup_running_deleted_instances {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 469.353031] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Getting list of instances from cluster (obj){ [ 469.353031] env[62066]: value = "domain-c8" [ 469.353031] env[62066]: _type = "ClusterComputeResource" [ 469.353031] env[62066]: } {{(pid=62066) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 469.353871] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-194e0800-f84c-4496-aa76-ddecc28a9a73 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 469.361731] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Got total of 0 instances {{(pid=62066) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 492.387798] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e2ac9da0-4290-4c63-a476-a96d29c6e38b tempest-ServerDiagnosticsNegativeTest-1631323339 tempest-ServerDiagnosticsNegativeTest-1631323339-project-member] Acquiring lock "8b29fc32-d27d-4474-84bc-f727f90fde09" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 492.388354] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e2ac9da0-4290-4c63-a476-a96d29c6e38b tempest-ServerDiagnosticsNegativeTest-1631323339 tempest-ServerDiagnosticsNegativeTest-1631323339-project-member] Lock "8b29fc32-d27d-4474-84bc-f727f90fde09" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 492.890670] env[62066]: DEBUG nova.compute.manager [None req-e2ac9da0-4290-4c63-a476-a96d29c6e38b tempest-ServerDiagnosticsNegativeTest-1631323339 tempest-ServerDiagnosticsNegativeTest-1631323339-project-member] [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 493.111012] env[62066]: DEBUG oslo_concurrency.lockutils [None req-dbc2121a-ef71-4ef4-8f94-3a34690382a8 tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] Acquiring lock "9fa0bae6-065d-49ad-b076-0956f1cc3677" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 493.111150] env[62066]: DEBUG oslo_concurrency.lockutils [None req-dbc2121a-ef71-4ef4-8f94-3a34690382a8 tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] Lock "9fa0bae6-065d-49ad-b076-0956f1cc3677" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 493.445527] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e2ac9da0-4290-4c63-a476-a96d29c6e38b tempest-ServerDiagnosticsNegativeTest-1631323339 tempest-ServerDiagnosticsNegativeTest-1631323339-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 493.447475] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e2ac9da0-4290-4c63-a476-a96d29c6e38b tempest-ServerDiagnosticsNegativeTest-1631323339 tempest-ServerDiagnosticsNegativeTest-1631323339-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 493.449474] env[62066]: INFO nova.compute.claims [None req-e2ac9da0-4290-4c63-a476-a96d29c6e38b tempest-ServerDiagnosticsNegativeTest-1631323339 tempest-ServerDiagnosticsNegativeTest-1631323339-project-member] [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 493.614566] env[62066]: DEBUG nova.compute.manager [None req-dbc2121a-ef71-4ef4-8f94-3a34690382a8 tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 494.149609] env[62066]: DEBUG oslo_concurrency.lockutils [None req-dbc2121a-ef71-4ef4-8f94-3a34690382a8 tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 494.175177] env[62066]: DEBUG oslo_concurrency.lockutils [None req-69d29524-cb64-4b08-8fc9-6a11810e0231 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] Acquiring lock "4b34ff57-e137-4abc-8aed-76dd2ec8d313" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 494.175774] env[62066]: DEBUG oslo_concurrency.lockutils [None req-69d29524-cb64-4b08-8fc9-6a11810e0231 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] Lock "4b34ff57-e137-4abc-8aed-76dd2ec8d313" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 494.593580] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9eabd10d-5afe-4c06-abad-b512a51e6251 tempest-ServersAdminNegativeTestJSON-732915097 tempest-ServersAdminNegativeTestJSON-732915097-project-member] Acquiring lock "21010917-f820-4163-848d-e4141c89b13d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 494.593876] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9eabd10d-5afe-4c06-abad-b512a51e6251 tempest-ServersAdminNegativeTestJSON-732915097 tempest-ServersAdminNegativeTestJSON-732915097-project-member] Lock "21010917-f820-4163-848d-e4141c89b13d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 494.613012] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca4a255a-2630-4be1-a70b-b947de88f08d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 494.622677] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccf56bc7-520b-4371-90e4-fdc050aba332 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 494.668054] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7da214ea-ac38-4b54-90fc-0da0f412f61e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 494.677785] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f35c576-8679-444a-85a0-aefcd70f8133 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 494.684855] env[62066]: DEBUG nova.compute.manager [None req-69d29524-cb64-4b08-8fc9-6a11810e0231 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] [instance: 4b34ff57-e137-4abc-8aed-76dd2ec8d313] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 494.701494] env[62066]: DEBUG nova.compute.provider_tree [None req-e2ac9da0-4290-4c63-a476-a96d29c6e38b tempest-ServerDiagnosticsNegativeTest-1631323339 tempest-ServerDiagnosticsNegativeTest-1631323339-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 494.979947] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1ca3716a-0388-4dd3-97b5-9e32fd241705 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Acquiring lock "56eee633-2c01-4df2-8d2e-58ad712942f6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 494.985623] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1ca3716a-0388-4dd3-97b5-9e32fd241705 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Lock "56eee633-2c01-4df2-8d2e-58ad712942f6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.004s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 495.096960] env[62066]: DEBUG nova.compute.manager [None req-9eabd10d-5afe-4c06-abad-b512a51e6251 tempest-ServersAdminNegativeTestJSON-732915097 tempest-ServersAdminNegativeTestJSON-732915097-project-member] [instance: 21010917-f820-4163-848d-e4141c89b13d] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 495.209807] env[62066]: DEBUG nova.scheduler.client.report [None req-e2ac9da0-4290-4c63-a476-a96d29c6e38b tempest-ServerDiagnosticsNegativeTest-1631323339 tempest-ServerDiagnosticsNegativeTest-1631323339-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 495.225800] env[62066]: DEBUG oslo_concurrency.lockutils [None req-69d29524-cb64-4b08-8fc9-6a11810e0231 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 495.486769] env[62066]: DEBUG nova.compute.manager [None req-1ca3716a-0388-4dd3-97b5-9e32fd241705 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] [instance: 56eee633-2c01-4df2-8d2e-58ad712942f6] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 495.629577] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9eabd10d-5afe-4c06-abad-b512a51e6251 tempest-ServersAdminNegativeTestJSON-732915097 tempest-ServersAdminNegativeTestJSON-732915097-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 495.714450] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e2ac9da0-4290-4c63-a476-a96d29c6e38b tempest-ServerDiagnosticsNegativeTest-1631323339 tempest-ServerDiagnosticsNegativeTest-1631323339-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.269s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 495.715500] env[62066]: DEBUG nova.compute.manager [None req-e2ac9da0-4290-4c63-a476-a96d29c6e38b tempest-ServerDiagnosticsNegativeTest-1631323339 tempest-ServerDiagnosticsNegativeTest-1631323339-project-member] [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 495.719074] env[62066]: DEBUG oslo_concurrency.lockutils [None req-dbc2121a-ef71-4ef4-8f94-3a34690382a8 tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.569s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 495.720012] env[62066]: INFO nova.compute.claims [None req-dbc2121a-ef71-4ef4-8f94-3a34690382a8 tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 496.031723] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1ca3716a-0388-4dd3-97b5-9e32fd241705 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 496.231873] env[62066]: DEBUG nova.compute.utils [None req-e2ac9da0-4290-4c63-a476-a96d29c6e38b tempest-ServerDiagnosticsNegativeTest-1631323339 tempest-ServerDiagnosticsNegativeTest-1631323339-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 496.233349] env[62066]: DEBUG nova.compute.manager [None req-e2ac9da0-4290-4c63-a476-a96d29c6e38b tempest-ServerDiagnosticsNegativeTest-1631323339 tempest-ServerDiagnosticsNegativeTest-1631323339-project-member] [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 496.233581] env[62066]: DEBUG nova.network.neutron [None req-e2ac9da0-4290-4c63-a476-a96d29c6e38b tempest-ServerDiagnosticsNegativeTest-1631323339 tempest-ServerDiagnosticsNegativeTest-1631323339-project-member] [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 496.504311] env[62066]: DEBUG oslo_concurrency.lockutils [None req-df85dbc4-0217-427e-9c09-82d1a0e6d774 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] Acquiring lock "1eeb879b-6d1e-4746-98b5-12c30d674fa9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 496.504553] env[62066]: DEBUG oslo_concurrency.lockutils [None req-df85dbc4-0217-427e-9c09-82d1a0e6d774 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] Lock "1eeb879b-6d1e-4746-98b5-12c30d674fa9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 496.506020] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c743a8d-c557-4d89-8fe0-1cc65e783f93 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] Acquiring lock "e8693f3d-ce6e-4c94-a719-4e6723c16de8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 496.506224] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c743a8d-c557-4d89-8fe0-1cc65e783f93 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] Lock "e8693f3d-ce6e-4c94-a719-4e6723c16de8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 496.744794] env[62066]: DEBUG nova.policy [None req-e2ac9da0-4290-4c63-a476-a96d29c6e38b tempest-ServerDiagnosticsNegativeTest-1631323339 tempest-ServerDiagnosticsNegativeTest-1631323339-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '91f2988d9b474be2a0fe23c7b63d65e3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '10e4162efab74f78af94efe8bf3d8f4c', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 496.746867] env[62066]: DEBUG nova.compute.manager [None req-e2ac9da0-4290-4c63-a476-a96d29c6e38b tempest-ServerDiagnosticsNegativeTest-1631323339 tempest-ServerDiagnosticsNegativeTest-1631323339-project-member] [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 496.897306] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e2362aa-a8e4-42b9-9d85-b66288f5e1e0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 496.907691] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69f181ec-874b-444c-b1ab-b03b1dbfdd84 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 496.947938] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5709dce5-b937-4602-a381-ad9a769f3bab {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 496.956313] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81896b61-687e-49d2-b838-bfbd8e7b95fe {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 496.971666] env[62066]: DEBUG nova.compute.provider_tree [None req-dbc2121a-ef71-4ef4-8f94-3a34690382a8 tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 497.009917] env[62066]: DEBUG nova.compute.manager [None req-df85dbc4-0217-427e-9c09-82d1a0e6d774 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 497.013162] env[62066]: DEBUG nova.compute.manager [None req-5c743a8d-c557-4d89-8fe0-1cc65e783f93 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 497.337600] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0e23bf5c-5716-4eba-9510-a3597f398645 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] Acquiring lock "095d69fe-bbd3-4850-9dcd-b4c2290b8352" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 497.337600] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0e23bf5c-5716-4eba-9510-a3597f398645 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] Lock "095d69fe-bbd3-4850-9dcd-b4c2290b8352" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 497.478717] env[62066]: DEBUG nova.scheduler.client.report [None req-dbc2121a-ef71-4ef4-8f94-3a34690382a8 tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 497.550399] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c743a8d-c557-4d89-8fe0-1cc65e783f93 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 497.555222] env[62066]: DEBUG oslo_concurrency.lockutils [None req-df85dbc4-0217-427e-9c09-82d1a0e6d774 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 497.742803] env[62066]: DEBUG nova.network.neutron [None req-e2ac9da0-4290-4c63-a476-a96d29c6e38b tempest-ServerDiagnosticsNegativeTest-1631323339 tempest-ServerDiagnosticsNegativeTest-1631323339-project-member] [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] Successfully created port: 91cab0bf-fb58-43b0-ab1a-a74661d807d0 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 497.770271] env[62066]: DEBUG nova.compute.manager [None req-e2ac9da0-4290-4c63-a476-a96d29c6e38b tempest-ServerDiagnosticsNegativeTest-1631323339 tempest-ServerDiagnosticsNegativeTest-1631323339-project-member] [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 497.811564] env[62066]: DEBUG nova.virt.hardware [None req-e2ac9da0-4290-4c63-a476-a96d29c6e38b tempest-ServerDiagnosticsNegativeTest-1631323339 tempest-ServerDiagnosticsNegativeTest-1631323339-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-23T13:40:41Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-23T13:40:23Z,direct_url=,disk_format='vmdk',id=50ff584c-3b50-4395-af07-3e66769bc9f7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='eb52f7069a374c61ae946f052007c6d9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-23T13:40:24Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 497.812756] env[62066]: DEBUG nova.virt.hardware [None req-e2ac9da0-4290-4c63-a476-a96d29c6e38b tempest-ServerDiagnosticsNegativeTest-1631323339 tempest-ServerDiagnosticsNegativeTest-1631323339-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 497.812756] env[62066]: DEBUG nova.virt.hardware [None req-e2ac9da0-4290-4c63-a476-a96d29c6e38b tempest-ServerDiagnosticsNegativeTest-1631323339 tempest-ServerDiagnosticsNegativeTest-1631323339-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 497.812756] env[62066]: DEBUG nova.virt.hardware [None req-e2ac9da0-4290-4c63-a476-a96d29c6e38b tempest-ServerDiagnosticsNegativeTest-1631323339 tempest-ServerDiagnosticsNegativeTest-1631323339-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 497.812756] env[62066]: DEBUG nova.virt.hardware [None req-e2ac9da0-4290-4c63-a476-a96d29c6e38b tempest-ServerDiagnosticsNegativeTest-1631323339 tempest-ServerDiagnosticsNegativeTest-1631323339-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 497.812756] env[62066]: DEBUG nova.virt.hardware [None req-e2ac9da0-4290-4c63-a476-a96d29c6e38b tempest-ServerDiagnosticsNegativeTest-1631323339 tempest-ServerDiagnosticsNegativeTest-1631323339-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 497.812976] env[62066]: DEBUG nova.virt.hardware [None req-e2ac9da0-4290-4c63-a476-a96d29c6e38b tempest-ServerDiagnosticsNegativeTest-1631323339 tempest-ServerDiagnosticsNegativeTest-1631323339-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 497.812976] env[62066]: DEBUG nova.virt.hardware [None req-e2ac9da0-4290-4c63-a476-a96d29c6e38b tempest-ServerDiagnosticsNegativeTest-1631323339 tempest-ServerDiagnosticsNegativeTest-1631323339-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 497.814123] env[62066]: DEBUG nova.virt.hardware [None req-e2ac9da0-4290-4c63-a476-a96d29c6e38b tempest-ServerDiagnosticsNegativeTest-1631323339 tempest-ServerDiagnosticsNegativeTest-1631323339-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 497.814313] env[62066]: DEBUG nova.virt.hardware [None req-e2ac9da0-4290-4c63-a476-a96d29c6e38b tempest-ServerDiagnosticsNegativeTest-1631323339 tempest-ServerDiagnosticsNegativeTest-1631323339-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 497.814491] env[62066]: DEBUG nova.virt.hardware [None req-e2ac9da0-4290-4c63-a476-a96d29c6e38b tempest-ServerDiagnosticsNegativeTest-1631323339 tempest-ServerDiagnosticsNegativeTest-1631323339-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 497.815396] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a74c0842-f792-4066-939a-d11e45c544c4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 497.826737] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ca0be17-0381-437d-a0c0-2846f02df7a8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 497.843380] env[62066]: DEBUG nova.compute.manager [None req-0e23bf5c-5716-4eba-9510-a3597f398645 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 497.848955] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2ccd627-390d-4d61-acc5-45dbb55caa52 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 497.985350] env[62066]: DEBUG oslo_concurrency.lockutils [None req-dbc2121a-ef71-4ef4-8f94-3a34690382a8 tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.263s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 497.985350] env[62066]: DEBUG nova.compute.manager [None req-dbc2121a-ef71-4ef4-8f94-3a34690382a8 tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 497.988620] env[62066]: DEBUG oslo_concurrency.lockutils [None req-69d29524-cb64-4b08-8fc9-6a11810e0231 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.762s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 497.989020] env[62066]: INFO nova.compute.claims [None req-69d29524-cb64-4b08-8fc9-6a11810e0231 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] [instance: 4b34ff57-e137-4abc-8aed-76dd2ec8d313] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 498.388979] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0e23bf5c-5716-4eba-9510-a3597f398645 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 498.495138] env[62066]: DEBUG nova.compute.utils [None req-dbc2121a-ef71-4ef4-8f94-3a34690382a8 tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 498.498807] env[62066]: DEBUG nova.compute.manager [None req-dbc2121a-ef71-4ef4-8f94-3a34690382a8 tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 498.499071] env[62066]: DEBUG nova.network.neutron [None req-dbc2121a-ef71-4ef4-8f94-3a34690382a8 tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 498.623445] env[62066]: DEBUG oslo_concurrency.lockutils [None req-68481586-e4e9-4c46-8b51-aa7845aca824 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] Acquiring lock "d522ee4a-9bd9-402b-b3bd-c36f93366249" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 498.623776] env[62066]: DEBUG oslo_concurrency.lockutils [None req-68481586-e4e9-4c46-8b51-aa7845aca824 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] Lock "d522ee4a-9bd9-402b-b3bd-c36f93366249" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 498.758216] env[62066]: DEBUG oslo_concurrency.lockutils [None req-99d923d5-ef6f-43f2-959c-4b252ac1d398 tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] Acquiring lock "68b65162-23a6-464f-a2f9-1635bea8786f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 498.758429] env[62066]: DEBUG oslo_concurrency.lockutils [None req-99d923d5-ef6f-43f2-959c-4b252ac1d398 tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] Lock "68b65162-23a6-464f-a2f9-1635bea8786f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 498.776491] env[62066]: DEBUG nova.policy [None req-dbc2121a-ef71-4ef4-8f94-3a34690382a8 tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '52f785738b1e4ea287bce066aab4db36', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '45d7a77f838249699a6f6f303b3d6670', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 499.002914] env[62066]: DEBUG nova.compute.manager [None req-dbc2121a-ef71-4ef4-8f94-3a34690382a8 tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 499.129462] env[62066]: DEBUG nova.compute.manager [None req-68481586-e4e9-4c46-8b51-aa7845aca824 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 499.261404] env[62066]: DEBUG nova.compute.manager [None req-99d923d5-ef6f-43f2-959c-4b252ac1d398 tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 499.413401] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f9cb86c-3d40-4086-8894-df20aa820dca {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 499.424773] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59869e1e-ad01-4e61-8236-0418273b315d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 499.458624] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5656456b-91a0-4b20-b9a4-8d08e05c64e5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 499.467291] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a01845da-a4b0-4566-9463-b9e6cbf2adf2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 499.480116] env[62066]: DEBUG nova.compute.provider_tree [None req-69d29524-cb64-4b08-8fc9-6a11810e0231 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 499.664374] env[62066]: DEBUG oslo_concurrency.lockutils [None req-68481586-e4e9-4c46-8b51-aa7845aca824 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 499.710651] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9838e0c6-e0b2-45d5-9c5b-a0a0b9c60219 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] Acquiring lock "64f9698e-16f5-447a-914a-9e936e05d101" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 499.711068] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9838e0c6-e0b2-45d5-9c5b-a0a0b9c60219 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] Lock "64f9698e-16f5-447a-914a-9e936e05d101" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 499.711426] env[62066]: DEBUG oslo_concurrency.lockutils [None req-67ee84a7-af8a-43fc-bff6-48212ddfc8ae tempest-ServersNegativeTestMultiTenantJSON-1325431503 tempest-ServersNegativeTestMultiTenantJSON-1325431503-project-member] Acquiring lock "95bb27b9-e5a3-418f-a078-ae22ebf40013" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 499.712082] env[62066]: DEBUG oslo_concurrency.lockutils [None req-67ee84a7-af8a-43fc-bff6-48212ddfc8ae tempest-ServersNegativeTestMultiTenantJSON-1325431503 tempest-ServersNegativeTestMultiTenantJSON-1325431503-project-member] Lock "95bb27b9-e5a3-418f-a078-ae22ebf40013" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 499.798262] env[62066]: DEBUG oslo_concurrency.lockutils [None req-99d923d5-ef6f-43f2-959c-4b252ac1d398 tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 499.803113] env[62066]: DEBUG nova.network.neutron [None req-dbc2121a-ef71-4ef4-8f94-3a34690382a8 tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] Successfully created port: 446f5348-3547-40f2-b27c-3762e9b219a7 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 499.984087] env[62066]: DEBUG nova.scheduler.client.report [None req-69d29524-cb64-4b08-8fc9-6a11810e0231 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 500.015108] env[62066]: DEBUG nova.compute.manager [None req-dbc2121a-ef71-4ef4-8f94-3a34690382a8 tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 500.045034] env[62066]: DEBUG nova.virt.hardware [None req-dbc2121a-ef71-4ef4-8f94-3a34690382a8 tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-23T13:40:41Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-23T13:40:23Z,direct_url=,disk_format='vmdk',id=50ff584c-3b50-4395-af07-3e66769bc9f7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='eb52f7069a374c61ae946f052007c6d9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-23T13:40:24Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 500.045195] env[62066]: DEBUG nova.virt.hardware [None req-dbc2121a-ef71-4ef4-8f94-3a34690382a8 tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 500.045709] env[62066]: DEBUG nova.virt.hardware [None req-dbc2121a-ef71-4ef4-8f94-3a34690382a8 tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 500.045859] env[62066]: DEBUG nova.virt.hardware [None req-dbc2121a-ef71-4ef4-8f94-3a34690382a8 tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 500.046466] env[62066]: DEBUG nova.virt.hardware [None req-dbc2121a-ef71-4ef4-8f94-3a34690382a8 tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 500.046466] env[62066]: DEBUG nova.virt.hardware [None req-dbc2121a-ef71-4ef4-8f94-3a34690382a8 tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 500.046466] env[62066]: DEBUG nova.virt.hardware [None req-dbc2121a-ef71-4ef4-8f94-3a34690382a8 tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 500.046617] env[62066]: DEBUG nova.virt.hardware [None req-dbc2121a-ef71-4ef4-8f94-3a34690382a8 tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 500.046658] env[62066]: DEBUG nova.virt.hardware [None req-dbc2121a-ef71-4ef4-8f94-3a34690382a8 tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 500.046821] env[62066]: DEBUG nova.virt.hardware [None req-dbc2121a-ef71-4ef4-8f94-3a34690382a8 tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 500.046980] env[62066]: DEBUG nova.virt.hardware [None req-dbc2121a-ef71-4ef4-8f94-3a34690382a8 tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 500.047928] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7214a159-649f-4a9d-8ce6-63d0b87c8554 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 500.056325] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9aab9e2-c6ae-4e95-b6ca-e09b84b96bf1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 500.491645] env[62066]: DEBUG oslo_concurrency.lockutils [None req-69d29524-cb64-4b08-8fc9-6a11810e0231 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.504s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 500.492133] env[62066]: DEBUG nova.compute.manager [None req-69d29524-cb64-4b08-8fc9-6a11810e0231 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] [instance: 4b34ff57-e137-4abc-8aed-76dd2ec8d313] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 500.495077] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9eabd10d-5afe-4c06-abad-b512a51e6251 tempest-ServersAdminNegativeTestJSON-732915097 tempest-ServersAdminNegativeTestJSON-732915097-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.865s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 500.496218] env[62066]: INFO nova.compute.claims [None req-9eabd10d-5afe-4c06-abad-b512a51e6251 tempest-ServersAdminNegativeTestJSON-732915097 tempest-ServersAdminNegativeTestJSON-732915097-project-member] [instance: 21010917-f820-4163-848d-e4141c89b13d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 501.008622] env[62066]: DEBUG nova.compute.utils [None req-69d29524-cb64-4b08-8fc9-6a11810e0231 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 501.010637] env[62066]: DEBUG nova.compute.manager [None req-69d29524-cb64-4b08-8fc9-6a11810e0231 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] [instance: 4b34ff57-e137-4abc-8aed-76dd2ec8d313] Not allocating networking since 'none' was specified. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 501.295570] env[62066]: ERROR nova.compute.manager [None req-e2ac9da0-4290-4c63-a476-a96d29c6e38b tempest-ServerDiagnosticsNegativeTest-1631323339 tempest-ServerDiagnosticsNegativeTest-1631323339-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 91cab0bf-fb58-43b0-ab1a-a74661d807d0, please check neutron logs for more information. [ 501.295570] env[62066]: ERROR nova.compute.manager Traceback (most recent call last): [ 501.295570] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 501.295570] env[62066]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 501.295570] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 501.295570] env[62066]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 501.295570] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 501.295570] env[62066]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 501.295570] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 501.295570] env[62066]: ERROR nova.compute.manager self.force_reraise() [ 501.295570] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 501.295570] env[62066]: ERROR nova.compute.manager raise self.value [ 501.295570] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 501.295570] env[62066]: ERROR nova.compute.manager updated_port = self._update_port( [ 501.295570] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 501.295570] env[62066]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 501.297101] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 501.297101] env[62066]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 501.297101] env[62066]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 91cab0bf-fb58-43b0-ab1a-a74661d807d0, please check neutron logs for more information. [ 501.297101] env[62066]: ERROR nova.compute.manager [ 501.297101] env[62066]: Traceback (most recent call last): [ 501.297101] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 501.297101] env[62066]: listener.cb(fileno) [ 501.297101] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 501.297101] env[62066]: result = function(*args, **kwargs) [ 501.297101] env[62066]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 501.297101] env[62066]: return func(*args, **kwargs) [ 501.297101] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 501.297101] env[62066]: raise e [ 501.297101] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 501.297101] env[62066]: nwinfo = self.network_api.allocate_for_instance( [ 501.297101] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 501.297101] env[62066]: created_port_ids = self._update_ports_for_instance( [ 501.297101] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 501.297101] env[62066]: with excutils.save_and_reraise_exception(): [ 501.297101] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 501.297101] env[62066]: self.force_reraise() [ 501.297101] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 501.297101] env[62066]: raise self.value [ 501.297101] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 501.297101] env[62066]: updated_port = self._update_port( [ 501.297101] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 501.297101] env[62066]: _ensure_no_port_binding_failure(port) [ 501.297101] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 501.297101] env[62066]: raise exception.PortBindingFailed(port_id=port['id']) [ 501.298627] env[62066]: nova.exception.PortBindingFailed: Binding failed for port 91cab0bf-fb58-43b0-ab1a-a74661d807d0, please check neutron logs for more information. [ 501.298627] env[62066]: Removing descriptor: 15 [ 501.298627] env[62066]: ERROR nova.compute.manager [None req-e2ac9da0-4290-4c63-a476-a96d29c6e38b tempest-ServerDiagnosticsNegativeTest-1631323339 tempest-ServerDiagnosticsNegativeTest-1631323339-project-member] [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 91cab0bf-fb58-43b0-ab1a-a74661d807d0, please check neutron logs for more information. [ 501.298627] env[62066]: ERROR nova.compute.manager [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] Traceback (most recent call last): [ 501.298627] env[62066]: ERROR nova.compute.manager [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 501.298627] env[62066]: ERROR nova.compute.manager [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] yield resources [ 501.298627] env[62066]: ERROR nova.compute.manager [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 501.298627] env[62066]: ERROR nova.compute.manager [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] self.driver.spawn(context, instance, image_meta, [ 501.298627] env[62066]: ERROR nova.compute.manager [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 501.298627] env[62066]: ERROR nova.compute.manager [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] self._vmops.spawn(context, instance, image_meta, injected_files, [ 501.298627] env[62066]: ERROR nova.compute.manager [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 501.298627] env[62066]: ERROR nova.compute.manager [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] vm_ref = self.build_virtual_machine(instance, [ 501.299231] env[62066]: ERROR nova.compute.manager [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 501.299231] env[62066]: ERROR nova.compute.manager [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] vif_infos = vmwarevif.get_vif_info(self._session, [ 501.299231] env[62066]: ERROR nova.compute.manager [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 501.299231] env[62066]: ERROR nova.compute.manager [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] for vif in network_info: [ 501.299231] env[62066]: ERROR nova.compute.manager [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 501.299231] env[62066]: ERROR nova.compute.manager [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] return self._sync_wrapper(fn, *args, **kwargs) [ 501.299231] env[62066]: ERROR nova.compute.manager [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 501.299231] env[62066]: ERROR nova.compute.manager [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] self.wait() [ 501.299231] env[62066]: ERROR nova.compute.manager [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 501.299231] env[62066]: ERROR nova.compute.manager [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] self[:] = self._gt.wait() [ 501.299231] env[62066]: ERROR nova.compute.manager [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 501.299231] env[62066]: ERROR nova.compute.manager [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] return self._exit_event.wait() [ 501.299231] env[62066]: ERROR nova.compute.manager [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 501.299793] env[62066]: ERROR nova.compute.manager [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] result = hub.switch() [ 501.299793] env[62066]: ERROR nova.compute.manager [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 501.299793] env[62066]: ERROR nova.compute.manager [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] return self.greenlet.switch() [ 501.299793] env[62066]: ERROR nova.compute.manager [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 501.299793] env[62066]: ERROR nova.compute.manager [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] result = function(*args, **kwargs) [ 501.299793] env[62066]: ERROR nova.compute.manager [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 501.299793] env[62066]: ERROR nova.compute.manager [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] return func(*args, **kwargs) [ 501.299793] env[62066]: ERROR nova.compute.manager [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 501.299793] env[62066]: ERROR nova.compute.manager [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] raise e [ 501.299793] env[62066]: ERROR nova.compute.manager [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 501.299793] env[62066]: ERROR nova.compute.manager [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] nwinfo = self.network_api.allocate_for_instance( [ 501.299793] env[62066]: ERROR nova.compute.manager [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 501.299793] env[62066]: ERROR nova.compute.manager [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] created_port_ids = self._update_ports_for_instance( [ 501.301940] env[62066]: ERROR nova.compute.manager [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 501.301940] env[62066]: ERROR nova.compute.manager [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] with excutils.save_and_reraise_exception(): [ 501.301940] env[62066]: ERROR nova.compute.manager [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 501.301940] env[62066]: ERROR nova.compute.manager [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] self.force_reraise() [ 501.301940] env[62066]: ERROR nova.compute.manager [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 501.301940] env[62066]: ERROR nova.compute.manager [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] raise self.value [ 501.301940] env[62066]: ERROR nova.compute.manager [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 501.301940] env[62066]: ERROR nova.compute.manager [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] updated_port = self._update_port( [ 501.301940] env[62066]: ERROR nova.compute.manager [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 501.301940] env[62066]: ERROR nova.compute.manager [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] _ensure_no_port_binding_failure(port) [ 501.301940] env[62066]: ERROR nova.compute.manager [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 501.301940] env[62066]: ERROR nova.compute.manager [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] raise exception.PortBindingFailed(port_id=port['id']) [ 501.302691] env[62066]: ERROR nova.compute.manager [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] nova.exception.PortBindingFailed: Binding failed for port 91cab0bf-fb58-43b0-ab1a-a74661d807d0, please check neutron logs for more information. [ 501.302691] env[62066]: ERROR nova.compute.manager [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] [ 501.302691] env[62066]: INFO nova.compute.manager [None req-e2ac9da0-4290-4c63-a476-a96d29c6e38b tempest-ServerDiagnosticsNegativeTest-1631323339 tempest-ServerDiagnosticsNegativeTest-1631323339-project-member] [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] Terminating instance [ 501.304325] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e2ac9da0-4290-4c63-a476-a96d29c6e38b tempest-ServerDiagnosticsNegativeTest-1631323339 tempest-ServerDiagnosticsNegativeTest-1631323339-project-member] Acquiring lock "refresh_cache-8b29fc32-d27d-4474-84bc-f727f90fde09" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 501.304602] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e2ac9da0-4290-4c63-a476-a96d29c6e38b tempest-ServerDiagnosticsNegativeTest-1631323339 tempest-ServerDiagnosticsNegativeTest-1631323339-project-member] Acquired lock "refresh_cache-8b29fc32-d27d-4474-84bc-f727f90fde09" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 501.304738] env[62066]: DEBUG nova.network.neutron [None req-e2ac9da0-4290-4c63-a476-a96d29c6e38b tempest-ServerDiagnosticsNegativeTest-1631323339 tempest-ServerDiagnosticsNegativeTest-1631323339-project-member] [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 501.518152] env[62066]: DEBUG nova.compute.manager [None req-69d29524-cb64-4b08-8fc9-6a11810e0231 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] [instance: 4b34ff57-e137-4abc-8aed-76dd2ec8d313] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 501.755685] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4807e6a-a23c-474e-bd08-9c3f4f0b2307 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 501.764144] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfbeeaea-efba-4d35-963a-6e87a3d893c8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 501.794668] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c497531-0f6a-40dd-a18f-5b7f98db103d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 501.802668] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a98dc560-508c-4084-a52f-8e822287ddaa {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 501.816562] env[62066]: DEBUG nova.compute.provider_tree [None req-9eabd10d-5afe-4c06-abad-b512a51e6251 tempest-ServersAdminNegativeTestJSON-732915097 tempest-ServersAdminNegativeTestJSON-732915097-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 501.843665] env[62066]: ERROR nova.compute.manager [None req-dbc2121a-ef71-4ef4-8f94-3a34690382a8 tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 446f5348-3547-40f2-b27c-3762e9b219a7, please check neutron logs for more information. [ 501.843665] env[62066]: ERROR nova.compute.manager Traceback (most recent call last): [ 501.843665] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 501.843665] env[62066]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 501.843665] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 501.843665] env[62066]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 501.843665] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 501.843665] env[62066]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 501.843665] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 501.843665] env[62066]: ERROR nova.compute.manager self.force_reraise() [ 501.843665] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 501.843665] env[62066]: ERROR nova.compute.manager raise self.value [ 501.843665] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 501.843665] env[62066]: ERROR nova.compute.manager updated_port = self._update_port( [ 501.843665] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 501.843665] env[62066]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 501.844257] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 501.844257] env[62066]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 501.844257] env[62066]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 446f5348-3547-40f2-b27c-3762e9b219a7, please check neutron logs for more information. [ 501.844257] env[62066]: ERROR nova.compute.manager [ 501.844257] env[62066]: Traceback (most recent call last): [ 501.844257] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 501.844257] env[62066]: listener.cb(fileno) [ 501.844257] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 501.844257] env[62066]: result = function(*args, **kwargs) [ 501.844257] env[62066]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 501.844257] env[62066]: return func(*args, **kwargs) [ 501.844257] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 501.844257] env[62066]: raise e [ 501.844257] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 501.844257] env[62066]: nwinfo = self.network_api.allocate_for_instance( [ 501.844257] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 501.844257] env[62066]: created_port_ids = self._update_ports_for_instance( [ 501.844257] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 501.844257] env[62066]: with excutils.save_and_reraise_exception(): [ 501.844257] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 501.844257] env[62066]: self.force_reraise() [ 501.844257] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 501.844257] env[62066]: raise self.value [ 501.844257] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 501.844257] env[62066]: updated_port = self._update_port( [ 501.844257] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 501.844257] env[62066]: _ensure_no_port_binding_failure(port) [ 501.844257] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 501.844257] env[62066]: raise exception.PortBindingFailed(port_id=port['id']) [ 501.845117] env[62066]: nova.exception.PortBindingFailed: Binding failed for port 446f5348-3547-40f2-b27c-3762e9b219a7, please check neutron logs for more information. [ 501.845117] env[62066]: Removing descriptor: 16 [ 501.845117] env[62066]: ERROR nova.compute.manager [None req-dbc2121a-ef71-4ef4-8f94-3a34690382a8 tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 446f5348-3547-40f2-b27c-3762e9b219a7, please check neutron logs for more information. [ 501.845117] env[62066]: ERROR nova.compute.manager [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] Traceback (most recent call last): [ 501.845117] env[62066]: ERROR nova.compute.manager [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 501.845117] env[62066]: ERROR nova.compute.manager [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] yield resources [ 501.845117] env[62066]: ERROR nova.compute.manager [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 501.845117] env[62066]: ERROR nova.compute.manager [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] self.driver.spawn(context, instance, image_meta, [ 501.845117] env[62066]: ERROR nova.compute.manager [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 501.845117] env[62066]: ERROR nova.compute.manager [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] self._vmops.spawn(context, instance, image_meta, injected_files, [ 501.845117] env[62066]: ERROR nova.compute.manager [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 501.845117] env[62066]: ERROR nova.compute.manager [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] vm_ref = self.build_virtual_machine(instance, [ 501.845460] env[62066]: ERROR nova.compute.manager [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 501.845460] env[62066]: ERROR nova.compute.manager [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] vif_infos = vmwarevif.get_vif_info(self._session, [ 501.845460] env[62066]: ERROR nova.compute.manager [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 501.845460] env[62066]: ERROR nova.compute.manager [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] for vif in network_info: [ 501.845460] env[62066]: ERROR nova.compute.manager [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 501.845460] env[62066]: ERROR nova.compute.manager [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] return self._sync_wrapper(fn, *args, **kwargs) [ 501.845460] env[62066]: ERROR nova.compute.manager [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 501.845460] env[62066]: ERROR nova.compute.manager [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] self.wait() [ 501.845460] env[62066]: ERROR nova.compute.manager [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 501.845460] env[62066]: ERROR nova.compute.manager [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] self[:] = self._gt.wait() [ 501.845460] env[62066]: ERROR nova.compute.manager [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 501.845460] env[62066]: ERROR nova.compute.manager [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] return self._exit_event.wait() [ 501.845460] env[62066]: ERROR nova.compute.manager [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 501.845876] env[62066]: ERROR nova.compute.manager [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] result = hub.switch() [ 501.845876] env[62066]: ERROR nova.compute.manager [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 501.845876] env[62066]: ERROR nova.compute.manager [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] return self.greenlet.switch() [ 501.845876] env[62066]: ERROR nova.compute.manager [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 501.845876] env[62066]: ERROR nova.compute.manager [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] result = function(*args, **kwargs) [ 501.845876] env[62066]: ERROR nova.compute.manager [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 501.845876] env[62066]: ERROR nova.compute.manager [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] return func(*args, **kwargs) [ 501.845876] env[62066]: ERROR nova.compute.manager [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 501.845876] env[62066]: ERROR nova.compute.manager [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] raise e [ 501.845876] env[62066]: ERROR nova.compute.manager [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 501.845876] env[62066]: ERROR nova.compute.manager [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] nwinfo = self.network_api.allocate_for_instance( [ 501.845876] env[62066]: ERROR nova.compute.manager [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 501.845876] env[62066]: ERROR nova.compute.manager [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] created_port_ids = self._update_ports_for_instance( [ 501.847337] env[62066]: ERROR nova.compute.manager [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 501.847337] env[62066]: ERROR nova.compute.manager [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] with excutils.save_and_reraise_exception(): [ 501.847337] env[62066]: ERROR nova.compute.manager [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 501.847337] env[62066]: ERROR nova.compute.manager [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] self.force_reraise() [ 501.847337] env[62066]: ERROR nova.compute.manager [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 501.847337] env[62066]: ERROR nova.compute.manager [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] raise self.value [ 501.847337] env[62066]: ERROR nova.compute.manager [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 501.847337] env[62066]: ERROR nova.compute.manager [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] updated_port = self._update_port( [ 501.847337] env[62066]: ERROR nova.compute.manager [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 501.847337] env[62066]: ERROR nova.compute.manager [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] _ensure_no_port_binding_failure(port) [ 501.847337] env[62066]: ERROR nova.compute.manager [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 501.847337] env[62066]: ERROR nova.compute.manager [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] raise exception.PortBindingFailed(port_id=port['id']) [ 501.847703] env[62066]: ERROR nova.compute.manager [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] nova.exception.PortBindingFailed: Binding failed for port 446f5348-3547-40f2-b27c-3762e9b219a7, please check neutron logs for more information. [ 501.847703] env[62066]: ERROR nova.compute.manager [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] [ 501.847703] env[62066]: INFO nova.compute.manager [None req-dbc2121a-ef71-4ef4-8f94-3a34690382a8 tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] Terminating instance [ 501.847703] env[62066]: DEBUG oslo_concurrency.lockutils [None req-dbc2121a-ef71-4ef4-8f94-3a34690382a8 tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] Acquiring lock "refresh_cache-9fa0bae6-065d-49ad-b076-0956f1cc3677" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 501.847703] env[62066]: DEBUG oslo_concurrency.lockutils [None req-dbc2121a-ef71-4ef4-8f94-3a34690382a8 tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] Acquired lock "refresh_cache-9fa0bae6-065d-49ad-b076-0956f1cc3677" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 501.847703] env[62066]: DEBUG nova.network.neutron [None req-dbc2121a-ef71-4ef4-8f94-3a34690382a8 tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 501.849973] env[62066]: DEBUG nova.network.neutron [None req-e2ac9da0-4290-4c63-a476-a96d29c6e38b tempest-ServerDiagnosticsNegativeTest-1631323339 tempest-ServerDiagnosticsNegativeTest-1631323339-project-member] [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 501.927087] env[62066]: DEBUG nova.network.neutron [None req-e2ac9da0-4290-4c63-a476-a96d29c6e38b tempest-ServerDiagnosticsNegativeTest-1631323339 tempest-ServerDiagnosticsNegativeTest-1631323339-project-member] [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 501.967699] env[62066]: DEBUG nova.compute.manager [req-835c40a1-0bf0-481b-9637-dd2704af4279 req-d3755735-c53d-4bb2-81d4-c482eaeb0c61 service nova] [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] Received event network-changed-91cab0bf-fb58-43b0-ab1a-a74661d807d0 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 501.967699] env[62066]: DEBUG nova.compute.manager [req-835c40a1-0bf0-481b-9637-dd2704af4279 req-d3755735-c53d-4bb2-81d4-c482eaeb0c61 service nova] [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] Refreshing instance network info cache due to event network-changed-91cab0bf-fb58-43b0-ab1a-a74661d807d0. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 501.967901] env[62066]: DEBUG oslo_concurrency.lockutils [req-835c40a1-0bf0-481b-9637-dd2704af4279 req-d3755735-c53d-4bb2-81d4-c482eaeb0c61 service nova] Acquiring lock "refresh_cache-8b29fc32-d27d-4474-84bc-f727f90fde09" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 502.321368] env[62066]: DEBUG nova.scheduler.client.report [None req-9eabd10d-5afe-4c06-abad-b512a51e6251 tempest-ServersAdminNegativeTestJSON-732915097 tempest-ServersAdminNegativeTestJSON-732915097-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 502.374779] env[62066]: DEBUG nova.network.neutron [None req-dbc2121a-ef71-4ef4-8f94-3a34690382a8 tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 502.436312] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e2ac9da0-4290-4c63-a476-a96d29c6e38b tempest-ServerDiagnosticsNegativeTest-1631323339 tempest-ServerDiagnosticsNegativeTest-1631323339-project-member] Releasing lock "refresh_cache-8b29fc32-d27d-4474-84bc-f727f90fde09" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 502.436312] env[62066]: DEBUG nova.compute.manager [None req-e2ac9da0-4290-4c63-a476-a96d29c6e38b tempest-ServerDiagnosticsNegativeTest-1631323339 tempest-ServerDiagnosticsNegativeTest-1631323339-project-member] [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 502.436312] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-e2ac9da0-4290-4c63-a476-a96d29c6e38b tempest-ServerDiagnosticsNegativeTest-1631323339 tempest-ServerDiagnosticsNegativeTest-1631323339-project-member] [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 502.436489] env[62066]: DEBUG oslo_concurrency.lockutils [req-835c40a1-0bf0-481b-9637-dd2704af4279 req-d3755735-c53d-4bb2-81d4-c482eaeb0c61 service nova] Acquired lock "refresh_cache-8b29fc32-d27d-4474-84bc-f727f90fde09" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 502.436652] env[62066]: DEBUG nova.network.neutron [req-835c40a1-0bf0-481b-9637-dd2704af4279 req-d3755735-c53d-4bb2-81d4-c482eaeb0c61 service nova] [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] Refreshing network info cache for port 91cab0bf-fb58-43b0-ab1a-a74661d807d0 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 502.437921] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-87224b09-50e9-4204-a9fc-0c9e94ae7f34 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 502.451142] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70c5580e-ff9f-490e-83a8-12a91868119e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 502.481187] env[62066]: WARNING nova.virt.vmwareapi.vmops [None req-e2ac9da0-4290-4c63-a476-a96d29c6e38b tempest-ServerDiagnosticsNegativeTest-1631323339 tempest-ServerDiagnosticsNegativeTest-1631323339-project-member] [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 8b29fc32-d27d-4474-84bc-f727f90fde09 could not be found. [ 502.481187] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-e2ac9da0-4290-4c63-a476-a96d29c6e38b tempest-ServerDiagnosticsNegativeTest-1631323339 tempest-ServerDiagnosticsNegativeTest-1631323339-project-member] [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 502.481187] env[62066]: INFO nova.compute.manager [None req-e2ac9da0-4290-4c63-a476-a96d29c6e38b tempest-ServerDiagnosticsNegativeTest-1631323339 tempest-ServerDiagnosticsNegativeTest-1631323339-project-member] [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] Took 0.04 seconds to destroy the instance on the hypervisor. [ 502.481187] env[62066]: DEBUG oslo.service.loopingcall [None req-e2ac9da0-4290-4c63-a476-a96d29c6e38b tempest-ServerDiagnosticsNegativeTest-1631323339 tempest-ServerDiagnosticsNegativeTest-1631323339-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 502.481187] env[62066]: DEBUG nova.network.neutron [None req-dbc2121a-ef71-4ef4-8f94-3a34690382a8 tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 502.482242] env[62066]: DEBUG nova.compute.manager [-] [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 502.482379] env[62066]: DEBUG nova.network.neutron [-] [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 502.531822] env[62066]: DEBUG nova.compute.manager [None req-69d29524-cb64-4b08-8fc9-6a11810e0231 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] [instance: 4b34ff57-e137-4abc-8aed-76dd2ec8d313] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 502.535226] env[62066]: DEBUG nova.network.neutron [-] [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 502.571611] env[62066]: DEBUG nova.virt.hardware [None req-69d29524-cb64-4b08-8fc9-6a11810e0231 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-23T13:40:41Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-23T13:40:23Z,direct_url=,disk_format='vmdk',id=50ff584c-3b50-4395-af07-3e66769bc9f7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='eb52f7069a374c61ae946f052007c6d9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-23T13:40:24Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 502.571855] env[62066]: DEBUG nova.virt.hardware [None req-69d29524-cb64-4b08-8fc9-6a11810e0231 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 502.571855] env[62066]: DEBUG nova.virt.hardware [None req-69d29524-cb64-4b08-8fc9-6a11810e0231 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 502.572037] env[62066]: DEBUG nova.virt.hardware [None req-69d29524-cb64-4b08-8fc9-6a11810e0231 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 502.578435] env[62066]: DEBUG nova.virt.hardware [None req-69d29524-cb64-4b08-8fc9-6a11810e0231 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 502.578435] env[62066]: DEBUG nova.virt.hardware [None req-69d29524-cb64-4b08-8fc9-6a11810e0231 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 502.578435] env[62066]: DEBUG nova.virt.hardware [None req-69d29524-cb64-4b08-8fc9-6a11810e0231 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 502.578435] env[62066]: DEBUG nova.virt.hardware [None req-69d29524-cb64-4b08-8fc9-6a11810e0231 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 502.578435] env[62066]: DEBUG nova.virt.hardware [None req-69d29524-cb64-4b08-8fc9-6a11810e0231 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 502.578788] env[62066]: DEBUG nova.virt.hardware [None req-69d29524-cb64-4b08-8fc9-6a11810e0231 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 502.578788] env[62066]: DEBUG nova.virt.hardware [None req-69d29524-cb64-4b08-8fc9-6a11810e0231 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 502.579910] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e95529d-f951-40db-9c67-ec135122047e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 502.591200] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fff21298-bb1c-4467-8406-3b6c227c094f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 502.607590] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-69d29524-cb64-4b08-8fc9-6a11810e0231 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] [instance: 4b34ff57-e137-4abc-8aed-76dd2ec8d313] Instance VIF info [] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 502.618038] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-69d29524-cb64-4b08-8fc9-6a11810e0231 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] Creating folder: OpenStack. Parent ref: group-v4. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 502.618332] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3692f431-b4f3-4dd0-b269-c13acbae248c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 502.632871] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-69d29524-cb64-4b08-8fc9-6a11810e0231 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] Created folder: OpenStack in parent group-v4. [ 502.632871] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-69d29524-cb64-4b08-8fc9-6a11810e0231 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] Creating folder: Project (f36620aa38c54139b748f8f5a97a0505). Parent ref: group-v285980. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 502.632871] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-419cb10b-5381-4112-9124-eda2ea810e31 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 502.645984] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-69d29524-cb64-4b08-8fc9-6a11810e0231 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] Created folder: Project (f36620aa38c54139b748f8f5a97a0505) in parent group-v285980. [ 502.646200] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-69d29524-cb64-4b08-8fc9-6a11810e0231 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] Creating folder: Instances. Parent ref: group-v285981. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 502.646636] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7e131197-1f93-4756-b8d1-f01ad6b949e5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 502.654891] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-69d29524-cb64-4b08-8fc9-6a11810e0231 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] Created folder: Instances in parent group-v285981. [ 502.655170] env[62066]: DEBUG oslo.service.loopingcall [None req-69d29524-cb64-4b08-8fc9-6a11810e0231 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 502.655504] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4b34ff57-e137-4abc-8aed-76dd2ec8d313] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 502.655569] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-161de96c-ce74-4c66-99d9-f71268834745 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 502.675134] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 502.675134] env[62066]: value = "task-1340601" [ 502.675134] env[62066]: _type = "Task" [ 502.675134] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 502.683641] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1340601, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 502.832195] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9eabd10d-5afe-4c06-abad-b512a51e6251 tempest-ServersAdminNegativeTestJSON-732915097 tempest-ServersAdminNegativeTestJSON-732915097-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.337s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 502.832688] env[62066]: DEBUG nova.compute.manager [None req-9eabd10d-5afe-4c06-abad-b512a51e6251 tempest-ServersAdminNegativeTestJSON-732915097 tempest-ServersAdminNegativeTestJSON-732915097-project-member] [instance: 21010917-f820-4163-848d-e4141c89b13d] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 502.835598] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1ca3716a-0388-4dd3-97b5-9e32fd241705 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.804s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 502.838253] env[62066]: INFO nova.compute.claims [None req-1ca3716a-0388-4dd3-97b5-9e32fd241705 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] [instance: 56eee633-2c01-4df2-8d2e-58ad712942f6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 502.985926] env[62066]: DEBUG oslo_concurrency.lockutils [None req-dbc2121a-ef71-4ef4-8f94-3a34690382a8 tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] Releasing lock "refresh_cache-9fa0bae6-065d-49ad-b076-0956f1cc3677" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 502.986372] env[62066]: DEBUG nova.compute.manager [None req-dbc2121a-ef71-4ef4-8f94-3a34690382a8 tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 502.988643] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-dbc2121a-ef71-4ef4-8f94-3a34690382a8 tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 502.989233] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d4ca6494-f2f8-4452-856c-ac043a367e88 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 502.996361] env[62066]: DEBUG nova.network.neutron [req-835c40a1-0bf0-481b-9637-dd2704af4279 req-d3755735-c53d-4bb2-81d4-c482eaeb0c61 service nova] [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 503.000285] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b319c201-ffbb-47e3-a7bf-ece571f089ba {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 503.022559] env[62066]: WARNING nova.virt.vmwareapi.vmops [None req-dbc2121a-ef71-4ef4-8f94-3a34690382a8 tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 9fa0bae6-065d-49ad-b076-0956f1cc3677 could not be found. [ 503.023089] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-dbc2121a-ef71-4ef4-8f94-3a34690382a8 tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 503.023298] env[62066]: INFO nova.compute.manager [None req-dbc2121a-ef71-4ef4-8f94-3a34690382a8 tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] Took 0.04 seconds to destroy the instance on the hypervisor. [ 503.023544] env[62066]: DEBUG oslo.service.loopingcall [None req-dbc2121a-ef71-4ef4-8f94-3a34690382a8 tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 503.023793] env[62066]: DEBUG nova.compute.manager [-] [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 503.023897] env[62066]: DEBUG nova.network.neutron [-] [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 503.043806] env[62066]: DEBUG nova.network.neutron [-] [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 503.069143] env[62066]: DEBUG nova.network.neutron [-] [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 503.164260] env[62066]: DEBUG nova.network.neutron [req-835c40a1-0bf0-481b-9637-dd2704af4279 req-d3755735-c53d-4bb2-81d4-c482eaeb0c61 service nova] [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 503.189470] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1340601, 'name': CreateVM_Task, 'duration_secs': 0.394316} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 503.189863] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4b34ff57-e137-4abc-8aed-76dd2ec8d313] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 503.191721] env[62066]: DEBUG oslo_vmware.service [None req-69d29524-cb64-4b08-8fc9-6a11810e0231 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ade1730c-fee5-4813-a9dc-7489407445bc {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 503.198294] env[62066]: DEBUG oslo_concurrency.lockutils [None req-69d29524-cb64-4b08-8fc9-6a11810e0231 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 503.198294] env[62066]: DEBUG oslo_concurrency.lockutils [None req-69d29524-cb64-4b08-8fc9-6a11810e0231 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] Acquired lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 503.198806] env[62066]: DEBUG oslo_concurrency.lockutils [None req-69d29524-cb64-4b08-8fc9-6a11810e0231 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 503.199354] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f0ccc368-8ceb-4c83-b081-4686df59bb13 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 503.204471] env[62066]: DEBUG oslo_vmware.api [None req-69d29524-cb64-4b08-8fc9-6a11810e0231 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] Waiting for the task: (returnval){ [ 503.204471] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]5233cadd-a66f-1595-596d-aadbad9a428c" [ 503.204471] env[62066]: _type = "Task" [ 503.204471] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 503.214323] env[62066]: DEBUG oslo_vmware.api [None req-69d29524-cb64-4b08-8fc9-6a11810e0231 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]5233cadd-a66f-1595-596d-aadbad9a428c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 503.340191] env[62066]: DEBUG oslo_service.periodic_task [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 503.340551] env[62066]: DEBUG oslo_service.periodic_task [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 503.340685] env[62066]: DEBUG nova.compute.manager [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Starting heal instance info cache {{(pid=62066) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 503.340805] env[62066]: DEBUG nova.compute.manager [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Rebuilding the list of instances to heal {{(pid=62066) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 503.344762] env[62066]: DEBUG nova.compute.utils [None req-9eabd10d-5afe-4c06-abad-b512a51e6251 tempest-ServersAdminNegativeTestJSON-732915097 tempest-ServersAdminNegativeTestJSON-732915097-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 503.348565] env[62066]: DEBUG nova.compute.manager [None req-9eabd10d-5afe-4c06-abad-b512a51e6251 tempest-ServersAdminNegativeTestJSON-732915097 tempest-ServersAdminNegativeTestJSON-732915097-project-member] [instance: 21010917-f820-4163-848d-e4141c89b13d] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 503.348754] env[62066]: DEBUG nova.network.neutron [None req-9eabd10d-5afe-4c06-abad-b512a51e6251 tempest-ServersAdminNegativeTestJSON-732915097 tempest-ServersAdminNegativeTestJSON-732915097-project-member] [instance: 21010917-f820-4163-848d-e4141c89b13d] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 503.485370] env[62066]: DEBUG nova.policy [None req-9eabd10d-5afe-4c06-abad-b512a51e6251 tempest-ServersAdminNegativeTestJSON-732915097 tempest-ServersAdminNegativeTestJSON-732915097-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '84a752059adc4cbab41a4203bc36d8be', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '248e2759e4f94b009d549e36a92d56ca', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 503.547637] env[62066]: INFO nova.compute.manager [-] [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] Took 1.07 seconds to deallocate network for instance. [ 503.553591] env[62066]: DEBUG nova.compute.claims [None req-e2ac9da0-4290-4c63-a476-a96d29c6e38b tempest-ServerDiagnosticsNegativeTest-1631323339 tempest-ServerDiagnosticsNegativeTest-1631323339-project-member] [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] Aborting claim: {{(pid=62066) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 503.554256] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e2ac9da0-4290-4c63-a476-a96d29c6e38b tempest-ServerDiagnosticsNegativeTest-1631323339 tempest-ServerDiagnosticsNegativeTest-1631323339-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 503.575879] env[62066]: DEBUG nova.network.neutron [-] [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 503.669276] env[62066]: DEBUG oslo_concurrency.lockutils [req-835c40a1-0bf0-481b-9637-dd2704af4279 req-d3755735-c53d-4bb2-81d4-c482eaeb0c61 service nova] Releasing lock "refresh_cache-8b29fc32-d27d-4474-84bc-f727f90fde09" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 503.716234] env[62066]: DEBUG oslo_concurrency.lockutils [None req-69d29524-cb64-4b08-8fc9-6a11810e0231 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] Releasing lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 503.716699] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-69d29524-cb64-4b08-8fc9-6a11810e0231 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] [instance: 4b34ff57-e137-4abc-8aed-76dd2ec8d313] Processing image 50ff584c-3b50-4395-af07-3e66769bc9f7 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 503.716778] env[62066]: DEBUG oslo_concurrency.lockutils [None req-69d29524-cb64-4b08-8fc9-6a11810e0231 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 503.716918] env[62066]: DEBUG oslo_concurrency.lockutils [None req-69d29524-cb64-4b08-8fc9-6a11810e0231 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] Acquired lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 503.717394] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-69d29524-cb64-4b08-8fc9-6a11810e0231 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 503.717687] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-802e174c-c3e7-4c21-b84f-7bbc073af16a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 503.742278] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-69d29524-cb64-4b08-8fc9-6a11810e0231 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 503.742278] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-69d29524-cb64-4b08-8fc9-6a11810e0231 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 503.742278] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a32cf5d-1fd1-4313-b468-f39f95db1cf8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 503.754065] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4413dea8-4343-4bf1-b79e-0dc8b125d9c6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 503.760184] env[62066]: DEBUG oslo_vmware.api [None req-69d29524-cb64-4b08-8fc9-6a11810e0231 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] Waiting for the task: (returnval){ [ 503.760184] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]52a82e80-aca7-1f04-697c-28b34064f5a4" [ 503.760184] env[62066]: _type = "Task" [ 503.760184] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 503.776087] env[62066]: DEBUG oslo_vmware.api [None req-69d29524-cb64-4b08-8fc9-6a11810e0231 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52a82e80-aca7-1f04-697c-28b34064f5a4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 503.849965] env[62066]: DEBUG nova.compute.manager [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] Skipping network cache update for instance because it is Building. {{(pid=62066) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 503.850227] env[62066]: DEBUG nova.compute.manager [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] Skipping network cache update for instance because it is Building. {{(pid=62066) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 503.850711] env[62066]: DEBUG nova.compute.manager [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] [instance: 4b34ff57-e137-4abc-8aed-76dd2ec8d313] Skipping network cache update for instance because it is Building. {{(pid=62066) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 503.850711] env[62066]: DEBUG nova.compute.manager [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] [instance: 21010917-f820-4163-848d-e4141c89b13d] Skipping network cache update for instance because it is Building. {{(pid=62066) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 503.850711] env[62066]: DEBUG nova.compute.manager [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] [instance: 56eee633-2c01-4df2-8d2e-58ad712942f6] Skipping network cache update for instance because it is Building. {{(pid=62066) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 503.850711] env[62066]: DEBUG nova.compute.manager [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Didn't find any instances for network info cache update. {{(pid=62066) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 503.851487] env[62066]: DEBUG nova.compute.manager [None req-9eabd10d-5afe-4c06-abad-b512a51e6251 tempest-ServersAdminNegativeTestJSON-732915097 tempest-ServersAdminNegativeTestJSON-732915097-project-member] [instance: 21010917-f820-4163-848d-e4141c89b13d] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 503.855671] env[62066]: DEBUG oslo_service.periodic_task [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 503.857111] env[62066]: DEBUG oslo_service.periodic_task [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 503.857111] env[62066]: DEBUG oslo_service.periodic_task [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 503.857111] env[62066]: DEBUG oslo_service.periodic_task [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 503.857111] env[62066]: DEBUG oslo_service.periodic_task [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 503.857111] env[62066]: DEBUG oslo_service.periodic_task [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 503.857111] env[62066]: DEBUG nova.compute.manager [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62066) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 503.857788] env[62066]: DEBUG oslo_service.periodic_task [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Running periodic task ComputeManager.update_available_resource {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 504.078894] env[62066]: INFO nova.compute.manager [-] [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] Took 1.05 seconds to deallocate network for instance. [ 504.085474] env[62066]: DEBUG nova.compute.claims [None req-dbc2121a-ef71-4ef4-8f94-3a34690382a8 tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] Aborting claim: {{(pid=62066) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 504.085938] env[62066]: DEBUG oslo_concurrency.lockutils [None req-dbc2121a-ef71-4ef4-8f94-3a34690382a8 tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 504.157991] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a00daa0b-84e7-4d84-8e6e-8a0b7fb35310 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 504.167613] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48244389-096b-4da0-a174-6734255155c5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 504.204544] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0d4a4a1-5f54-41eb-a734-eeeeadf79055 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 504.213603] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26bbe79f-c820-4bfd-87d7-4db569ed9ff0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 504.227541] env[62066]: DEBUG nova.compute.provider_tree [None req-1ca3716a-0388-4dd3-97b5-9e32fd241705 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 504.272125] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-69d29524-cb64-4b08-8fc9-6a11810e0231 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] [instance: 4b34ff57-e137-4abc-8aed-76dd2ec8d313] Preparing fetch location {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 504.274449] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-69d29524-cb64-4b08-8fc9-6a11810e0231 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] Creating directory with path [datastore2] vmware_temp/5600ab98-a0cc-467a-8243-420a5f1dc15f/50ff584c-3b50-4395-af07-3e66769bc9f7 {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 504.274449] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-cce47ac0-9c9f-4e42-a185-9a9b53d86ea8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 504.297973] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-69d29524-cb64-4b08-8fc9-6a11810e0231 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] Created directory with path [datastore2] vmware_temp/5600ab98-a0cc-467a-8243-420a5f1dc15f/50ff584c-3b50-4395-af07-3e66769bc9f7 {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 504.297973] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-69d29524-cb64-4b08-8fc9-6a11810e0231 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] [instance: 4b34ff57-e137-4abc-8aed-76dd2ec8d313] Fetch image to [datastore2] vmware_temp/5600ab98-a0cc-467a-8243-420a5f1dc15f/50ff584c-3b50-4395-af07-3e66769bc9f7/tmp-sparse.vmdk {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 504.297973] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-69d29524-cb64-4b08-8fc9-6a11810e0231 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] [instance: 4b34ff57-e137-4abc-8aed-76dd2ec8d313] Downloading image file data 50ff584c-3b50-4395-af07-3e66769bc9f7 to [datastore2] vmware_temp/5600ab98-a0cc-467a-8243-420a5f1dc15f/50ff584c-3b50-4395-af07-3e66769bc9f7/tmp-sparse.vmdk on the data store datastore2 {{(pid=62066) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 504.298796] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8456d46c-0b5f-4ab5-95d7-91e2b2c4e1b5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 504.313553] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4e70d64-c197-4685-a541-2f0d826e042b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 504.331786] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7440239-a84f-49c3-ac72-1fc7788da8d6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 504.373754] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 504.375317] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d8c7b12-44e7-4f29-8884-609c53724ffe {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 504.381747] env[62066]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-0113863f-9f2d-4828-bc9a-ba7c8a3ea915 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 504.463715] env[62066]: DEBUG nova.compute.manager [req-76a36dd5-4bfa-47e2-84eb-bf46adc7c8ba req-6ba99504-128e-4bfa-93b4-d414365a1de4 service nova] [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] Received event network-vif-deleted-91cab0bf-fb58-43b0-ab1a-a74661d807d0 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 504.463972] env[62066]: DEBUG nova.compute.manager [req-76a36dd5-4bfa-47e2-84eb-bf46adc7c8ba req-6ba99504-128e-4bfa-93b4-d414365a1de4 service nova] [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] Received event network-changed-446f5348-3547-40f2-b27c-3762e9b219a7 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 504.465925] env[62066]: DEBUG nova.compute.manager [req-76a36dd5-4bfa-47e2-84eb-bf46adc7c8ba req-6ba99504-128e-4bfa-93b4-d414365a1de4 service nova] [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] Refreshing instance network info cache due to event network-changed-446f5348-3547-40f2-b27c-3762e9b219a7. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 504.469209] env[62066]: DEBUG oslo_concurrency.lockutils [req-76a36dd5-4bfa-47e2-84eb-bf46adc7c8ba req-6ba99504-128e-4bfa-93b4-d414365a1de4 service nova] Acquiring lock "refresh_cache-9fa0bae6-065d-49ad-b076-0956f1cc3677" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 504.469209] env[62066]: DEBUG oslo_concurrency.lockutils [req-76a36dd5-4bfa-47e2-84eb-bf46adc7c8ba req-6ba99504-128e-4bfa-93b4-d414365a1de4 service nova] Acquired lock "refresh_cache-9fa0bae6-065d-49ad-b076-0956f1cc3677" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 504.469209] env[62066]: DEBUG nova.network.neutron [req-76a36dd5-4bfa-47e2-84eb-bf46adc7c8ba req-6ba99504-128e-4bfa-93b4-d414365a1de4 service nova] [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] Refreshing network info cache for port 446f5348-3547-40f2-b27c-3762e9b219a7 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 504.478028] env[62066]: DEBUG nova.virt.vmwareapi.images [None req-69d29524-cb64-4b08-8fc9-6a11810e0231 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] [instance: 4b34ff57-e137-4abc-8aed-76dd2ec8d313] Downloading image file data 50ff584c-3b50-4395-af07-3e66769bc9f7 to the data store datastore2 {{(pid=62066) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 504.581100] env[62066]: DEBUG oslo_vmware.rw_handles [None req-69d29524-cb64-4b08-8fc9-6a11810e0231 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c1n3.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/5600ab98-a0cc-467a-8243-420a5f1dc15f/50ff584c-3b50-4395-af07-3e66769bc9f7/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=62066) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 504.734875] env[62066]: DEBUG nova.scheduler.client.report [None req-1ca3716a-0388-4dd3-97b5-9e32fd241705 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 504.876828] env[62066]: DEBUG nova.compute.manager [None req-9eabd10d-5afe-4c06-abad-b512a51e6251 tempest-ServersAdminNegativeTestJSON-732915097 tempest-ServersAdminNegativeTestJSON-732915097-project-member] [instance: 21010917-f820-4163-848d-e4141c89b13d] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 504.908169] env[62066]: DEBUG nova.virt.hardware [None req-9eabd10d-5afe-4c06-abad-b512a51e6251 tempest-ServersAdminNegativeTestJSON-732915097 tempest-ServersAdminNegativeTestJSON-732915097-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-23T13:40:41Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-23T13:40:23Z,direct_url=,disk_format='vmdk',id=50ff584c-3b50-4395-af07-3e66769bc9f7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='eb52f7069a374c61ae946f052007c6d9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-23T13:40:24Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 504.908169] env[62066]: DEBUG nova.virt.hardware [None req-9eabd10d-5afe-4c06-abad-b512a51e6251 tempest-ServersAdminNegativeTestJSON-732915097 tempest-ServersAdminNegativeTestJSON-732915097-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 504.908169] env[62066]: DEBUG nova.virt.hardware [None req-9eabd10d-5afe-4c06-abad-b512a51e6251 tempest-ServersAdminNegativeTestJSON-732915097 tempest-ServersAdminNegativeTestJSON-732915097-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 504.911925] env[62066]: DEBUG nova.virt.hardware [None req-9eabd10d-5afe-4c06-abad-b512a51e6251 tempest-ServersAdminNegativeTestJSON-732915097 tempest-ServersAdminNegativeTestJSON-732915097-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 504.911925] env[62066]: DEBUG nova.virt.hardware [None req-9eabd10d-5afe-4c06-abad-b512a51e6251 tempest-ServersAdminNegativeTestJSON-732915097 tempest-ServersAdminNegativeTestJSON-732915097-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 504.911925] env[62066]: DEBUG nova.virt.hardware [None req-9eabd10d-5afe-4c06-abad-b512a51e6251 tempest-ServersAdminNegativeTestJSON-732915097 tempest-ServersAdminNegativeTestJSON-732915097-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 504.911925] env[62066]: DEBUG nova.virt.hardware [None req-9eabd10d-5afe-4c06-abad-b512a51e6251 tempest-ServersAdminNegativeTestJSON-732915097 tempest-ServersAdminNegativeTestJSON-732915097-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 504.911925] env[62066]: DEBUG nova.virt.hardware [None req-9eabd10d-5afe-4c06-abad-b512a51e6251 tempest-ServersAdminNegativeTestJSON-732915097 tempest-ServersAdminNegativeTestJSON-732915097-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 504.912291] env[62066]: DEBUG nova.virt.hardware [None req-9eabd10d-5afe-4c06-abad-b512a51e6251 tempest-ServersAdminNegativeTestJSON-732915097 tempest-ServersAdminNegativeTestJSON-732915097-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 504.912291] env[62066]: DEBUG nova.virt.hardware [None req-9eabd10d-5afe-4c06-abad-b512a51e6251 tempest-ServersAdminNegativeTestJSON-732915097 tempest-ServersAdminNegativeTestJSON-732915097-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 504.912291] env[62066]: DEBUG nova.virt.hardware [None req-9eabd10d-5afe-4c06-abad-b512a51e6251 tempest-ServersAdminNegativeTestJSON-732915097 tempest-ServersAdminNegativeTestJSON-732915097-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 504.912291] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c674a905-9b1a-40a3-9156-8567609141a5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 504.932674] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aed668e6-a0bf-495c-900b-96dd140e7847 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 505.109575] env[62066]: DEBUG nova.network.neutron [req-76a36dd5-4bfa-47e2-84eb-bf46adc7c8ba req-6ba99504-128e-4bfa-93b4-d414365a1de4 service nova] [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 505.243839] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1ca3716a-0388-4dd3-97b5-9e32fd241705 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.408s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 505.244461] env[62066]: DEBUG nova.compute.manager [None req-1ca3716a-0388-4dd3-97b5-9e32fd241705 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] [instance: 56eee633-2c01-4df2-8d2e-58ad712942f6] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 505.247540] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c743a8d-c557-4d89-8fe0-1cc65e783f93 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.697s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 505.249498] env[62066]: INFO nova.compute.claims [None req-5c743a8d-c557-4d89-8fe0-1cc65e783f93 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 505.290218] env[62066]: DEBUG oslo_vmware.rw_handles [None req-69d29524-cb64-4b08-8fc9-6a11810e0231 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] Completed reading data from the image iterator. {{(pid=62066) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 505.290730] env[62066]: DEBUG oslo_vmware.rw_handles [None req-69d29524-cb64-4b08-8fc9-6a11810e0231 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] Closing write handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/5600ab98-a0cc-467a-8243-420a5f1dc15f/50ff584c-3b50-4395-af07-3e66769bc9f7/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=62066) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 505.411022] env[62066]: DEBUG nova.network.neutron [req-76a36dd5-4bfa-47e2-84eb-bf46adc7c8ba req-6ba99504-128e-4bfa-93b4-d414365a1de4 service nova] [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 505.426797] env[62066]: DEBUG nova.network.neutron [None req-9eabd10d-5afe-4c06-abad-b512a51e6251 tempest-ServersAdminNegativeTestJSON-732915097 tempest-ServersAdminNegativeTestJSON-732915097-project-member] [instance: 21010917-f820-4163-848d-e4141c89b13d] Successfully created port: 358d9a83-0d84-4a7b-8e89-6754aadd6c84 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 505.435021] env[62066]: DEBUG nova.virt.vmwareapi.images [None req-69d29524-cb64-4b08-8fc9-6a11810e0231 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] [instance: 4b34ff57-e137-4abc-8aed-76dd2ec8d313] Downloaded image file data 50ff584c-3b50-4395-af07-3e66769bc9f7 to vmware_temp/5600ab98-a0cc-467a-8243-420a5f1dc15f/50ff584c-3b50-4395-af07-3e66769bc9f7/tmp-sparse.vmdk on the data store datastore2 {{(pid=62066) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 505.436826] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-69d29524-cb64-4b08-8fc9-6a11810e0231 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] [instance: 4b34ff57-e137-4abc-8aed-76dd2ec8d313] Caching image {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 505.437146] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-69d29524-cb64-4b08-8fc9-6a11810e0231 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] Copying Virtual Disk [datastore2] vmware_temp/5600ab98-a0cc-467a-8243-420a5f1dc15f/50ff584c-3b50-4395-af07-3e66769bc9f7/tmp-sparse.vmdk to [datastore2] vmware_temp/5600ab98-a0cc-467a-8243-420a5f1dc15f/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 505.437412] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-09102eb5-495c-44d3-881f-cf0692ccd6de {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 505.449886] env[62066]: DEBUG oslo_vmware.api [None req-69d29524-cb64-4b08-8fc9-6a11810e0231 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] Waiting for the task: (returnval){ [ 505.449886] env[62066]: value = "task-1340602" [ 505.449886] env[62066]: _type = "Task" [ 505.449886] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 505.459614] env[62066]: DEBUG oslo_vmware.api [None req-69d29524-cb64-4b08-8fc9-6a11810e0231 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] Task: {'id': task-1340602, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 505.756527] env[62066]: DEBUG nova.compute.utils [None req-1ca3716a-0388-4dd3-97b5-9e32fd241705 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 505.763797] env[62066]: DEBUG nova.compute.manager [None req-1ca3716a-0388-4dd3-97b5-9e32fd241705 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] [instance: 56eee633-2c01-4df2-8d2e-58ad712942f6] Not allocating networking since 'none' was specified. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 505.914209] env[62066]: DEBUG oslo_concurrency.lockutils [req-76a36dd5-4bfa-47e2-84eb-bf46adc7c8ba req-6ba99504-128e-4bfa-93b4-d414365a1de4 service nova] Releasing lock "refresh_cache-9fa0bae6-065d-49ad-b076-0956f1cc3677" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 505.914488] env[62066]: DEBUG nova.compute.manager [req-76a36dd5-4bfa-47e2-84eb-bf46adc7c8ba req-6ba99504-128e-4bfa-93b4-d414365a1de4 service nova] [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] Received event network-vif-deleted-446f5348-3547-40f2-b27c-3762e9b219a7 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 505.966604] env[62066]: DEBUG oslo_vmware.api [None req-69d29524-cb64-4b08-8fc9-6a11810e0231 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] Task: {'id': task-1340602, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 506.266029] env[62066]: DEBUG nova.compute.manager [None req-1ca3716a-0388-4dd3-97b5-9e32fd241705 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] [instance: 56eee633-2c01-4df2-8d2e-58ad712942f6] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 506.468461] env[62066]: DEBUG oslo_vmware.api [None req-69d29524-cb64-4b08-8fc9-6a11810e0231 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] Task: {'id': task-1340602, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.69335} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 506.468779] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-69d29524-cb64-4b08-8fc9-6a11810e0231 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] Copied Virtual Disk [datastore2] vmware_temp/5600ab98-a0cc-467a-8243-420a5f1dc15f/50ff584c-3b50-4395-af07-3e66769bc9f7/tmp-sparse.vmdk to [datastore2] vmware_temp/5600ab98-a0cc-467a-8243-420a5f1dc15f/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 506.468973] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-69d29524-cb64-4b08-8fc9-6a11810e0231 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] Deleting the datastore file [datastore2] vmware_temp/5600ab98-a0cc-467a-8243-420a5f1dc15f/50ff584c-3b50-4395-af07-3e66769bc9f7/tmp-sparse.vmdk {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 506.469321] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d502a7eb-35e1-4242-8669-e63d9aa1aa2a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 506.473896] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9deff003-26ca-4891-a615-026fcf120104 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 506.480245] env[62066]: DEBUG oslo_vmware.api [None req-69d29524-cb64-4b08-8fc9-6a11810e0231 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] Waiting for the task: (returnval){ [ 506.480245] env[62066]: value = "task-1340603" [ 506.480245] env[62066]: _type = "Task" [ 506.480245] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 506.493917] env[62066]: DEBUG oslo_vmware.api [None req-69d29524-cb64-4b08-8fc9-6a11810e0231 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] Task: {'id': task-1340603, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 506.499911] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a6f2fa8-5299-4cb1-bc1b-c82f80626ef8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 506.531734] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7399c870-23eb-4902-868f-c5c9230bfcb8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 506.539934] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bba11255-4d46-4b19-a7a1-063e58d4dd27 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 506.553723] env[62066]: DEBUG nova.compute.provider_tree [None req-5c743a8d-c557-4d89-8fe0-1cc65e783f93 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 506.993221] env[62066]: DEBUG oslo_vmware.api [None req-69d29524-cb64-4b08-8fc9-6a11810e0231 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] Task: {'id': task-1340603, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.024231} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 506.993480] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-69d29524-cb64-4b08-8fc9-6a11810e0231 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 506.993687] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-69d29524-cb64-4b08-8fc9-6a11810e0231 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] Moving file from [datastore2] vmware_temp/5600ab98-a0cc-467a-8243-420a5f1dc15f/50ff584c-3b50-4395-af07-3e66769bc9f7 to [datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7. {{(pid=62066) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 506.994107] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-360cc11d-d4d0-4aae-af22-40e40516e2e2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 507.001843] env[62066]: DEBUG oslo_vmware.api [None req-69d29524-cb64-4b08-8fc9-6a11810e0231 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] Waiting for the task: (returnval){ [ 507.001843] env[62066]: value = "task-1340604" [ 507.001843] env[62066]: _type = "Task" [ 507.001843] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 507.016290] env[62066]: DEBUG oslo_vmware.api [None req-69d29524-cb64-4b08-8fc9-6a11810e0231 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] Task: {'id': task-1340604, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 507.059020] env[62066]: DEBUG nova.scheduler.client.report [None req-5c743a8d-c557-4d89-8fe0-1cc65e783f93 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 507.278457] env[62066]: DEBUG nova.compute.manager [None req-1ca3716a-0388-4dd3-97b5-9e32fd241705 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] [instance: 56eee633-2c01-4df2-8d2e-58ad712942f6] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 507.314591] env[62066]: DEBUG nova.virt.hardware [None req-1ca3716a-0388-4dd3-97b5-9e32fd241705 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-23T13:40:41Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-23T13:40:23Z,direct_url=,disk_format='vmdk',id=50ff584c-3b50-4395-af07-3e66769bc9f7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='eb52f7069a374c61ae946f052007c6d9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-23T13:40:24Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 507.314817] env[62066]: DEBUG nova.virt.hardware [None req-1ca3716a-0388-4dd3-97b5-9e32fd241705 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 507.314950] env[62066]: DEBUG nova.virt.hardware [None req-1ca3716a-0388-4dd3-97b5-9e32fd241705 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 507.315138] env[62066]: DEBUG nova.virt.hardware [None req-1ca3716a-0388-4dd3-97b5-9e32fd241705 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 507.315316] env[62066]: DEBUG nova.virt.hardware [None req-1ca3716a-0388-4dd3-97b5-9e32fd241705 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 507.315417] env[62066]: DEBUG nova.virt.hardware [None req-1ca3716a-0388-4dd3-97b5-9e32fd241705 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 507.315799] env[62066]: DEBUG nova.virt.hardware [None req-1ca3716a-0388-4dd3-97b5-9e32fd241705 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 507.315935] env[62066]: DEBUG nova.virt.hardware [None req-1ca3716a-0388-4dd3-97b5-9e32fd241705 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 507.317537] env[62066]: DEBUG nova.virt.hardware [None req-1ca3716a-0388-4dd3-97b5-9e32fd241705 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 507.319863] env[62066]: DEBUG nova.virt.hardware [None req-1ca3716a-0388-4dd3-97b5-9e32fd241705 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 507.319863] env[62066]: DEBUG nova.virt.hardware [None req-1ca3716a-0388-4dd3-97b5-9e32fd241705 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 507.319863] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd9f0208-7043-43d8-a837-0e288cc8aa09 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 507.327986] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53bdfe10-0c21-4c31-94b9-9bf969890cc5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 507.347357] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-1ca3716a-0388-4dd3-97b5-9e32fd241705 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] [instance: 56eee633-2c01-4df2-8d2e-58ad712942f6] Instance VIF info [] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 507.353464] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-1ca3716a-0388-4dd3-97b5-9e32fd241705 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Creating folder: Project (6ccd43d711494bedb6504e3217e3983d). Parent ref: group-v285980. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 507.356080] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6c60b1ff-f5e3-47bc-844b-2b86b14c09f9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 507.361264] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d274d4c0-1d30-44ad-b71d-9f542d291186 tempest-VolumesAssistedSnapshotsTest-1416743605 tempest-VolumesAssistedSnapshotsTest-1416743605-project-member] Acquiring lock "1fde7ae0-c88c-4d7f-b657-355c20e2cb10" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 507.361863] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d274d4c0-1d30-44ad-b71d-9f542d291186 tempest-VolumesAssistedSnapshotsTest-1416743605 tempest-VolumesAssistedSnapshotsTest-1416743605-project-member] Lock "1fde7ae0-c88c-4d7f-b657-355c20e2cb10" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 507.372707] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-1ca3716a-0388-4dd3-97b5-9e32fd241705 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Created folder: Project (6ccd43d711494bedb6504e3217e3983d) in parent group-v285980. [ 507.372911] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-1ca3716a-0388-4dd3-97b5-9e32fd241705 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Creating folder: Instances. Parent ref: group-v285984. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 507.373393] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f2fa11f4-d701-4dc8-a794-e2cd2f32273f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 507.383157] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-1ca3716a-0388-4dd3-97b5-9e32fd241705 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Created folder: Instances in parent group-v285984. [ 507.384323] env[62066]: DEBUG oslo.service.loopingcall [None req-1ca3716a-0388-4dd3-97b5-9e32fd241705 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 507.384323] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 56eee633-2c01-4df2-8d2e-58ad712942f6] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 507.384323] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-33fcc6ae-5d92-4043-9d2e-08378caaa3de {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 507.403358] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 507.403358] env[62066]: value = "task-1340607" [ 507.403358] env[62066]: _type = "Task" [ 507.403358] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 507.411766] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1340607, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 507.516019] env[62066]: DEBUG oslo_vmware.api [None req-69d29524-cb64-4b08-8fc9-6a11810e0231 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] Task: {'id': task-1340604, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.025921} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 507.516495] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-69d29524-cb64-4b08-8fc9-6a11810e0231 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] File moved {{(pid=62066) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 507.516495] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-69d29524-cb64-4b08-8fc9-6a11810e0231 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] [instance: 4b34ff57-e137-4abc-8aed-76dd2ec8d313] Cleaning up location [datastore2] vmware_temp/5600ab98-a0cc-467a-8243-420a5f1dc15f {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 507.517176] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-69d29524-cb64-4b08-8fc9-6a11810e0231 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] Deleting the datastore file [datastore2] vmware_temp/5600ab98-a0cc-467a-8243-420a5f1dc15f {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 507.517176] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-949321ed-e99a-4295-b16f-0a1304613dc9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 507.525159] env[62066]: DEBUG oslo_vmware.api [None req-69d29524-cb64-4b08-8fc9-6a11810e0231 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] Waiting for the task: (returnval){ [ 507.525159] env[62066]: value = "task-1340608" [ 507.525159] env[62066]: _type = "Task" [ 507.525159] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 507.537112] env[62066]: DEBUG oslo_vmware.api [None req-69d29524-cb64-4b08-8fc9-6a11810e0231 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] Task: {'id': task-1340608, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 507.563141] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c743a8d-c557-4d89-8fe0-1cc65e783f93 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.315s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 507.563141] env[62066]: DEBUG nova.compute.manager [None req-5c743a8d-c557-4d89-8fe0-1cc65e783f93 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 507.566514] env[62066]: DEBUG oslo_concurrency.lockutils [None req-df85dbc4-0217-427e-9c09-82d1a0e6d774 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.012s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 507.568711] env[62066]: INFO nova.compute.claims [None req-df85dbc4-0217-427e-9c09-82d1a0e6d774 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 507.915903] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1340607, 'name': CreateVM_Task} progress is 99%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 507.943171] env[62066]: ERROR nova.compute.manager [None req-9eabd10d-5afe-4c06-abad-b512a51e6251 tempest-ServersAdminNegativeTestJSON-732915097 tempest-ServersAdminNegativeTestJSON-732915097-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 358d9a83-0d84-4a7b-8e89-6754aadd6c84, please check neutron logs for more information. [ 507.943171] env[62066]: ERROR nova.compute.manager Traceback (most recent call last): [ 507.943171] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 507.943171] env[62066]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 507.943171] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 507.943171] env[62066]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 507.943171] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 507.943171] env[62066]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 507.943171] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 507.943171] env[62066]: ERROR nova.compute.manager self.force_reraise() [ 507.943171] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 507.943171] env[62066]: ERROR nova.compute.manager raise self.value [ 507.943171] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 507.943171] env[62066]: ERROR nova.compute.manager updated_port = self._update_port( [ 507.943171] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 507.943171] env[62066]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 507.943608] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 507.943608] env[62066]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 507.943608] env[62066]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 358d9a83-0d84-4a7b-8e89-6754aadd6c84, please check neutron logs for more information. [ 507.943608] env[62066]: ERROR nova.compute.manager [ 507.943608] env[62066]: Traceback (most recent call last): [ 507.943608] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 507.943608] env[62066]: listener.cb(fileno) [ 507.943608] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 507.943608] env[62066]: result = function(*args, **kwargs) [ 507.943608] env[62066]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 507.943608] env[62066]: return func(*args, **kwargs) [ 507.943608] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 507.943608] env[62066]: raise e [ 507.943608] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 507.943608] env[62066]: nwinfo = self.network_api.allocate_for_instance( [ 507.943608] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 507.943608] env[62066]: created_port_ids = self._update_ports_for_instance( [ 507.943608] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 507.943608] env[62066]: with excutils.save_and_reraise_exception(): [ 507.943608] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 507.943608] env[62066]: self.force_reraise() [ 507.943608] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 507.943608] env[62066]: raise self.value [ 507.943608] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 507.943608] env[62066]: updated_port = self._update_port( [ 507.943608] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 507.943608] env[62066]: _ensure_no_port_binding_failure(port) [ 507.943608] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 507.943608] env[62066]: raise exception.PortBindingFailed(port_id=port['id']) [ 507.944645] env[62066]: nova.exception.PortBindingFailed: Binding failed for port 358d9a83-0d84-4a7b-8e89-6754aadd6c84, please check neutron logs for more information. [ 507.944645] env[62066]: Removing descriptor: 15 [ 507.944645] env[62066]: ERROR nova.compute.manager [None req-9eabd10d-5afe-4c06-abad-b512a51e6251 tempest-ServersAdminNegativeTestJSON-732915097 tempest-ServersAdminNegativeTestJSON-732915097-project-member] [instance: 21010917-f820-4163-848d-e4141c89b13d] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 358d9a83-0d84-4a7b-8e89-6754aadd6c84, please check neutron logs for more information. [ 507.944645] env[62066]: ERROR nova.compute.manager [instance: 21010917-f820-4163-848d-e4141c89b13d] Traceback (most recent call last): [ 507.944645] env[62066]: ERROR nova.compute.manager [instance: 21010917-f820-4163-848d-e4141c89b13d] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 507.944645] env[62066]: ERROR nova.compute.manager [instance: 21010917-f820-4163-848d-e4141c89b13d] yield resources [ 507.944645] env[62066]: ERROR nova.compute.manager [instance: 21010917-f820-4163-848d-e4141c89b13d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 507.944645] env[62066]: ERROR nova.compute.manager [instance: 21010917-f820-4163-848d-e4141c89b13d] self.driver.spawn(context, instance, image_meta, [ 507.944645] env[62066]: ERROR nova.compute.manager [instance: 21010917-f820-4163-848d-e4141c89b13d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 507.944645] env[62066]: ERROR nova.compute.manager [instance: 21010917-f820-4163-848d-e4141c89b13d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 507.944645] env[62066]: ERROR nova.compute.manager [instance: 21010917-f820-4163-848d-e4141c89b13d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 507.944645] env[62066]: ERROR nova.compute.manager [instance: 21010917-f820-4163-848d-e4141c89b13d] vm_ref = self.build_virtual_machine(instance, [ 507.945586] env[62066]: ERROR nova.compute.manager [instance: 21010917-f820-4163-848d-e4141c89b13d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 507.945586] env[62066]: ERROR nova.compute.manager [instance: 21010917-f820-4163-848d-e4141c89b13d] vif_infos = vmwarevif.get_vif_info(self._session, [ 507.945586] env[62066]: ERROR nova.compute.manager [instance: 21010917-f820-4163-848d-e4141c89b13d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 507.945586] env[62066]: ERROR nova.compute.manager [instance: 21010917-f820-4163-848d-e4141c89b13d] for vif in network_info: [ 507.945586] env[62066]: ERROR nova.compute.manager [instance: 21010917-f820-4163-848d-e4141c89b13d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 507.945586] env[62066]: ERROR nova.compute.manager [instance: 21010917-f820-4163-848d-e4141c89b13d] return self._sync_wrapper(fn, *args, **kwargs) [ 507.945586] env[62066]: ERROR nova.compute.manager [instance: 21010917-f820-4163-848d-e4141c89b13d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 507.945586] env[62066]: ERROR nova.compute.manager [instance: 21010917-f820-4163-848d-e4141c89b13d] self.wait() [ 507.945586] env[62066]: ERROR nova.compute.manager [instance: 21010917-f820-4163-848d-e4141c89b13d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 507.945586] env[62066]: ERROR nova.compute.manager [instance: 21010917-f820-4163-848d-e4141c89b13d] self[:] = self._gt.wait() [ 507.945586] env[62066]: ERROR nova.compute.manager [instance: 21010917-f820-4163-848d-e4141c89b13d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 507.945586] env[62066]: ERROR nova.compute.manager [instance: 21010917-f820-4163-848d-e4141c89b13d] return self._exit_event.wait() [ 507.945586] env[62066]: ERROR nova.compute.manager [instance: 21010917-f820-4163-848d-e4141c89b13d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 507.946141] env[62066]: ERROR nova.compute.manager [instance: 21010917-f820-4163-848d-e4141c89b13d] result = hub.switch() [ 507.946141] env[62066]: ERROR nova.compute.manager [instance: 21010917-f820-4163-848d-e4141c89b13d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 507.946141] env[62066]: ERROR nova.compute.manager [instance: 21010917-f820-4163-848d-e4141c89b13d] return self.greenlet.switch() [ 507.946141] env[62066]: ERROR nova.compute.manager [instance: 21010917-f820-4163-848d-e4141c89b13d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 507.946141] env[62066]: ERROR nova.compute.manager [instance: 21010917-f820-4163-848d-e4141c89b13d] result = function(*args, **kwargs) [ 507.946141] env[62066]: ERROR nova.compute.manager [instance: 21010917-f820-4163-848d-e4141c89b13d] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 507.946141] env[62066]: ERROR nova.compute.manager [instance: 21010917-f820-4163-848d-e4141c89b13d] return func(*args, **kwargs) [ 507.946141] env[62066]: ERROR nova.compute.manager [instance: 21010917-f820-4163-848d-e4141c89b13d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 507.946141] env[62066]: ERROR nova.compute.manager [instance: 21010917-f820-4163-848d-e4141c89b13d] raise e [ 507.946141] env[62066]: ERROR nova.compute.manager [instance: 21010917-f820-4163-848d-e4141c89b13d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 507.946141] env[62066]: ERROR nova.compute.manager [instance: 21010917-f820-4163-848d-e4141c89b13d] nwinfo = self.network_api.allocate_for_instance( [ 507.946141] env[62066]: ERROR nova.compute.manager [instance: 21010917-f820-4163-848d-e4141c89b13d] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 507.946141] env[62066]: ERROR nova.compute.manager [instance: 21010917-f820-4163-848d-e4141c89b13d] created_port_ids = self._update_ports_for_instance( [ 507.946628] env[62066]: ERROR nova.compute.manager [instance: 21010917-f820-4163-848d-e4141c89b13d] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 507.946628] env[62066]: ERROR nova.compute.manager [instance: 21010917-f820-4163-848d-e4141c89b13d] with excutils.save_and_reraise_exception(): [ 507.946628] env[62066]: ERROR nova.compute.manager [instance: 21010917-f820-4163-848d-e4141c89b13d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 507.946628] env[62066]: ERROR nova.compute.manager [instance: 21010917-f820-4163-848d-e4141c89b13d] self.force_reraise() [ 507.946628] env[62066]: ERROR nova.compute.manager [instance: 21010917-f820-4163-848d-e4141c89b13d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 507.946628] env[62066]: ERROR nova.compute.manager [instance: 21010917-f820-4163-848d-e4141c89b13d] raise self.value [ 507.946628] env[62066]: ERROR nova.compute.manager [instance: 21010917-f820-4163-848d-e4141c89b13d] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 507.946628] env[62066]: ERROR nova.compute.manager [instance: 21010917-f820-4163-848d-e4141c89b13d] updated_port = self._update_port( [ 507.946628] env[62066]: ERROR nova.compute.manager [instance: 21010917-f820-4163-848d-e4141c89b13d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 507.946628] env[62066]: ERROR nova.compute.manager [instance: 21010917-f820-4163-848d-e4141c89b13d] _ensure_no_port_binding_failure(port) [ 507.946628] env[62066]: ERROR nova.compute.manager [instance: 21010917-f820-4163-848d-e4141c89b13d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 507.946628] env[62066]: ERROR nova.compute.manager [instance: 21010917-f820-4163-848d-e4141c89b13d] raise exception.PortBindingFailed(port_id=port['id']) [ 507.947128] env[62066]: ERROR nova.compute.manager [instance: 21010917-f820-4163-848d-e4141c89b13d] nova.exception.PortBindingFailed: Binding failed for port 358d9a83-0d84-4a7b-8e89-6754aadd6c84, please check neutron logs for more information. [ 507.947128] env[62066]: ERROR nova.compute.manager [instance: 21010917-f820-4163-848d-e4141c89b13d] [ 507.947128] env[62066]: INFO nova.compute.manager [None req-9eabd10d-5afe-4c06-abad-b512a51e6251 tempest-ServersAdminNegativeTestJSON-732915097 tempest-ServersAdminNegativeTestJSON-732915097-project-member] [instance: 21010917-f820-4163-848d-e4141c89b13d] Terminating instance [ 507.948592] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9eabd10d-5afe-4c06-abad-b512a51e6251 tempest-ServersAdminNegativeTestJSON-732915097 tempest-ServersAdminNegativeTestJSON-732915097-project-member] Acquiring lock "refresh_cache-21010917-f820-4163-848d-e4141c89b13d" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 507.949113] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9eabd10d-5afe-4c06-abad-b512a51e6251 tempest-ServersAdminNegativeTestJSON-732915097 tempest-ServersAdminNegativeTestJSON-732915097-project-member] Acquired lock "refresh_cache-21010917-f820-4163-848d-e4141c89b13d" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 507.949113] env[62066]: DEBUG nova.network.neutron [None req-9eabd10d-5afe-4c06-abad-b512a51e6251 tempest-ServersAdminNegativeTestJSON-732915097 tempest-ServersAdminNegativeTestJSON-732915097-project-member] [instance: 21010917-f820-4163-848d-e4141c89b13d] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 508.036463] env[62066]: DEBUG oslo_vmware.api [None req-69d29524-cb64-4b08-8fc9-6a11810e0231 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] Task: {'id': task-1340608, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.028215} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 508.036769] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-69d29524-cb64-4b08-8fc9-6a11810e0231 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 508.037628] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8cd3cdf5-974e-4e98-8c1d-4f86bba8665a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 508.047438] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c4a0feae-1b12-4166-bd78-82a96fe8ac4c tempest-ServersWithSpecificFlavorTestJSON-184391817 tempest-ServersWithSpecificFlavorTestJSON-184391817-project-member] Acquiring lock "4c67e82d-b51b-4820-af81-e6eefc7bcb36" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 508.047438] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c4a0feae-1b12-4166-bd78-82a96fe8ac4c tempest-ServersWithSpecificFlavorTestJSON-184391817 tempest-ServersWithSpecificFlavorTestJSON-184391817-project-member] Lock "4c67e82d-b51b-4820-af81-e6eefc7bcb36" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 508.048571] env[62066]: DEBUG oslo_vmware.api [None req-69d29524-cb64-4b08-8fc9-6a11810e0231 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] Waiting for the task: (returnval){ [ 508.048571] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]52de9488-9513-bfd2-33fd-ff9909458c3f" [ 508.048571] env[62066]: _type = "Task" [ 508.048571] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 508.056881] env[62066]: DEBUG oslo_vmware.api [None req-69d29524-cb64-4b08-8fc9-6a11810e0231 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52de9488-9513-bfd2-33fd-ff9909458c3f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 508.075614] env[62066]: DEBUG nova.compute.utils [None req-5c743a8d-c557-4d89-8fe0-1cc65e783f93 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 508.077210] env[62066]: DEBUG nova.compute.manager [None req-5c743a8d-c557-4d89-8fe0-1cc65e783f93 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 508.077336] env[62066]: DEBUG nova.network.neutron [None req-5c743a8d-c557-4d89-8fe0-1cc65e783f93 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 508.153444] env[62066]: DEBUG nova.policy [None req-5c743a8d-c557-4d89-8fe0-1cc65e783f93 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '22afd2db22164c9da7ecbc76c58c68dc', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '975c88ff0be74b81a20da760ffc8e667', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 508.244460] env[62066]: DEBUG nova.compute.manager [req-a2ce1daa-e637-4b10-b90b-4838b4906140 req-3ba0bdfa-f812-4aef-bb94-75179080dada service nova] [instance: 21010917-f820-4163-848d-e4141c89b13d] Received event network-changed-358d9a83-0d84-4a7b-8e89-6754aadd6c84 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 508.244569] env[62066]: DEBUG nova.compute.manager [req-a2ce1daa-e637-4b10-b90b-4838b4906140 req-3ba0bdfa-f812-4aef-bb94-75179080dada service nova] [instance: 21010917-f820-4163-848d-e4141c89b13d] Refreshing instance network info cache due to event network-changed-358d9a83-0d84-4a7b-8e89-6754aadd6c84. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 508.245080] env[62066]: DEBUG oslo_concurrency.lockutils [req-a2ce1daa-e637-4b10-b90b-4838b4906140 req-3ba0bdfa-f812-4aef-bb94-75179080dada service nova] Acquiring lock "refresh_cache-21010917-f820-4163-848d-e4141c89b13d" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 508.418440] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1340607, 'name': CreateVM_Task} progress is 99%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 508.487994] env[62066]: DEBUG nova.network.neutron [None req-9eabd10d-5afe-4c06-abad-b512a51e6251 tempest-ServersAdminNegativeTestJSON-732915097 tempest-ServersAdminNegativeTestJSON-732915097-project-member] [instance: 21010917-f820-4163-848d-e4141c89b13d] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 508.564406] env[62066]: DEBUG oslo_vmware.api [None req-69d29524-cb64-4b08-8fc9-6a11810e0231 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52de9488-9513-bfd2-33fd-ff9909458c3f, 'name': SearchDatastore_Task, 'duration_secs': 0.009986} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 508.564756] env[62066]: DEBUG oslo_concurrency.lockutils [None req-69d29524-cb64-4b08-8fc9-6a11810e0231 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] Releasing lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 508.564756] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-69d29524-cb64-4b08-8fc9-6a11810e0231 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk to [datastore2] 4b34ff57-e137-4abc-8aed-76dd2ec8d313/4b34ff57-e137-4abc-8aed-76dd2ec8d313.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 508.567256] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-38abce08-918c-477b-a1c9-75048e930972 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 508.573513] env[62066]: DEBUG oslo_vmware.api [None req-69d29524-cb64-4b08-8fc9-6a11810e0231 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] Waiting for the task: (returnval){ [ 508.573513] env[62066]: value = "task-1340609" [ 508.573513] env[62066]: _type = "Task" [ 508.573513] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 508.581374] env[62066]: DEBUG nova.compute.manager [None req-5c743a8d-c557-4d89-8fe0-1cc65e783f93 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 508.588149] env[62066]: DEBUG oslo_vmware.api [None req-69d29524-cb64-4b08-8fc9-6a11810e0231 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] Task: {'id': task-1340609, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 508.620311] env[62066]: DEBUG nova.network.neutron [None req-9eabd10d-5afe-4c06-abad-b512a51e6251 tempest-ServersAdminNegativeTestJSON-732915097 tempest-ServersAdminNegativeTestJSON-732915097-project-member] [instance: 21010917-f820-4163-848d-e4141c89b13d] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 508.630971] env[62066]: DEBUG nova.network.neutron [None req-5c743a8d-c557-4d89-8fe0-1cc65e783f93 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] Successfully created port: 33397b47-1617-4402-98e4-8e41bebcc4a5 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 508.867265] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb80b909-2c54-4c37-be8b-264d5265dbbd {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 508.875992] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e5e8a61-1515-4259-a4df-0624492ed570 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 508.915691] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fc7da46-60b0-497f-9898-949cd1896606 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 508.923934] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1340607, 'name': CreateVM_Task, 'duration_secs': 1.326215} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 508.926125] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 56eee633-2c01-4df2-8d2e-58ad712942f6] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 508.926634] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1ca3716a-0388-4dd3-97b5-9e32fd241705 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 508.926796] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1ca3716a-0388-4dd3-97b5-9e32fd241705 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Acquired lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 508.927133] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1ca3716a-0388-4dd3-97b5-9e32fd241705 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 508.928361] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84dc7383-dfc1-4e26-99ef-e4e5055450cc {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 508.932376] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-eba6d00b-c288-4080-aac3-554e8058a775 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 508.945328] env[62066]: DEBUG nova.compute.provider_tree [None req-df85dbc4-0217-427e-9c09-82d1a0e6d774 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 508.948116] env[62066]: DEBUG oslo_vmware.api [None req-1ca3716a-0388-4dd3-97b5-9e32fd241705 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Waiting for the task: (returnval){ [ 508.948116] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]52f8dab8-a3d7-25e9-63bd-f5eaad7dc77b" [ 508.948116] env[62066]: _type = "Task" [ 508.948116] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 508.956536] env[62066]: DEBUG oslo_vmware.api [None req-1ca3716a-0388-4dd3-97b5-9e32fd241705 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52f8dab8-a3d7-25e9-63bd-f5eaad7dc77b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 509.093187] env[62066]: DEBUG oslo_vmware.api [None req-69d29524-cb64-4b08-8fc9-6a11810e0231 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] Task: {'id': task-1340609, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 509.125726] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9eabd10d-5afe-4c06-abad-b512a51e6251 tempest-ServersAdminNegativeTestJSON-732915097 tempest-ServersAdminNegativeTestJSON-732915097-project-member] Releasing lock "refresh_cache-21010917-f820-4163-848d-e4141c89b13d" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 509.126054] env[62066]: DEBUG nova.compute.manager [None req-9eabd10d-5afe-4c06-abad-b512a51e6251 tempest-ServersAdminNegativeTestJSON-732915097 tempest-ServersAdminNegativeTestJSON-732915097-project-member] [instance: 21010917-f820-4163-848d-e4141c89b13d] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 509.126251] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-9eabd10d-5afe-4c06-abad-b512a51e6251 tempest-ServersAdminNegativeTestJSON-732915097 tempest-ServersAdminNegativeTestJSON-732915097-project-member] [instance: 21010917-f820-4163-848d-e4141c89b13d] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 509.126577] env[62066]: DEBUG oslo_concurrency.lockutils [req-a2ce1daa-e637-4b10-b90b-4838b4906140 req-3ba0bdfa-f812-4aef-bb94-75179080dada service nova] Acquired lock "refresh_cache-21010917-f820-4163-848d-e4141c89b13d" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 509.126744] env[62066]: DEBUG nova.network.neutron [req-a2ce1daa-e637-4b10-b90b-4838b4906140 req-3ba0bdfa-f812-4aef-bb94-75179080dada service nova] [instance: 21010917-f820-4163-848d-e4141c89b13d] Refreshing network info cache for port 358d9a83-0d84-4a7b-8e89-6754aadd6c84 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 509.127840] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-da773829-39b3-4f02-8bc1-24d8c9b212bb {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 509.143838] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6ca174c-91a3-4339-9771-76701795f84f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 509.176375] env[62066]: WARNING nova.virt.vmwareapi.vmops [None req-9eabd10d-5afe-4c06-abad-b512a51e6251 tempest-ServersAdminNegativeTestJSON-732915097 tempest-ServersAdminNegativeTestJSON-732915097-project-member] [instance: 21010917-f820-4163-848d-e4141c89b13d] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 21010917-f820-4163-848d-e4141c89b13d could not be found. [ 509.176607] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-9eabd10d-5afe-4c06-abad-b512a51e6251 tempest-ServersAdminNegativeTestJSON-732915097 tempest-ServersAdminNegativeTestJSON-732915097-project-member] [instance: 21010917-f820-4163-848d-e4141c89b13d] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 509.176779] env[62066]: INFO nova.compute.manager [None req-9eabd10d-5afe-4c06-abad-b512a51e6251 tempest-ServersAdminNegativeTestJSON-732915097 tempest-ServersAdminNegativeTestJSON-732915097-project-member] [instance: 21010917-f820-4163-848d-e4141c89b13d] Took 0.05 seconds to destroy the instance on the hypervisor. [ 509.177025] env[62066]: DEBUG oslo.service.loopingcall [None req-9eabd10d-5afe-4c06-abad-b512a51e6251 tempest-ServersAdminNegativeTestJSON-732915097 tempest-ServersAdminNegativeTestJSON-732915097-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 509.177472] env[62066]: DEBUG nova.compute.manager [-] [instance: 21010917-f820-4163-848d-e4141c89b13d] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 509.177560] env[62066]: DEBUG nova.network.neutron [-] [instance: 21010917-f820-4163-848d-e4141c89b13d] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 509.208554] env[62066]: DEBUG nova.network.neutron [-] [instance: 21010917-f820-4163-848d-e4141c89b13d] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 509.449956] env[62066]: DEBUG nova.scheduler.client.report [None req-df85dbc4-0217-427e-9c09-82d1a0e6d774 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 509.465665] env[62066]: DEBUG oslo_vmware.api [None req-1ca3716a-0388-4dd3-97b5-9e32fd241705 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52f8dab8-a3d7-25e9-63bd-f5eaad7dc77b, 'name': SearchDatastore_Task, 'duration_secs': 0.069582} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 509.466075] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1ca3716a-0388-4dd3-97b5-9e32fd241705 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Releasing lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 509.466349] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-1ca3716a-0388-4dd3-97b5-9e32fd241705 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] [instance: 56eee633-2c01-4df2-8d2e-58ad712942f6] Processing image 50ff584c-3b50-4395-af07-3e66769bc9f7 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 509.466617] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1ca3716a-0388-4dd3-97b5-9e32fd241705 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 509.466801] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1ca3716a-0388-4dd3-97b5-9e32fd241705 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Acquired lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 509.467027] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-1ca3716a-0388-4dd3-97b5-9e32fd241705 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 509.467343] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-194b1485-d0f4-4ce9-b14b-fd12aedb5444 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 509.475530] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-1ca3716a-0388-4dd3-97b5-9e32fd241705 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 509.475706] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-1ca3716a-0388-4dd3-97b5-9e32fd241705 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 509.477903] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-45b846fd-938e-4000-a869-b4d7c409b670 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 509.484272] env[62066]: DEBUG oslo_vmware.api [None req-1ca3716a-0388-4dd3-97b5-9e32fd241705 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Waiting for the task: (returnval){ [ 509.484272] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]5286c588-6aae-14db-15e1-6894660acc6c" [ 509.484272] env[62066]: _type = "Task" [ 509.484272] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 509.493901] env[62066]: DEBUG oslo_vmware.api [None req-1ca3716a-0388-4dd3-97b5-9e32fd241705 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]5286c588-6aae-14db-15e1-6894660acc6c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 509.586219] env[62066]: DEBUG oslo_vmware.api [None req-69d29524-cb64-4b08-8fc9-6a11810e0231 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] Task: {'id': task-1340609, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.560393} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 509.587378] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-69d29524-cb64-4b08-8fc9-6a11810e0231 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk to [datastore2] 4b34ff57-e137-4abc-8aed-76dd2ec8d313/4b34ff57-e137-4abc-8aed-76dd2ec8d313.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 509.587622] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-69d29524-cb64-4b08-8fc9-6a11810e0231 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] [instance: 4b34ff57-e137-4abc-8aed-76dd2ec8d313] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 509.587974] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-95edf750-19ca-4462-bef7-2a8d5903f481 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 509.598152] env[62066]: DEBUG nova.compute.manager [None req-5c743a8d-c557-4d89-8fe0-1cc65e783f93 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 509.600640] env[62066]: DEBUG oslo_vmware.api [None req-69d29524-cb64-4b08-8fc9-6a11810e0231 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] Waiting for the task: (returnval){ [ 509.600640] env[62066]: value = "task-1340610" [ 509.600640] env[62066]: _type = "Task" [ 509.600640] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 509.611219] env[62066]: DEBUG oslo_vmware.api [None req-69d29524-cb64-4b08-8fc9-6a11810e0231 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] Task: {'id': task-1340610, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 509.627522] env[62066]: DEBUG nova.virt.hardware [None req-5c743a8d-c557-4d89-8fe0-1cc65e783f93 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-23T13:40:41Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-23T13:40:23Z,direct_url=,disk_format='vmdk',id=50ff584c-3b50-4395-af07-3e66769bc9f7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='eb52f7069a374c61ae946f052007c6d9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-23T13:40:24Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 509.627522] env[62066]: DEBUG nova.virt.hardware [None req-5c743a8d-c557-4d89-8fe0-1cc65e783f93 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 509.627522] env[62066]: DEBUG nova.virt.hardware [None req-5c743a8d-c557-4d89-8fe0-1cc65e783f93 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 509.627899] env[62066]: DEBUG nova.virt.hardware [None req-5c743a8d-c557-4d89-8fe0-1cc65e783f93 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 509.628143] env[62066]: DEBUG nova.virt.hardware [None req-5c743a8d-c557-4d89-8fe0-1cc65e783f93 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 509.628143] env[62066]: DEBUG nova.virt.hardware [None req-5c743a8d-c557-4d89-8fe0-1cc65e783f93 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 509.628788] env[62066]: DEBUG nova.virt.hardware [None req-5c743a8d-c557-4d89-8fe0-1cc65e783f93 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 509.628788] env[62066]: DEBUG nova.virt.hardware [None req-5c743a8d-c557-4d89-8fe0-1cc65e783f93 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 509.628788] env[62066]: DEBUG nova.virt.hardware [None req-5c743a8d-c557-4d89-8fe0-1cc65e783f93 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 509.628963] env[62066]: DEBUG nova.virt.hardware [None req-5c743a8d-c557-4d89-8fe0-1cc65e783f93 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 509.629262] env[62066]: DEBUG nova.virt.hardware [None req-5c743a8d-c557-4d89-8fe0-1cc65e783f93 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 509.629973] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4097b836-426d-470c-842a-f5dadd6abdde {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 509.640514] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0655e82b-faae-485a-88ae-2ba0a4598e93 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 509.706932] env[62066]: DEBUG nova.network.neutron [req-a2ce1daa-e637-4b10-b90b-4838b4906140 req-3ba0bdfa-f812-4aef-bb94-75179080dada service nova] [instance: 21010917-f820-4163-848d-e4141c89b13d] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 509.711185] env[62066]: DEBUG nova.network.neutron [-] [instance: 21010917-f820-4163-848d-e4141c89b13d] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 509.908483] env[62066]: DEBUG nova.network.neutron [req-a2ce1daa-e637-4b10-b90b-4838b4906140 req-3ba0bdfa-f812-4aef-bb94-75179080dada service nova] [instance: 21010917-f820-4163-848d-e4141c89b13d] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 509.958464] env[62066]: DEBUG oslo_concurrency.lockutils [None req-df85dbc4-0217-427e-9c09-82d1a0e6d774 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.392s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 509.959776] env[62066]: DEBUG nova.compute.manager [None req-df85dbc4-0217-427e-9c09-82d1a0e6d774 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 509.962646] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0e23bf5c-5716-4eba-9510-a3597f398645 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.574s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 509.966040] env[62066]: INFO nova.compute.claims [None req-0e23bf5c-5716-4eba-9510-a3597f398645 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 510.000084] env[62066]: DEBUG oslo_vmware.api [None req-1ca3716a-0388-4dd3-97b5-9e32fd241705 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]5286c588-6aae-14db-15e1-6894660acc6c, 'name': SearchDatastore_Task, 'duration_secs': 0.008567} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 510.001206] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-beef1718-f053-4fd4-8213-5fa2921da165 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 510.012150] env[62066]: DEBUG oslo_vmware.api [None req-1ca3716a-0388-4dd3-97b5-9e32fd241705 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Waiting for the task: (returnval){ [ 510.012150] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]523e7d82-da0e-9bb1-b60c-0f556f486b5f" [ 510.012150] env[62066]: _type = "Task" [ 510.012150] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 510.018959] env[62066]: DEBUG oslo_vmware.api [None req-1ca3716a-0388-4dd3-97b5-9e32fd241705 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]523e7d82-da0e-9bb1-b60c-0f556f486b5f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 510.119018] env[62066]: DEBUG oslo_vmware.api [None req-69d29524-cb64-4b08-8fc9-6a11810e0231 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] Task: {'id': task-1340610, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.077369} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 510.120668] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-69d29524-cb64-4b08-8fc9-6a11810e0231 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] [instance: 4b34ff57-e137-4abc-8aed-76dd2ec8d313] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 510.124181] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a70552e1-fa74-4f2c-8329-4dbd01be5dcd {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 510.171029] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-69d29524-cb64-4b08-8fc9-6a11810e0231 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] [instance: 4b34ff57-e137-4abc-8aed-76dd2ec8d313] Reconfiguring VM instance instance-00000003 to attach disk [datastore2] 4b34ff57-e137-4abc-8aed-76dd2ec8d313/4b34ff57-e137-4abc-8aed-76dd2ec8d313.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 510.171029] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-dd8796bf-099e-4235-9e88-b7fe997414a2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 510.200455] env[62066]: DEBUG oslo_vmware.api [None req-69d29524-cb64-4b08-8fc9-6a11810e0231 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] Waiting for the task: (returnval){ [ 510.200455] env[62066]: value = "task-1340611" [ 510.200455] env[62066]: _type = "Task" [ 510.200455] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 510.211259] env[62066]: DEBUG oslo_vmware.api [None req-69d29524-cb64-4b08-8fc9-6a11810e0231 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] Task: {'id': task-1340611, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 510.214562] env[62066]: INFO nova.compute.manager [-] [instance: 21010917-f820-4163-848d-e4141c89b13d] Took 1.04 seconds to deallocate network for instance. [ 510.217335] env[62066]: DEBUG nova.compute.claims [None req-9eabd10d-5afe-4c06-abad-b512a51e6251 tempest-ServersAdminNegativeTestJSON-732915097 tempest-ServersAdminNegativeTestJSON-732915097-project-member] [instance: 21010917-f820-4163-848d-e4141c89b13d] Aborting claim: {{(pid=62066) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 510.217571] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9eabd10d-5afe-4c06-abad-b512a51e6251 tempest-ServersAdminNegativeTestJSON-732915097 tempest-ServersAdminNegativeTestJSON-732915097-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 510.411454] env[62066]: DEBUG oslo_concurrency.lockutils [req-a2ce1daa-e637-4b10-b90b-4838b4906140 req-3ba0bdfa-f812-4aef-bb94-75179080dada service nova] Releasing lock "refresh_cache-21010917-f820-4163-848d-e4141c89b13d" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 510.468571] env[62066]: DEBUG nova.compute.utils [None req-df85dbc4-0217-427e-9c09-82d1a0e6d774 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 510.473764] env[62066]: DEBUG nova.compute.manager [None req-df85dbc4-0217-427e-9c09-82d1a0e6d774 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 510.473888] env[62066]: DEBUG nova.network.neutron [None req-df85dbc4-0217-427e-9c09-82d1a0e6d774 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 510.530316] env[62066]: DEBUG oslo_vmware.api [None req-1ca3716a-0388-4dd3-97b5-9e32fd241705 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]523e7d82-da0e-9bb1-b60c-0f556f486b5f, 'name': SearchDatastore_Task, 'duration_secs': 0.008867} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 510.531569] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1ca3716a-0388-4dd3-97b5-9e32fd241705 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Releasing lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 510.531569] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-1ca3716a-0388-4dd3-97b5-9e32fd241705 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk to [datastore2] 56eee633-2c01-4df2-8d2e-58ad712942f6/56eee633-2c01-4df2-8d2e-58ad712942f6.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 510.531569] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-fdefb584-a2a8-436f-8711-ac619885c420 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 510.541825] env[62066]: DEBUG oslo_vmware.api [None req-1ca3716a-0388-4dd3-97b5-9e32fd241705 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Waiting for the task: (returnval){ [ 510.541825] env[62066]: value = "task-1340612" [ 510.541825] env[62066]: _type = "Task" [ 510.541825] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 510.554221] env[62066]: DEBUG oslo_vmware.api [None req-1ca3716a-0388-4dd3-97b5-9e32fd241705 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Task: {'id': task-1340612, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 510.581328] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cb29a201-afba-4eb8-8336-fe27a681c8f8 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] Acquiring lock "69d99063-a6a7-4832-a864-dc9fd6024ea8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 510.581554] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cb29a201-afba-4eb8-8336-fe27a681c8f8 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] Lock "69d99063-a6a7-4832-a864-dc9fd6024ea8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 510.618241] env[62066]: DEBUG nova.policy [None req-df85dbc4-0217-427e-9c09-82d1a0e6d774 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ef2bdd5efd6b4a4e8287c63007defdc4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a7fb7655514d47fa88c7fccb28665caa', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 510.714429] env[62066]: DEBUG oslo_vmware.api [None req-69d29524-cb64-4b08-8fc9-6a11810e0231 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] Task: {'id': task-1340611, 'name': ReconfigVM_Task, 'duration_secs': 0.294685} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 510.715100] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-69d29524-cb64-4b08-8fc9-6a11810e0231 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] [instance: 4b34ff57-e137-4abc-8aed-76dd2ec8d313] Reconfigured VM instance instance-00000003 to attach disk [datastore2] 4b34ff57-e137-4abc-8aed-76dd2ec8d313/4b34ff57-e137-4abc-8aed-76dd2ec8d313.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 510.716090] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f567b9ff-ef06-412e-9153-409cac846f83 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 510.726231] env[62066]: DEBUG oslo_vmware.api [None req-69d29524-cb64-4b08-8fc9-6a11810e0231 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] Waiting for the task: (returnval){ [ 510.726231] env[62066]: value = "task-1340613" [ 510.726231] env[62066]: _type = "Task" [ 510.726231] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 510.734540] env[62066]: DEBUG oslo_vmware.api [None req-69d29524-cb64-4b08-8fc9-6a11810e0231 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] Task: {'id': task-1340613, 'name': Rename_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 510.896433] env[62066]: DEBUG nova.compute.manager [req-992f08c5-7e94-4fd9-90d8-5f3d7471ca56 req-df2eb372-b48e-47f9-80d8-6531cf63011d service nova] [instance: 21010917-f820-4163-848d-e4141c89b13d] Received event network-vif-deleted-358d9a83-0d84-4a7b-8e89-6754aadd6c84 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 510.977257] env[62066]: DEBUG nova.compute.manager [None req-df85dbc4-0217-427e-9c09-82d1a0e6d774 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 511.056930] env[62066]: DEBUG oslo_vmware.api [None req-1ca3716a-0388-4dd3-97b5-9e32fd241705 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Task: {'id': task-1340612, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.502664} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 511.057278] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-1ca3716a-0388-4dd3-97b5-9e32fd241705 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk to [datastore2] 56eee633-2c01-4df2-8d2e-58ad712942f6/56eee633-2c01-4df2-8d2e-58ad712942f6.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 511.057560] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-1ca3716a-0388-4dd3-97b5-9e32fd241705 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] [instance: 56eee633-2c01-4df2-8d2e-58ad712942f6] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 511.057837] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d20ff843-8d3c-434b-8c8a-0a972dd3351d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 511.067889] env[62066]: DEBUG oslo_vmware.api [None req-1ca3716a-0388-4dd3-97b5-9e32fd241705 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Waiting for the task: (returnval){ [ 511.067889] env[62066]: value = "task-1340614" [ 511.067889] env[62066]: _type = "Task" [ 511.067889] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 511.077439] env[62066]: DEBUG oslo_vmware.api [None req-1ca3716a-0388-4dd3-97b5-9e32fd241705 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Task: {'id': task-1340614, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 511.240894] env[62066]: DEBUG oslo_vmware.api [None req-69d29524-cb64-4b08-8fc9-6a11810e0231 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] Task: {'id': task-1340613, 'name': Rename_Task, 'duration_secs': 0.222085} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 511.241175] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-69d29524-cb64-4b08-8fc9-6a11810e0231 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] [instance: 4b34ff57-e137-4abc-8aed-76dd2ec8d313] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 511.241409] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-96dc39ba-a4a1-452d-bc82-6475c2d2224e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 511.252632] env[62066]: DEBUG oslo_vmware.api [None req-69d29524-cb64-4b08-8fc9-6a11810e0231 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] Waiting for the task: (returnval){ [ 511.252632] env[62066]: value = "task-1340615" [ 511.252632] env[62066]: _type = "Task" [ 511.252632] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 511.253224] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-052b739a-793f-4800-9f32-ac6ba3740e3c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 511.269465] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cde27e59-c576-4d4d-b2d8-30b095254f44 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 511.272612] env[62066]: DEBUG oslo_vmware.api [None req-69d29524-cb64-4b08-8fc9-6a11810e0231 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] Task: {'id': task-1340615, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 511.302712] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a2982a4-71df-4d0f-8479-e113a83f5293 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 511.310250] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4378b7c2-3470-485b-8268-d9fd5a901b9e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 511.324569] env[62066]: DEBUG nova.compute.provider_tree [None req-0e23bf5c-5716-4eba-9510-a3597f398645 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 511.581996] env[62066]: DEBUG oslo_vmware.api [None req-1ca3716a-0388-4dd3-97b5-9e32fd241705 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Task: {'id': task-1340614, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070511} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 511.583540] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-1ca3716a-0388-4dd3-97b5-9e32fd241705 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] [instance: 56eee633-2c01-4df2-8d2e-58ad712942f6] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 511.587156] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85549f67-a3f9-4b73-aab5-501877d3a67d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 511.619746] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-1ca3716a-0388-4dd3-97b5-9e32fd241705 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] [instance: 56eee633-2c01-4df2-8d2e-58ad712942f6] Reconfiguring VM instance instance-00000005 to attach disk [datastore2] 56eee633-2c01-4df2-8d2e-58ad712942f6/56eee633-2c01-4df2-8d2e-58ad712942f6.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 511.621088] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1b56d79d-239d-4546-9ae9-69f251d20662 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 511.646484] env[62066]: DEBUG oslo_vmware.api [None req-1ca3716a-0388-4dd3-97b5-9e32fd241705 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Waiting for the task: (returnval){ [ 511.646484] env[62066]: value = "task-1340616" [ 511.646484] env[62066]: _type = "Task" [ 511.646484] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 511.658034] env[62066]: DEBUG oslo_vmware.api [None req-1ca3716a-0388-4dd3-97b5-9e32fd241705 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Task: {'id': task-1340616, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 511.754950] env[62066]: DEBUG nova.network.neutron [None req-df85dbc4-0217-427e-9c09-82d1a0e6d774 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] Successfully created port: e279d463-9a92-4d4b-913e-d08e912e8b3e {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 511.774744] env[62066]: DEBUG oslo_vmware.api [None req-69d29524-cb64-4b08-8fc9-6a11810e0231 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] Task: {'id': task-1340615, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 511.828611] env[62066]: DEBUG nova.scheduler.client.report [None req-0e23bf5c-5716-4eba-9510-a3597f398645 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 511.991531] env[62066]: DEBUG nova.compute.manager [None req-df85dbc4-0217-427e-9c09-82d1a0e6d774 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 512.022454] env[62066]: DEBUG nova.virt.hardware [None req-df85dbc4-0217-427e-9c09-82d1a0e6d774 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-23T13:40:41Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-23T13:40:23Z,direct_url=,disk_format='vmdk',id=50ff584c-3b50-4395-af07-3e66769bc9f7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='eb52f7069a374c61ae946f052007c6d9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-23T13:40:24Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 512.022454] env[62066]: DEBUG nova.virt.hardware [None req-df85dbc4-0217-427e-9c09-82d1a0e6d774 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 512.022454] env[62066]: DEBUG nova.virt.hardware [None req-df85dbc4-0217-427e-9c09-82d1a0e6d774 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 512.022725] env[62066]: DEBUG nova.virt.hardware [None req-df85dbc4-0217-427e-9c09-82d1a0e6d774 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 512.022771] env[62066]: DEBUG nova.virt.hardware [None req-df85dbc4-0217-427e-9c09-82d1a0e6d774 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 512.022926] env[62066]: DEBUG nova.virt.hardware [None req-df85dbc4-0217-427e-9c09-82d1a0e6d774 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 512.023504] env[62066]: DEBUG nova.virt.hardware [None req-df85dbc4-0217-427e-9c09-82d1a0e6d774 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 512.023732] env[62066]: DEBUG nova.virt.hardware [None req-df85dbc4-0217-427e-9c09-82d1a0e6d774 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 512.023842] env[62066]: DEBUG nova.virt.hardware [None req-df85dbc4-0217-427e-9c09-82d1a0e6d774 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 512.024368] env[62066]: DEBUG nova.virt.hardware [None req-df85dbc4-0217-427e-9c09-82d1a0e6d774 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 512.024442] env[62066]: DEBUG nova.virt.hardware [None req-df85dbc4-0217-427e-9c09-82d1a0e6d774 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 512.025695] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0efc0347-ff75-4a6f-be9b-3e2bc133aa0d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 512.038616] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fac1e5a-8ba6-47f4-b427-5e2df63ae075 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 512.160439] env[62066]: DEBUG oslo_vmware.api [None req-1ca3716a-0388-4dd3-97b5-9e32fd241705 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Task: {'id': task-1340616, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 512.275585] env[62066]: DEBUG oslo_vmware.api [None req-69d29524-cb64-4b08-8fc9-6a11810e0231 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] Task: {'id': task-1340615, 'name': PowerOnVM_Task, 'duration_secs': 0.545311} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 512.276821] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-69d29524-cb64-4b08-8fc9-6a11810e0231 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] [instance: 4b34ff57-e137-4abc-8aed-76dd2ec8d313] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 512.278485] env[62066]: INFO nova.compute.manager [None req-69d29524-cb64-4b08-8fc9-6a11810e0231 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] [instance: 4b34ff57-e137-4abc-8aed-76dd2ec8d313] Took 9.75 seconds to spawn the instance on the hypervisor. [ 512.278485] env[62066]: DEBUG nova.compute.manager [None req-69d29524-cb64-4b08-8fc9-6a11810e0231 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] [instance: 4b34ff57-e137-4abc-8aed-76dd2ec8d313] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 512.281844] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e1497b4-4e07-4e46-b1e0-190201e704a2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 512.286690] env[62066]: DEBUG oslo_concurrency.lockutils [None req-65bf05da-7d56-4074-8d13-63fde71b9619 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] Acquiring lock "eacdd472-5bb0-48bc-8603-9e1d08b41f55" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 512.287228] env[62066]: DEBUG oslo_concurrency.lockutils [None req-65bf05da-7d56-4074-8d13-63fde71b9619 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] Lock "eacdd472-5bb0-48bc-8603-9e1d08b41f55" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 512.335644] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0e23bf5c-5716-4eba-9510-a3597f398645 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.373s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 512.336193] env[62066]: DEBUG nova.compute.manager [None req-0e23bf5c-5716-4eba-9510-a3597f398645 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 512.340166] env[62066]: DEBUG oslo_concurrency.lockutils [None req-68481586-e4e9-4c46-8b51-aa7845aca824 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.676s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 512.341816] env[62066]: INFO nova.compute.claims [None req-68481586-e4e9-4c46-8b51-aa7845aca824 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 512.667992] env[62066]: DEBUG oslo_vmware.api [None req-1ca3716a-0388-4dd3-97b5-9e32fd241705 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Task: {'id': task-1340616, 'name': ReconfigVM_Task, 'duration_secs': 0.570593} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 512.670570] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-1ca3716a-0388-4dd3-97b5-9e32fd241705 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] [instance: 56eee633-2c01-4df2-8d2e-58ad712942f6] Reconfigured VM instance instance-00000005 to attach disk [datastore2] 56eee633-2c01-4df2-8d2e-58ad712942f6/56eee633-2c01-4df2-8d2e-58ad712942f6.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 512.670570] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-efbf4f1e-46ff-4a08-8b7f-9ec22915a814 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 512.682138] env[62066]: DEBUG oslo_vmware.api [None req-1ca3716a-0388-4dd3-97b5-9e32fd241705 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Waiting for the task: (returnval){ [ 512.682138] env[62066]: value = "task-1340617" [ 512.682138] env[62066]: _type = "Task" [ 512.682138] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 512.694276] env[62066]: DEBUG oslo_vmware.api [None req-1ca3716a-0388-4dd3-97b5-9e32fd241705 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Task: {'id': task-1340617, 'name': Rename_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 512.809477] env[62066]: INFO nova.compute.manager [None req-69d29524-cb64-4b08-8fc9-6a11810e0231 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] [instance: 4b34ff57-e137-4abc-8aed-76dd2ec8d313] Took 17.61 seconds to build instance. [ 512.847724] env[62066]: DEBUG nova.compute.utils [None req-0e23bf5c-5716-4eba-9510-a3597f398645 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 512.856945] env[62066]: DEBUG nova.compute.manager [None req-0e23bf5c-5716-4eba-9510-a3597f398645 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 512.856945] env[62066]: DEBUG nova.network.neutron [None req-0e23bf5c-5716-4eba-9510-a3597f398645 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 513.105492] env[62066]: DEBUG nova.policy [None req-0e23bf5c-5716-4eba-9510-a3597f398645 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fef942e00fa5418ca11215a109e5ca59', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c358d8ed80254a148ce1d26749e64951', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 513.192452] env[62066]: DEBUG oslo_vmware.api [None req-1ca3716a-0388-4dd3-97b5-9e32fd241705 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Task: {'id': task-1340617, 'name': Rename_Task, 'duration_secs': 0.293191} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 513.192895] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-1ca3716a-0388-4dd3-97b5-9e32fd241705 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] [instance: 56eee633-2c01-4df2-8d2e-58ad712942f6] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 513.193243] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-18989fd2-5ac1-46b3-84b2-e49d9325fd2d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 513.202227] env[62066]: DEBUG oslo_vmware.api [None req-1ca3716a-0388-4dd3-97b5-9e32fd241705 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Waiting for the task: (returnval){ [ 513.202227] env[62066]: value = "task-1340618" [ 513.202227] env[62066]: _type = "Task" [ 513.202227] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 513.214574] env[62066]: DEBUG oslo_vmware.api [None req-1ca3716a-0388-4dd3-97b5-9e32fd241705 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Task: {'id': task-1340618, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 513.237588] env[62066]: ERROR nova.compute.manager [None req-5c743a8d-c557-4d89-8fe0-1cc65e783f93 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 33397b47-1617-4402-98e4-8e41bebcc4a5, please check neutron logs for more information. [ 513.237588] env[62066]: ERROR nova.compute.manager Traceback (most recent call last): [ 513.237588] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 513.237588] env[62066]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 513.237588] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 513.237588] env[62066]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 513.237588] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 513.237588] env[62066]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 513.237588] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 513.237588] env[62066]: ERROR nova.compute.manager self.force_reraise() [ 513.237588] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 513.237588] env[62066]: ERROR nova.compute.manager raise self.value [ 513.237588] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 513.237588] env[62066]: ERROR nova.compute.manager updated_port = self._update_port( [ 513.237588] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 513.237588] env[62066]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 513.238127] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 513.238127] env[62066]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 513.238127] env[62066]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 33397b47-1617-4402-98e4-8e41bebcc4a5, please check neutron logs for more information. [ 513.238127] env[62066]: ERROR nova.compute.manager [ 513.238127] env[62066]: Traceback (most recent call last): [ 513.238127] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 513.238127] env[62066]: listener.cb(fileno) [ 513.238127] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 513.238127] env[62066]: result = function(*args, **kwargs) [ 513.238127] env[62066]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 513.238127] env[62066]: return func(*args, **kwargs) [ 513.238127] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 513.238127] env[62066]: raise e [ 513.238127] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 513.238127] env[62066]: nwinfo = self.network_api.allocate_for_instance( [ 513.238127] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 513.238127] env[62066]: created_port_ids = self._update_ports_for_instance( [ 513.238127] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 513.238127] env[62066]: with excutils.save_and_reraise_exception(): [ 513.238127] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 513.238127] env[62066]: self.force_reraise() [ 513.238127] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 513.238127] env[62066]: raise self.value [ 513.238127] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 513.238127] env[62066]: updated_port = self._update_port( [ 513.238127] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 513.238127] env[62066]: _ensure_no_port_binding_failure(port) [ 513.238127] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 513.238127] env[62066]: raise exception.PortBindingFailed(port_id=port['id']) [ 513.238921] env[62066]: nova.exception.PortBindingFailed: Binding failed for port 33397b47-1617-4402-98e4-8e41bebcc4a5, please check neutron logs for more information. [ 513.238921] env[62066]: Removing descriptor: 15 [ 513.238921] env[62066]: ERROR nova.compute.manager [None req-5c743a8d-c557-4d89-8fe0-1cc65e783f93 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 33397b47-1617-4402-98e4-8e41bebcc4a5, please check neutron logs for more information. [ 513.238921] env[62066]: ERROR nova.compute.manager [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] Traceback (most recent call last): [ 513.238921] env[62066]: ERROR nova.compute.manager [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 513.238921] env[62066]: ERROR nova.compute.manager [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] yield resources [ 513.238921] env[62066]: ERROR nova.compute.manager [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 513.238921] env[62066]: ERROR nova.compute.manager [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] self.driver.spawn(context, instance, image_meta, [ 513.238921] env[62066]: ERROR nova.compute.manager [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 513.238921] env[62066]: ERROR nova.compute.manager [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] self._vmops.spawn(context, instance, image_meta, injected_files, [ 513.238921] env[62066]: ERROR nova.compute.manager [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 513.238921] env[62066]: ERROR nova.compute.manager [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] vm_ref = self.build_virtual_machine(instance, [ 513.239282] env[62066]: ERROR nova.compute.manager [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 513.239282] env[62066]: ERROR nova.compute.manager [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] vif_infos = vmwarevif.get_vif_info(self._session, [ 513.239282] env[62066]: ERROR nova.compute.manager [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 513.239282] env[62066]: ERROR nova.compute.manager [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] for vif in network_info: [ 513.239282] env[62066]: ERROR nova.compute.manager [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 513.239282] env[62066]: ERROR nova.compute.manager [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] return self._sync_wrapper(fn, *args, **kwargs) [ 513.239282] env[62066]: ERROR nova.compute.manager [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 513.239282] env[62066]: ERROR nova.compute.manager [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] self.wait() [ 513.239282] env[62066]: ERROR nova.compute.manager [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 513.239282] env[62066]: ERROR nova.compute.manager [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] self[:] = self._gt.wait() [ 513.239282] env[62066]: ERROR nova.compute.manager [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 513.239282] env[62066]: ERROR nova.compute.manager [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] return self._exit_event.wait() [ 513.239282] env[62066]: ERROR nova.compute.manager [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 513.239625] env[62066]: ERROR nova.compute.manager [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] result = hub.switch() [ 513.239625] env[62066]: ERROR nova.compute.manager [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 513.239625] env[62066]: ERROR nova.compute.manager [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] return self.greenlet.switch() [ 513.239625] env[62066]: ERROR nova.compute.manager [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 513.239625] env[62066]: ERROR nova.compute.manager [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] result = function(*args, **kwargs) [ 513.239625] env[62066]: ERROR nova.compute.manager [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 513.239625] env[62066]: ERROR nova.compute.manager [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] return func(*args, **kwargs) [ 513.239625] env[62066]: ERROR nova.compute.manager [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 513.239625] env[62066]: ERROR nova.compute.manager [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] raise e [ 513.239625] env[62066]: ERROR nova.compute.manager [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 513.239625] env[62066]: ERROR nova.compute.manager [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] nwinfo = self.network_api.allocate_for_instance( [ 513.239625] env[62066]: ERROR nova.compute.manager [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 513.239625] env[62066]: ERROR nova.compute.manager [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] created_port_ids = self._update_ports_for_instance( [ 513.239979] env[62066]: ERROR nova.compute.manager [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 513.239979] env[62066]: ERROR nova.compute.manager [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] with excutils.save_and_reraise_exception(): [ 513.239979] env[62066]: ERROR nova.compute.manager [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 513.239979] env[62066]: ERROR nova.compute.manager [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] self.force_reraise() [ 513.239979] env[62066]: ERROR nova.compute.manager [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 513.239979] env[62066]: ERROR nova.compute.manager [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] raise self.value [ 513.239979] env[62066]: ERROR nova.compute.manager [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 513.239979] env[62066]: ERROR nova.compute.manager [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] updated_port = self._update_port( [ 513.239979] env[62066]: ERROR nova.compute.manager [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 513.239979] env[62066]: ERROR nova.compute.manager [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] _ensure_no_port_binding_failure(port) [ 513.239979] env[62066]: ERROR nova.compute.manager [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 513.239979] env[62066]: ERROR nova.compute.manager [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] raise exception.PortBindingFailed(port_id=port['id']) [ 513.240333] env[62066]: ERROR nova.compute.manager [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] nova.exception.PortBindingFailed: Binding failed for port 33397b47-1617-4402-98e4-8e41bebcc4a5, please check neutron logs for more information. [ 513.240333] env[62066]: ERROR nova.compute.manager [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] [ 513.240333] env[62066]: INFO nova.compute.manager [None req-5c743a8d-c557-4d89-8fe0-1cc65e783f93 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] Terminating instance [ 513.244679] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c743a8d-c557-4d89-8fe0-1cc65e783f93 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] Acquiring lock "refresh_cache-e8693f3d-ce6e-4c94-a719-4e6723c16de8" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 513.244864] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c743a8d-c557-4d89-8fe0-1cc65e783f93 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] Acquired lock "refresh_cache-e8693f3d-ce6e-4c94-a719-4e6723c16de8" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 513.245121] env[62066]: DEBUG nova.network.neutron [None req-5c743a8d-c557-4d89-8fe0-1cc65e783f93 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 513.311385] env[62066]: DEBUG oslo_concurrency.lockutils [None req-69d29524-cb64-4b08-8fc9-6a11810e0231 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] Lock "4b34ff57-e137-4abc-8aed-76dd2ec8d313" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.136s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 513.354598] env[62066]: DEBUG nova.compute.manager [None req-0e23bf5c-5716-4eba-9510-a3597f398645 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 513.693585] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6eb5264-efd3-49ad-acdd-0fa749bd394b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 513.709324] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c00856cb-df4d-4e83-aa8a-264c66a7f399 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 513.718733] env[62066]: DEBUG oslo_vmware.api [None req-1ca3716a-0388-4dd3-97b5-9e32fd241705 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Task: {'id': task-1340618, 'name': PowerOnVM_Task} progress is 64%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 513.745925] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f774288-c0cc-4ac3-99d7-f3b54652c896 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 513.756635] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c54f8344-04e7-48eb-a5cb-d7453982b914 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 513.772141] env[62066]: DEBUG nova.compute.provider_tree [None req-68481586-e4e9-4c46-8b51-aa7845aca824 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 513.820735] env[62066]: DEBUG nova.compute.manager [None req-9838e0c6-e0b2-45d5-9c5b-a0a0b9c60219 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] [instance: 64f9698e-16f5-447a-914a-9e936e05d101] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 513.855979] env[62066]: DEBUG nova.network.neutron [None req-5c743a8d-c557-4d89-8fe0-1cc65e783f93 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 513.983160] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4ae0c9b4-adca-4b03-8ac7-409a4ee035f6 tempest-InstanceActionsTestJSON-1008103477 tempest-InstanceActionsTestJSON-1008103477-project-member] Acquiring lock "454a790d-3cb2-4ef2-995f-97e0eb1e2469" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 513.983344] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4ae0c9b4-adca-4b03-8ac7-409a4ee035f6 tempest-InstanceActionsTestJSON-1008103477 tempest-InstanceActionsTestJSON-1008103477-project-member] Lock "454a790d-3cb2-4ef2-995f-97e0eb1e2469" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 514.214304] env[62066]: DEBUG oslo_vmware.api [None req-1ca3716a-0388-4dd3-97b5-9e32fd241705 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Task: {'id': task-1340618, 'name': PowerOnVM_Task, 'duration_secs': 0.798057} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 514.214642] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-1ca3716a-0388-4dd3-97b5-9e32fd241705 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] [instance: 56eee633-2c01-4df2-8d2e-58ad712942f6] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 514.215192] env[62066]: INFO nova.compute.manager [None req-1ca3716a-0388-4dd3-97b5-9e32fd241705 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] [instance: 56eee633-2c01-4df2-8d2e-58ad712942f6] Took 6.94 seconds to spawn the instance on the hypervisor. [ 514.215361] env[62066]: DEBUG nova.compute.manager [None req-1ca3716a-0388-4dd3-97b5-9e32fd241705 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] [instance: 56eee633-2c01-4df2-8d2e-58ad712942f6] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 514.216416] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d75e1776-a863-40ec-85a2-5946bd772453 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 514.275490] env[62066]: DEBUG nova.scheduler.client.report [None req-68481586-e4e9-4c46-8b51-aa7845aca824 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 514.308409] env[62066]: DEBUG nova.network.neutron [None req-5c743a8d-c557-4d89-8fe0-1cc65e783f93 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 514.348617] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9838e0c6-e0b2-45d5-9c5b-a0a0b9c60219 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 514.364994] env[62066]: DEBUG nova.compute.manager [None req-0e23bf5c-5716-4eba-9510-a3597f398645 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 514.397186] env[62066]: DEBUG nova.virt.hardware [None req-0e23bf5c-5716-4eba-9510-a3597f398645 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-23T13:40:41Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-23T13:40:23Z,direct_url=,disk_format='vmdk',id=50ff584c-3b50-4395-af07-3e66769bc9f7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='eb52f7069a374c61ae946f052007c6d9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-23T13:40:24Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 514.397186] env[62066]: DEBUG nova.virt.hardware [None req-0e23bf5c-5716-4eba-9510-a3597f398645 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 514.397186] env[62066]: DEBUG nova.virt.hardware [None req-0e23bf5c-5716-4eba-9510-a3597f398645 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 514.397186] env[62066]: DEBUG nova.virt.hardware [None req-0e23bf5c-5716-4eba-9510-a3597f398645 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 514.397424] env[62066]: DEBUG nova.virt.hardware [None req-0e23bf5c-5716-4eba-9510-a3597f398645 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 514.397424] env[62066]: DEBUG nova.virt.hardware [None req-0e23bf5c-5716-4eba-9510-a3597f398645 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 514.397424] env[62066]: DEBUG nova.virt.hardware [None req-0e23bf5c-5716-4eba-9510-a3597f398645 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 514.397595] env[62066]: DEBUG nova.virt.hardware [None req-0e23bf5c-5716-4eba-9510-a3597f398645 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 514.397890] env[62066]: DEBUG nova.virt.hardware [None req-0e23bf5c-5716-4eba-9510-a3597f398645 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 514.397965] env[62066]: DEBUG nova.virt.hardware [None req-0e23bf5c-5716-4eba-9510-a3597f398645 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 514.398219] env[62066]: DEBUG nova.virt.hardware [None req-0e23bf5c-5716-4eba-9510-a3597f398645 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 514.399181] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ccdf7fd-b273-4cbd-95b6-2605224d96ee {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 514.411195] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8db36f5a-f343-4356-bdaa-00d2ad94484c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 514.512676] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cba923da-ec24-4fd6-9695-23f810adaa32 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] Acquiring lock "0a08bc46-0eea-4802-b0be-eb24f3507f73" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 514.512926] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cba923da-ec24-4fd6-9695-23f810adaa32 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] Lock "0a08bc46-0eea-4802-b0be-eb24f3507f73" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 514.742842] env[62066]: INFO nova.compute.manager [None req-1ca3716a-0388-4dd3-97b5-9e32fd241705 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] [instance: 56eee633-2c01-4df2-8d2e-58ad712942f6] Took 18.75 seconds to build instance. [ 514.781272] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4fa8ac7b-ee00-4ef3-97aa-07a2c2ae6378 tempest-ServerExternalEventsTest-41922728 tempest-ServerExternalEventsTest-41922728-project-member] Acquiring lock "39cd204a-49c2-4309-9030-555056ab125e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 514.781272] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4fa8ac7b-ee00-4ef3-97aa-07a2c2ae6378 tempest-ServerExternalEventsTest-41922728 tempest-ServerExternalEventsTest-41922728-project-member] Lock "39cd204a-49c2-4309-9030-555056ab125e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 514.785312] env[62066]: DEBUG oslo_concurrency.lockutils [None req-68481586-e4e9-4c46-8b51-aa7845aca824 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.445s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 514.785312] env[62066]: DEBUG nova.compute.manager [None req-68481586-e4e9-4c46-8b51-aa7845aca824 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 514.789423] env[62066]: DEBUG oslo_concurrency.lockutils [None req-99d923d5-ef6f-43f2-959c-4b252ac1d398 tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.991s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 514.790914] env[62066]: INFO nova.compute.claims [None req-99d923d5-ef6f-43f2-959c-4b252ac1d398 tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 514.812220] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c743a8d-c557-4d89-8fe0-1cc65e783f93 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] Releasing lock "refresh_cache-e8693f3d-ce6e-4c94-a719-4e6723c16de8" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 514.812645] env[62066]: DEBUG nova.compute.manager [None req-5c743a8d-c557-4d89-8fe0-1cc65e783f93 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 514.812841] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-5c743a8d-c557-4d89-8fe0-1cc65e783f93 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 514.813905] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ca3c55cf-66fd-4275-90d8-654e723260e5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 514.826343] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28fe7946-987e-4b5e-9154-60ebea652930 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 514.856937] env[62066]: WARNING nova.virt.vmwareapi.vmops [None req-5c743a8d-c557-4d89-8fe0-1cc65e783f93 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance e8693f3d-ce6e-4c94-a719-4e6723c16de8 could not be found. [ 514.857363] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-5c743a8d-c557-4d89-8fe0-1cc65e783f93 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 514.857466] env[62066]: INFO nova.compute.manager [None req-5c743a8d-c557-4d89-8fe0-1cc65e783f93 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] Took 0.04 seconds to destroy the instance on the hypervisor. [ 514.857689] env[62066]: DEBUG oslo.service.loopingcall [None req-5c743a8d-c557-4d89-8fe0-1cc65e783f93 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 514.858613] env[62066]: DEBUG nova.compute.manager [-] [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 514.858700] env[62066]: DEBUG nova.network.neutron [-] [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 514.931858] env[62066]: DEBUG nova.network.neutron [-] [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 515.062510] env[62066]: DEBUG nova.network.neutron [None req-0e23bf5c-5716-4eba-9510-a3597f398645 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] Successfully created port: 78745978-2782-4cbd-8fbd-fbdf247cc830 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 515.122881] env[62066]: DEBUG nova.compute.manager [req-0b5f3160-3b0a-41b4-9fd3-b617aeca31ed req-77875ba9-1a40-4bcc-b7e8-c7375a3f712c service nova] [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] Received event network-changed-33397b47-1617-4402-98e4-8e41bebcc4a5 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 515.123093] env[62066]: DEBUG nova.compute.manager [req-0b5f3160-3b0a-41b4-9fd3-b617aeca31ed req-77875ba9-1a40-4bcc-b7e8-c7375a3f712c service nova] [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] Refreshing instance network info cache due to event network-changed-33397b47-1617-4402-98e4-8e41bebcc4a5. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 515.123308] env[62066]: DEBUG oslo_concurrency.lockutils [req-0b5f3160-3b0a-41b4-9fd3-b617aeca31ed req-77875ba9-1a40-4bcc-b7e8-c7375a3f712c service nova] Acquiring lock "refresh_cache-e8693f3d-ce6e-4c94-a719-4e6723c16de8" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 515.123451] env[62066]: DEBUG oslo_concurrency.lockutils [req-0b5f3160-3b0a-41b4-9fd3-b617aeca31ed req-77875ba9-1a40-4bcc-b7e8-c7375a3f712c service nova] Acquired lock "refresh_cache-e8693f3d-ce6e-4c94-a719-4e6723c16de8" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 515.123759] env[62066]: DEBUG nova.network.neutron [req-0b5f3160-3b0a-41b4-9fd3-b617aeca31ed req-77875ba9-1a40-4bcc-b7e8-c7375a3f712c service nova] [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] Refreshing network info cache for port 33397b47-1617-4402-98e4-8e41bebcc4a5 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 515.245719] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1ca3716a-0388-4dd3-97b5-9e32fd241705 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Lock "56eee633-2c01-4df2-8d2e-58ad712942f6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.262s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 515.262297] env[62066]: DEBUG nova.compute.manager [None req-b73516e3-0c66-48ce-8d0f-63023a332685 tempest-ServerDiagnosticsV248Test-1738630200 tempest-ServerDiagnosticsV248Test-1738630200-project-admin] [instance: 4b34ff57-e137-4abc-8aed-76dd2ec8d313] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 515.262297] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d8c9b98-4978-4e01-b410-121a05d67865 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 515.272019] env[62066]: INFO nova.compute.manager [None req-b73516e3-0c66-48ce-8d0f-63023a332685 tempest-ServerDiagnosticsV248Test-1738630200 tempest-ServerDiagnosticsV248Test-1738630200-project-admin] [instance: 4b34ff57-e137-4abc-8aed-76dd2ec8d313] Retrieving diagnostics [ 515.272019] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b2d03c2-debb-49dc-b636-2c41b6a0a122 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 515.313437] env[62066]: DEBUG nova.compute.utils [None req-68481586-e4e9-4c46-8b51-aa7845aca824 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 515.318897] env[62066]: DEBUG nova.compute.manager [None req-68481586-e4e9-4c46-8b51-aa7845aca824 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 515.318897] env[62066]: DEBUG nova.network.neutron [None req-68481586-e4e9-4c46-8b51-aa7845aca824 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 515.434636] env[62066]: DEBUG nova.network.neutron [-] [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 515.508473] env[62066]: DEBUG nova.policy [None req-68481586-e4e9-4c46-8b51-aa7845aca824 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '22afd2db22164c9da7ecbc76c58c68dc', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '975c88ff0be74b81a20da760ffc8e667', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 515.682996] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1e3e30d6-8bb5-4347-b6c2-c2d648c37b67 tempest-ImagesOneServerNegativeTestJSON-310236551 tempest-ImagesOneServerNegativeTestJSON-310236551-project-member] Acquiring lock "194588b4-fe40-4286-8036-874a7c410327" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 515.683337] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1e3e30d6-8bb5-4347-b6c2-c2d648c37b67 tempest-ImagesOneServerNegativeTestJSON-310236551 tempest-ImagesOneServerNegativeTestJSON-310236551-project-member] Lock "194588b4-fe40-4286-8036-874a7c410327" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 515.712485] env[62066]: DEBUG nova.network.neutron [req-0b5f3160-3b0a-41b4-9fd3-b617aeca31ed req-77875ba9-1a40-4bcc-b7e8-c7375a3f712c service nova] [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 515.748129] env[62066]: DEBUG nova.compute.manager [None req-67ee84a7-af8a-43fc-bff6-48212ddfc8ae tempest-ServersNegativeTestMultiTenantJSON-1325431503 tempest-ServersNegativeTestMultiTenantJSON-1325431503-project-member] [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 515.824812] env[62066]: DEBUG nova.compute.manager [None req-68481586-e4e9-4c46-8b51-aa7845aca824 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 515.938367] env[62066]: INFO nova.compute.manager [-] [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] Took 1.08 seconds to deallocate network for instance. [ 515.945355] env[62066]: DEBUG nova.compute.claims [None req-5c743a8d-c557-4d89-8fe0-1cc65e783f93 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] Aborting claim: {{(pid=62066) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 515.945657] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c743a8d-c557-4d89-8fe0-1cc65e783f93 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 516.270217] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a176157e-98e6-4923-8543-a1c2fa207653 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 516.272995] env[62066]: DEBUG oslo_concurrency.lockutils [None req-67ee84a7-af8a-43fc-bff6-48212ddfc8ae tempest-ServersNegativeTestMultiTenantJSON-1325431503 tempest-ServersNegativeTestMultiTenantJSON-1325431503-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 516.279460] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6957f114-ca8e-40a7-b707-aa39b014d40b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 516.314296] env[62066]: DEBUG nova.network.neutron [req-0b5f3160-3b0a-41b4-9fd3-b617aeca31ed req-77875ba9-1a40-4bcc-b7e8-c7375a3f712c service nova] [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 516.316015] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3db639df-a202-4145-a19d-1dc8dd0b6c68 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 516.325022] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40c1556f-4ab4-4a32-865f-bc8380a631b3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 516.347687] env[62066]: DEBUG nova.compute.provider_tree [None req-99d923d5-ef6f-43f2-959c-4b252ac1d398 tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 516.822123] env[62066]: DEBUG oslo_concurrency.lockutils [req-0b5f3160-3b0a-41b4-9fd3-b617aeca31ed req-77875ba9-1a40-4bcc-b7e8-c7375a3f712c service nova] Releasing lock "refresh_cache-e8693f3d-ce6e-4c94-a719-4e6723c16de8" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 516.838656] env[62066]: DEBUG nova.compute.manager [None req-68481586-e4e9-4c46-8b51-aa7845aca824 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 516.857720] env[62066]: DEBUG nova.scheduler.client.report [None req-99d923d5-ef6f-43f2-959c-4b252ac1d398 tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 516.872432] env[62066]: DEBUG nova.virt.hardware [None req-68481586-e4e9-4c46-8b51-aa7845aca824 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-23T13:40:41Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-23T13:40:23Z,direct_url=,disk_format='vmdk',id=50ff584c-3b50-4395-af07-3e66769bc9f7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='eb52f7069a374c61ae946f052007c6d9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-23T13:40:24Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 516.873055] env[62066]: DEBUG nova.virt.hardware [None req-68481586-e4e9-4c46-8b51-aa7845aca824 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 516.873788] env[62066]: DEBUG nova.virt.hardware [None req-68481586-e4e9-4c46-8b51-aa7845aca824 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 516.873788] env[62066]: DEBUG nova.virt.hardware [None req-68481586-e4e9-4c46-8b51-aa7845aca824 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 516.873788] env[62066]: DEBUG nova.virt.hardware [None req-68481586-e4e9-4c46-8b51-aa7845aca824 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 516.874224] env[62066]: DEBUG nova.virt.hardware [None req-68481586-e4e9-4c46-8b51-aa7845aca824 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 516.874624] env[62066]: DEBUG nova.virt.hardware [None req-68481586-e4e9-4c46-8b51-aa7845aca824 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 516.876393] env[62066]: DEBUG nova.virt.hardware [None req-68481586-e4e9-4c46-8b51-aa7845aca824 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 516.876393] env[62066]: DEBUG nova.virt.hardware [None req-68481586-e4e9-4c46-8b51-aa7845aca824 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 516.876393] env[62066]: DEBUG nova.virt.hardware [None req-68481586-e4e9-4c46-8b51-aa7845aca824 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 516.876393] env[62066]: DEBUG nova.virt.hardware [None req-68481586-e4e9-4c46-8b51-aa7845aca824 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 516.876580] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5adb9b6e-91e8-4973-bfeb-6dcce7e6041e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 516.886522] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e21b1638-7ed7-4027-9148-2fa7a5f627ca {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 517.363897] env[62066]: DEBUG oslo_concurrency.lockutils [None req-99d923d5-ef6f-43f2-959c-4b252ac1d398 tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.574s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 517.364698] env[62066]: DEBUG nova.compute.manager [None req-99d923d5-ef6f-43f2-959c-4b252ac1d398 tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 517.369481] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e2ac9da0-4290-4c63-a476-a96d29c6e38b tempest-ServerDiagnosticsNegativeTest-1631323339 tempest-ServerDiagnosticsNegativeTest-1631323339-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.816s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 517.713403] env[62066]: DEBUG nova.network.neutron [None req-68481586-e4e9-4c46-8b51-aa7845aca824 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] Successfully created port: 5d597d92-a2ed-414a-babb-5eacc2e2e5c2 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 517.871788] env[62066]: DEBUG nova.compute.utils [None req-99d923d5-ef6f-43f2-959c-4b252ac1d398 tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 517.877454] env[62066]: DEBUG nova.compute.manager [None req-99d923d5-ef6f-43f2-959c-4b252ac1d398 tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 517.877659] env[62066]: DEBUG nova.network.neutron [None req-99d923d5-ef6f-43f2-959c-4b252ac1d398 tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 517.967299] env[62066]: INFO nova.compute.manager [None req-70656e85-7c6a-408b-a415-8e3cff7746a0 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] [instance: 56eee633-2c01-4df2-8d2e-58ad712942f6] Rebuilding instance [ 518.064953] env[62066]: DEBUG nova.compute.manager [None req-70656e85-7c6a-408b-a415-8e3cff7746a0 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] [instance: 56eee633-2c01-4df2-8d2e-58ad712942f6] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 518.065856] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47185eac-9da4-4ae3-951c-4976ae5213b8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 518.130324] env[62066]: DEBUG nova.policy [None req-99d923d5-ef6f-43f2-959c-4b252ac1d398 tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a00e7a5187d949e9a8924e9fda469dff', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'dd07c5b9369f4803900c3974049ecf33', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 518.306586] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f915c247-9097-4fa0-bc14-dd112b1d94d9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 518.317581] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffdad7c4-13cd-43f7-b843-f41727155d52 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 518.353880] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6936f439-3c29-42c1-9092-995e65041b4b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 518.362056] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91161858-7d78-44d6-95c7-13c25963d794 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 518.376612] env[62066]: DEBUG nova.compute.provider_tree [None req-e2ac9da0-4290-4c63-a476-a96d29c6e38b tempest-ServerDiagnosticsNegativeTest-1631323339 tempest-ServerDiagnosticsNegativeTest-1631323339-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 518.378540] env[62066]: DEBUG nova.compute.manager [None req-99d923d5-ef6f-43f2-959c-4b252ac1d398 tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 518.428925] env[62066]: DEBUG nova.compute.manager [req-8007c6d4-3db2-407c-9207-99e0833b04ea req-4f9bafd5-3c6b-42c6-a9bd-9cad6506b3a3 service nova] [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] Received event network-vif-deleted-33397b47-1617-4402-98e4-8e41bebcc4a5 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 518.428925] env[62066]: DEBUG nova.compute.manager [req-8007c6d4-3db2-407c-9207-99e0833b04ea req-4f9bafd5-3c6b-42c6-a9bd-9cad6506b3a3 service nova] [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] Received event network-changed-e279d463-9a92-4d4b-913e-d08e912e8b3e {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 518.428925] env[62066]: DEBUG nova.compute.manager [req-8007c6d4-3db2-407c-9207-99e0833b04ea req-4f9bafd5-3c6b-42c6-a9bd-9cad6506b3a3 service nova] [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] Refreshing instance network info cache due to event network-changed-e279d463-9a92-4d4b-913e-d08e912e8b3e. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 518.429619] env[62066]: DEBUG oslo_concurrency.lockutils [req-8007c6d4-3db2-407c-9207-99e0833b04ea req-4f9bafd5-3c6b-42c6-a9bd-9cad6506b3a3 service nova] Acquiring lock "refresh_cache-1eeb879b-6d1e-4746-98b5-12c30d674fa9" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 518.430186] env[62066]: DEBUG oslo_concurrency.lockutils [req-8007c6d4-3db2-407c-9207-99e0833b04ea req-4f9bafd5-3c6b-42c6-a9bd-9cad6506b3a3 service nova] Acquired lock "refresh_cache-1eeb879b-6d1e-4746-98b5-12c30d674fa9" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 518.430574] env[62066]: DEBUG nova.network.neutron [req-8007c6d4-3db2-407c-9207-99e0833b04ea req-4f9bafd5-3c6b-42c6-a9bd-9cad6506b3a3 service nova] [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] Refreshing network info cache for port e279d463-9a92-4d4b-913e-d08e912e8b3e {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 518.582525] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-70656e85-7c6a-408b-a415-8e3cff7746a0 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] [instance: 56eee633-2c01-4df2-8d2e-58ad712942f6] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 518.583896] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e9160a40-34d2-4db3-a500-5d58c66d1724 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 518.591718] env[62066]: DEBUG oslo_vmware.api [None req-70656e85-7c6a-408b-a415-8e3cff7746a0 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Waiting for the task: (returnval){ [ 518.591718] env[62066]: value = "task-1340619" [ 518.591718] env[62066]: _type = "Task" [ 518.591718] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 518.609517] env[62066]: DEBUG oslo_vmware.api [None req-70656e85-7c6a-408b-a415-8e3cff7746a0 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Task: {'id': task-1340619, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 518.646850] env[62066]: ERROR nova.compute.manager [None req-df85dbc4-0217-427e-9c09-82d1a0e6d774 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port e279d463-9a92-4d4b-913e-d08e912e8b3e, please check neutron logs for more information. [ 518.646850] env[62066]: ERROR nova.compute.manager Traceback (most recent call last): [ 518.646850] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 518.646850] env[62066]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 518.646850] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 518.646850] env[62066]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 518.646850] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 518.646850] env[62066]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 518.646850] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 518.646850] env[62066]: ERROR nova.compute.manager self.force_reraise() [ 518.646850] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 518.646850] env[62066]: ERROR nova.compute.manager raise self.value [ 518.646850] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 518.646850] env[62066]: ERROR nova.compute.manager updated_port = self._update_port( [ 518.646850] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 518.646850] env[62066]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 518.647322] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 518.647322] env[62066]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 518.647322] env[62066]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port e279d463-9a92-4d4b-913e-d08e912e8b3e, please check neutron logs for more information. [ 518.647322] env[62066]: ERROR nova.compute.manager [ 518.647322] env[62066]: Traceback (most recent call last): [ 518.647322] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 518.647322] env[62066]: listener.cb(fileno) [ 518.647322] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 518.647322] env[62066]: result = function(*args, **kwargs) [ 518.647322] env[62066]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 518.647322] env[62066]: return func(*args, **kwargs) [ 518.647322] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 518.647322] env[62066]: raise e [ 518.647322] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 518.647322] env[62066]: nwinfo = self.network_api.allocate_for_instance( [ 518.647322] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 518.647322] env[62066]: created_port_ids = self._update_ports_for_instance( [ 518.647322] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 518.647322] env[62066]: with excutils.save_and_reraise_exception(): [ 518.647322] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 518.647322] env[62066]: self.force_reraise() [ 518.647322] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 518.647322] env[62066]: raise self.value [ 518.647322] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 518.647322] env[62066]: updated_port = self._update_port( [ 518.647322] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 518.647322] env[62066]: _ensure_no_port_binding_failure(port) [ 518.647322] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 518.647322] env[62066]: raise exception.PortBindingFailed(port_id=port['id']) [ 518.648273] env[62066]: nova.exception.PortBindingFailed: Binding failed for port e279d463-9a92-4d4b-913e-d08e912e8b3e, please check neutron logs for more information. [ 518.648273] env[62066]: Removing descriptor: 16 [ 518.648273] env[62066]: ERROR nova.compute.manager [None req-df85dbc4-0217-427e-9c09-82d1a0e6d774 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port e279d463-9a92-4d4b-913e-d08e912e8b3e, please check neutron logs for more information. [ 518.648273] env[62066]: ERROR nova.compute.manager [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] Traceback (most recent call last): [ 518.648273] env[62066]: ERROR nova.compute.manager [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 518.648273] env[62066]: ERROR nova.compute.manager [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] yield resources [ 518.648273] env[62066]: ERROR nova.compute.manager [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 518.648273] env[62066]: ERROR nova.compute.manager [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] self.driver.spawn(context, instance, image_meta, [ 518.648273] env[62066]: ERROR nova.compute.manager [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 518.648273] env[62066]: ERROR nova.compute.manager [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 518.648273] env[62066]: ERROR nova.compute.manager [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 518.648273] env[62066]: ERROR nova.compute.manager [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] vm_ref = self.build_virtual_machine(instance, [ 518.648659] env[62066]: ERROR nova.compute.manager [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 518.648659] env[62066]: ERROR nova.compute.manager [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] vif_infos = vmwarevif.get_vif_info(self._session, [ 518.648659] env[62066]: ERROR nova.compute.manager [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 518.648659] env[62066]: ERROR nova.compute.manager [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] for vif in network_info: [ 518.648659] env[62066]: ERROR nova.compute.manager [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 518.648659] env[62066]: ERROR nova.compute.manager [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] return self._sync_wrapper(fn, *args, **kwargs) [ 518.648659] env[62066]: ERROR nova.compute.manager [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 518.648659] env[62066]: ERROR nova.compute.manager [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] self.wait() [ 518.648659] env[62066]: ERROR nova.compute.manager [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 518.648659] env[62066]: ERROR nova.compute.manager [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] self[:] = self._gt.wait() [ 518.648659] env[62066]: ERROR nova.compute.manager [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 518.648659] env[62066]: ERROR nova.compute.manager [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] return self._exit_event.wait() [ 518.648659] env[62066]: ERROR nova.compute.manager [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 518.649023] env[62066]: ERROR nova.compute.manager [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] result = hub.switch() [ 518.649023] env[62066]: ERROR nova.compute.manager [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 518.649023] env[62066]: ERROR nova.compute.manager [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] return self.greenlet.switch() [ 518.649023] env[62066]: ERROR nova.compute.manager [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 518.649023] env[62066]: ERROR nova.compute.manager [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] result = function(*args, **kwargs) [ 518.649023] env[62066]: ERROR nova.compute.manager [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 518.649023] env[62066]: ERROR nova.compute.manager [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] return func(*args, **kwargs) [ 518.649023] env[62066]: ERROR nova.compute.manager [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 518.649023] env[62066]: ERROR nova.compute.manager [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] raise e [ 518.649023] env[62066]: ERROR nova.compute.manager [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 518.649023] env[62066]: ERROR nova.compute.manager [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] nwinfo = self.network_api.allocate_for_instance( [ 518.649023] env[62066]: ERROR nova.compute.manager [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 518.649023] env[62066]: ERROR nova.compute.manager [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] created_port_ids = self._update_ports_for_instance( [ 518.649348] env[62066]: ERROR nova.compute.manager [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 518.649348] env[62066]: ERROR nova.compute.manager [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] with excutils.save_and_reraise_exception(): [ 518.649348] env[62066]: ERROR nova.compute.manager [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 518.649348] env[62066]: ERROR nova.compute.manager [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] self.force_reraise() [ 518.649348] env[62066]: ERROR nova.compute.manager [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 518.649348] env[62066]: ERROR nova.compute.manager [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] raise self.value [ 518.649348] env[62066]: ERROR nova.compute.manager [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 518.649348] env[62066]: ERROR nova.compute.manager [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] updated_port = self._update_port( [ 518.649348] env[62066]: ERROR nova.compute.manager [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 518.649348] env[62066]: ERROR nova.compute.manager [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] _ensure_no_port_binding_failure(port) [ 518.649348] env[62066]: ERROR nova.compute.manager [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 518.649348] env[62066]: ERROR nova.compute.manager [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] raise exception.PortBindingFailed(port_id=port['id']) [ 518.649659] env[62066]: ERROR nova.compute.manager [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] nova.exception.PortBindingFailed: Binding failed for port e279d463-9a92-4d4b-913e-d08e912e8b3e, please check neutron logs for more information. [ 518.649659] env[62066]: ERROR nova.compute.manager [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] [ 518.649659] env[62066]: INFO nova.compute.manager [None req-df85dbc4-0217-427e-9c09-82d1a0e6d774 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] Terminating instance [ 518.656403] env[62066]: DEBUG oslo_concurrency.lockutils [None req-df85dbc4-0217-427e-9c09-82d1a0e6d774 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] Acquiring lock "refresh_cache-1eeb879b-6d1e-4746-98b5-12c30d674fa9" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 518.884018] env[62066]: DEBUG nova.scheduler.client.report [None req-e2ac9da0-4290-4c63-a476-a96d29c6e38b tempest-ServerDiagnosticsNegativeTest-1631323339 tempest-ServerDiagnosticsNegativeTest-1631323339-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 519.006865] env[62066]: DEBUG nova.network.neutron [req-8007c6d4-3db2-407c-9207-99e0833b04ea req-4f9bafd5-3c6b-42c6-a9bd-9cad6506b3a3 service nova] [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 519.105820] env[62066]: DEBUG oslo_vmware.api [None req-70656e85-7c6a-408b-a415-8e3cff7746a0 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Task: {'id': task-1340619, 'name': PowerOffVM_Task, 'duration_secs': 0.142794} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 519.106110] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-70656e85-7c6a-408b-a415-8e3cff7746a0 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] [instance: 56eee633-2c01-4df2-8d2e-58ad712942f6] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 519.106403] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-70656e85-7c6a-408b-a415-8e3cff7746a0 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] [instance: 56eee633-2c01-4df2-8d2e-58ad712942f6] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 519.108599] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7a97c05-c2c6-48a8-8f14-05f87c63f0d6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 519.120370] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-70656e85-7c6a-408b-a415-8e3cff7746a0 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] [instance: 56eee633-2c01-4df2-8d2e-58ad712942f6] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 519.120764] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-680c6b63-5f8d-49df-88a0-fa65e1940b13 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 519.151643] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-70656e85-7c6a-408b-a415-8e3cff7746a0 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] [instance: 56eee633-2c01-4df2-8d2e-58ad712942f6] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 519.152138] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-70656e85-7c6a-408b-a415-8e3cff7746a0 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] [instance: 56eee633-2c01-4df2-8d2e-58ad712942f6] Deleting contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 519.152138] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-70656e85-7c6a-408b-a415-8e3cff7746a0 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Deleting the datastore file [datastore2] 56eee633-2c01-4df2-8d2e-58ad712942f6 {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 519.152310] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6f68f2b0-7b7f-4587-829e-0eb096d1274c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 519.164165] env[62066]: DEBUG oslo_vmware.api [None req-70656e85-7c6a-408b-a415-8e3cff7746a0 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Waiting for the task: (returnval){ [ 519.164165] env[62066]: value = "task-1340621" [ 519.164165] env[62066]: _type = "Task" [ 519.164165] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 519.174604] env[62066]: DEBUG oslo_vmware.api [None req-70656e85-7c6a-408b-a415-8e3cff7746a0 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Task: {'id': task-1340621, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 519.394864] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e2ac9da0-4290-4c63-a476-a96d29c6e38b tempest-ServerDiagnosticsNegativeTest-1631323339 tempest-ServerDiagnosticsNegativeTest-1631323339-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.025s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 519.395891] env[62066]: ERROR nova.compute.manager [None req-e2ac9da0-4290-4c63-a476-a96d29c6e38b tempest-ServerDiagnosticsNegativeTest-1631323339 tempest-ServerDiagnosticsNegativeTest-1631323339-project-member] [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 91cab0bf-fb58-43b0-ab1a-a74661d807d0, please check neutron logs for more information. [ 519.395891] env[62066]: ERROR nova.compute.manager [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] Traceback (most recent call last): [ 519.395891] env[62066]: ERROR nova.compute.manager [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 519.395891] env[62066]: ERROR nova.compute.manager [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] self.driver.spawn(context, instance, image_meta, [ 519.395891] env[62066]: ERROR nova.compute.manager [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 519.395891] env[62066]: ERROR nova.compute.manager [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] self._vmops.spawn(context, instance, image_meta, injected_files, [ 519.395891] env[62066]: ERROR nova.compute.manager [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 519.395891] env[62066]: ERROR nova.compute.manager [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] vm_ref = self.build_virtual_machine(instance, [ 519.395891] env[62066]: ERROR nova.compute.manager [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 519.395891] env[62066]: ERROR nova.compute.manager [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] vif_infos = vmwarevif.get_vif_info(self._session, [ 519.395891] env[62066]: ERROR nova.compute.manager [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 519.396276] env[62066]: ERROR nova.compute.manager [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] for vif in network_info: [ 519.396276] env[62066]: ERROR nova.compute.manager [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 519.396276] env[62066]: ERROR nova.compute.manager [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] return self._sync_wrapper(fn, *args, **kwargs) [ 519.396276] env[62066]: ERROR nova.compute.manager [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 519.396276] env[62066]: ERROR nova.compute.manager [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] self.wait() [ 519.396276] env[62066]: ERROR nova.compute.manager [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 519.396276] env[62066]: ERROR nova.compute.manager [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] self[:] = self._gt.wait() [ 519.396276] env[62066]: ERROR nova.compute.manager [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 519.396276] env[62066]: ERROR nova.compute.manager [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] return self._exit_event.wait() [ 519.396276] env[62066]: ERROR nova.compute.manager [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 519.396276] env[62066]: ERROR nova.compute.manager [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] result = hub.switch() [ 519.396276] env[62066]: ERROR nova.compute.manager [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 519.396276] env[62066]: ERROR nova.compute.manager [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] return self.greenlet.switch() [ 519.396627] env[62066]: ERROR nova.compute.manager [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 519.396627] env[62066]: ERROR nova.compute.manager [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] result = function(*args, **kwargs) [ 519.396627] env[62066]: ERROR nova.compute.manager [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 519.396627] env[62066]: ERROR nova.compute.manager [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] return func(*args, **kwargs) [ 519.396627] env[62066]: ERROR nova.compute.manager [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 519.396627] env[62066]: ERROR nova.compute.manager [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] raise e [ 519.396627] env[62066]: ERROR nova.compute.manager [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 519.396627] env[62066]: ERROR nova.compute.manager [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] nwinfo = self.network_api.allocate_for_instance( [ 519.396627] env[62066]: ERROR nova.compute.manager [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 519.396627] env[62066]: ERROR nova.compute.manager [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] created_port_ids = self._update_ports_for_instance( [ 519.396627] env[62066]: ERROR nova.compute.manager [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 519.396627] env[62066]: ERROR nova.compute.manager [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] with excutils.save_and_reraise_exception(): [ 519.396627] env[62066]: ERROR nova.compute.manager [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 519.396961] env[62066]: ERROR nova.compute.manager [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] self.force_reraise() [ 519.396961] env[62066]: ERROR nova.compute.manager [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 519.396961] env[62066]: ERROR nova.compute.manager [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] raise self.value [ 519.396961] env[62066]: ERROR nova.compute.manager [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 519.396961] env[62066]: ERROR nova.compute.manager [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] updated_port = self._update_port( [ 519.396961] env[62066]: ERROR nova.compute.manager [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 519.396961] env[62066]: ERROR nova.compute.manager [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] _ensure_no_port_binding_failure(port) [ 519.396961] env[62066]: ERROR nova.compute.manager [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 519.396961] env[62066]: ERROR nova.compute.manager [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] raise exception.PortBindingFailed(port_id=port['id']) [ 519.396961] env[62066]: ERROR nova.compute.manager [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] nova.exception.PortBindingFailed: Binding failed for port 91cab0bf-fb58-43b0-ab1a-a74661d807d0, please check neutron logs for more information. [ 519.396961] env[62066]: ERROR nova.compute.manager [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] [ 519.398450] env[62066]: DEBUG nova.compute.utils [None req-e2ac9da0-4290-4c63-a476-a96d29c6e38b tempest-ServerDiagnosticsNegativeTest-1631323339 tempest-ServerDiagnosticsNegativeTest-1631323339-project-member] [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] Binding failed for port 91cab0bf-fb58-43b0-ab1a-a74661d807d0, please check neutron logs for more information. {{(pid=62066) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 519.400397] env[62066]: DEBUG nova.compute.manager [None req-99d923d5-ef6f-43f2-959c-4b252ac1d398 tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 519.403175] env[62066]: DEBUG oslo_concurrency.lockutils [None req-dbc2121a-ef71-4ef4-8f94-3a34690382a8 tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.317s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 519.409600] env[62066]: DEBUG nova.compute.manager [None req-e2ac9da0-4290-4c63-a476-a96d29c6e38b tempest-ServerDiagnosticsNegativeTest-1631323339 tempest-ServerDiagnosticsNegativeTest-1631323339-project-member] [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] Build of instance 8b29fc32-d27d-4474-84bc-f727f90fde09 was re-scheduled: Binding failed for port 91cab0bf-fb58-43b0-ab1a-a74661d807d0, please check neutron logs for more information. {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 519.412021] env[62066]: DEBUG nova.compute.manager [None req-e2ac9da0-4290-4c63-a476-a96d29c6e38b tempest-ServerDiagnosticsNegativeTest-1631323339 tempest-ServerDiagnosticsNegativeTest-1631323339-project-member] [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] Unplugging VIFs for instance {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 519.412021] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e2ac9da0-4290-4c63-a476-a96d29c6e38b tempest-ServerDiagnosticsNegativeTest-1631323339 tempest-ServerDiagnosticsNegativeTest-1631323339-project-member] Acquiring lock "refresh_cache-8b29fc32-d27d-4474-84bc-f727f90fde09" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 519.412021] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e2ac9da0-4290-4c63-a476-a96d29c6e38b tempest-ServerDiagnosticsNegativeTest-1631323339 tempest-ServerDiagnosticsNegativeTest-1631323339-project-member] Acquired lock "refresh_cache-8b29fc32-d27d-4474-84bc-f727f90fde09" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 519.412021] env[62066]: DEBUG nova.network.neutron [None req-e2ac9da0-4290-4c63-a476-a96d29c6e38b tempest-ServerDiagnosticsNegativeTest-1631323339 tempest-ServerDiagnosticsNegativeTest-1631323339-project-member] [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 519.445837] env[62066]: DEBUG nova.virt.hardware [None req-99d923d5-ef6f-43f2-959c-4b252ac1d398 tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-23T13:40:41Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-23T13:40:23Z,direct_url=,disk_format='vmdk',id=50ff584c-3b50-4395-af07-3e66769bc9f7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='eb52f7069a374c61ae946f052007c6d9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-23T13:40:24Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 519.447071] env[62066]: DEBUG nova.virt.hardware [None req-99d923d5-ef6f-43f2-959c-4b252ac1d398 tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 519.447682] env[62066]: DEBUG nova.virt.hardware [None req-99d923d5-ef6f-43f2-959c-4b252ac1d398 tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 519.447682] env[62066]: DEBUG nova.virt.hardware [None req-99d923d5-ef6f-43f2-959c-4b252ac1d398 tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 519.447682] env[62066]: DEBUG nova.virt.hardware [None req-99d923d5-ef6f-43f2-959c-4b252ac1d398 tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 519.447891] env[62066]: DEBUG nova.virt.hardware [None req-99d923d5-ef6f-43f2-959c-4b252ac1d398 tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 519.448775] env[62066]: DEBUG nova.virt.hardware [None req-99d923d5-ef6f-43f2-959c-4b252ac1d398 tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 519.448775] env[62066]: DEBUG nova.virt.hardware [None req-99d923d5-ef6f-43f2-959c-4b252ac1d398 tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 519.448775] env[62066]: DEBUG nova.virt.hardware [None req-99d923d5-ef6f-43f2-959c-4b252ac1d398 tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 519.448775] env[62066]: DEBUG nova.virt.hardware [None req-99d923d5-ef6f-43f2-959c-4b252ac1d398 tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 519.448775] env[62066]: DEBUG nova.virt.hardware [None req-99d923d5-ef6f-43f2-959c-4b252ac1d398 tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 519.450263] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94372c6d-b435-4c77-b924-3ec20c5035ca {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 519.460993] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3df5c968-3b89-40f3-89f4-a7ecbb0d0d42 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 519.526826] env[62066]: DEBUG nova.network.neutron [req-8007c6d4-3db2-407c-9207-99e0833b04ea req-4f9bafd5-3c6b-42c6-a9bd-9cad6506b3a3 service nova] [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 519.679935] env[62066]: DEBUG oslo_vmware.api [None req-70656e85-7c6a-408b-a415-8e3cff7746a0 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Task: {'id': task-1340621, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.22035} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 519.680112] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-70656e85-7c6a-408b-a415-8e3cff7746a0 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 519.680327] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-70656e85-7c6a-408b-a415-8e3cff7746a0 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] [instance: 56eee633-2c01-4df2-8d2e-58ad712942f6] Deleted contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 519.680447] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-70656e85-7c6a-408b-a415-8e3cff7746a0 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] [instance: 56eee633-2c01-4df2-8d2e-58ad712942f6] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 519.978258] env[62066]: DEBUG nova.network.neutron [None req-e2ac9da0-4290-4c63-a476-a96d29c6e38b tempest-ServerDiagnosticsNegativeTest-1631323339 tempest-ServerDiagnosticsNegativeTest-1631323339-project-member] [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 520.030324] env[62066]: DEBUG oslo_concurrency.lockutils [req-8007c6d4-3db2-407c-9207-99e0833b04ea req-4f9bafd5-3c6b-42c6-a9bd-9cad6506b3a3 service nova] Releasing lock "refresh_cache-1eeb879b-6d1e-4746-98b5-12c30d674fa9" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 520.031654] env[62066]: DEBUG oslo_concurrency.lockutils [None req-df85dbc4-0217-427e-9c09-82d1a0e6d774 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] Acquired lock "refresh_cache-1eeb879b-6d1e-4746-98b5-12c30d674fa9" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 520.031654] env[62066]: DEBUG nova.network.neutron [None req-df85dbc4-0217-427e-9c09-82d1a0e6d774 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 520.134236] env[62066]: DEBUG nova.network.neutron [None req-99d923d5-ef6f-43f2-959c-4b252ac1d398 tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] Successfully created port: 04cc7ab7-caaa-4afc-afd7-b4230a6e2326 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 520.284536] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9764d867-20f3-43f0-a859-c7ed3ddcb023 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 520.297045] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f584359-f87f-43ff-a8ef-8b5d0cb1c66f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 520.303037] env[62066]: DEBUG nova.network.neutron [None req-e2ac9da0-4290-4c63-a476-a96d29c6e38b tempest-ServerDiagnosticsNegativeTest-1631323339 tempest-ServerDiagnosticsNegativeTest-1631323339-project-member] [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 520.336703] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27437aa8-f25e-4085-ab47-93fe0acd921e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 520.345224] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58e6cd34-f42a-4be4-907d-ebbcb77d8f71 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 520.361970] env[62066]: DEBUG nova.compute.provider_tree [None req-dbc2121a-ef71-4ef4-8f94-3a34690382a8 tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 520.597328] env[62066]: DEBUG nova.network.neutron [None req-df85dbc4-0217-427e-9c09-82d1a0e6d774 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 520.730112] env[62066]: DEBUG nova.virt.hardware [None req-70656e85-7c6a-408b-a415-8e3cff7746a0 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-23T13:40:41Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-23T13:40:23Z,direct_url=,disk_format='vmdk',id=50ff584c-3b50-4395-af07-3e66769bc9f7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='eb52f7069a374c61ae946f052007c6d9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-23T13:40:24Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 520.730405] env[62066]: DEBUG nova.virt.hardware [None req-70656e85-7c6a-408b-a415-8e3cff7746a0 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 520.730782] env[62066]: DEBUG nova.virt.hardware [None req-70656e85-7c6a-408b-a415-8e3cff7746a0 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 520.730782] env[62066]: DEBUG nova.virt.hardware [None req-70656e85-7c6a-408b-a415-8e3cff7746a0 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 520.730866] env[62066]: DEBUG nova.virt.hardware [None req-70656e85-7c6a-408b-a415-8e3cff7746a0 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 520.731032] env[62066]: DEBUG nova.virt.hardware [None req-70656e85-7c6a-408b-a415-8e3cff7746a0 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 520.731280] env[62066]: DEBUG nova.virt.hardware [None req-70656e85-7c6a-408b-a415-8e3cff7746a0 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 520.731439] env[62066]: DEBUG nova.virt.hardware [None req-70656e85-7c6a-408b-a415-8e3cff7746a0 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 520.731605] env[62066]: DEBUG nova.virt.hardware [None req-70656e85-7c6a-408b-a415-8e3cff7746a0 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 520.731778] env[62066]: DEBUG nova.virt.hardware [None req-70656e85-7c6a-408b-a415-8e3cff7746a0 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 520.731936] env[62066]: DEBUG nova.virt.hardware [None req-70656e85-7c6a-408b-a415-8e3cff7746a0 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 520.732883] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3168e92b-4795-47ff-bc85-33d150aaa3f1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 520.743622] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9af048f0-4a5e-45b7-9435-595c1af0f973 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 520.757682] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-70656e85-7c6a-408b-a415-8e3cff7746a0 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] [instance: 56eee633-2c01-4df2-8d2e-58ad712942f6] Instance VIF info [] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 520.765093] env[62066]: DEBUG oslo.service.loopingcall [None req-70656e85-7c6a-408b-a415-8e3cff7746a0 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 520.767908] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 56eee633-2c01-4df2-8d2e-58ad712942f6] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 520.768096] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0481a4ee-4fac-4e10-b966-26f23c328d3d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 520.786961] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 520.786961] env[62066]: value = "task-1340622" [ 520.786961] env[62066]: _type = "Task" [ 520.786961] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 520.795610] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1340622, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 520.805626] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e2ac9da0-4290-4c63-a476-a96d29c6e38b tempest-ServerDiagnosticsNegativeTest-1631323339 tempest-ServerDiagnosticsNegativeTest-1631323339-project-member] Releasing lock "refresh_cache-8b29fc32-d27d-4474-84bc-f727f90fde09" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 520.805865] env[62066]: DEBUG nova.compute.manager [None req-e2ac9da0-4290-4c63-a476-a96d29c6e38b tempest-ServerDiagnosticsNegativeTest-1631323339 tempest-ServerDiagnosticsNegativeTest-1631323339-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 520.806173] env[62066]: DEBUG nova.compute.manager [None req-e2ac9da0-4290-4c63-a476-a96d29c6e38b tempest-ServerDiagnosticsNegativeTest-1631323339 tempest-ServerDiagnosticsNegativeTest-1631323339-project-member] [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 520.806253] env[62066]: DEBUG nova.network.neutron [None req-e2ac9da0-4290-4c63-a476-a96d29c6e38b tempest-ServerDiagnosticsNegativeTest-1631323339 tempest-ServerDiagnosticsNegativeTest-1631323339-project-member] [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 520.857711] env[62066]: DEBUG nova.network.neutron [None req-e2ac9da0-4290-4c63-a476-a96d29c6e38b tempest-ServerDiagnosticsNegativeTest-1631323339 tempest-ServerDiagnosticsNegativeTest-1631323339-project-member] [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 520.865409] env[62066]: DEBUG nova.scheduler.client.report [None req-dbc2121a-ef71-4ef4-8f94-3a34690382a8 tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 520.933081] env[62066]: DEBUG nova.network.neutron [None req-df85dbc4-0217-427e-9c09-82d1a0e6d774 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 521.298931] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1340622, 'name': CreateVM_Task, 'duration_secs': 0.324033} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 521.299203] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 56eee633-2c01-4df2-8d2e-58ad712942f6] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 521.299814] env[62066]: DEBUG oslo_concurrency.lockutils [None req-70656e85-7c6a-408b-a415-8e3cff7746a0 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 521.302367] env[62066]: DEBUG oslo_concurrency.lockutils [None req-70656e85-7c6a-408b-a415-8e3cff7746a0 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Acquired lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 521.302367] env[62066]: DEBUG oslo_concurrency.lockutils [None req-70656e85-7c6a-408b-a415-8e3cff7746a0 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 521.302367] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e765362f-09a0-477a-818e-29059c0e4ea1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 521.308724] env[62066]: DEBUG oslo_vmware.api [None req-70656e85-7c6a-408b-a415-8e3cff7746a0 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Waiting for the task: (returnval){ [ 521.308724] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]52bfa806-538f-61d1-73f1-00d3811cac29" [ 521.308724] env[62066]: _type = "Task" [ 521.308724] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 521.316826] env[62066]: DEBUG oslo_vmware.api [None req-70656e85-7c6a-408b-a415-8e3cff7746a0 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52bfa806-538f-61d1-73f1-00d3811cac29, 'name': SearchDatastore_Task} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 521.317436] env[62066]: DEBUG oslo_concurrency.lockutils [None req-70656e85-7c6a-408b-a415-8e3cff7746a0 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Releasing lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 521.317822] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-70656e85-7c6a-408b-a415-8e3cff7746a0 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] [instance: 56eee633-2c01-4df2-8d2e-58ad712942f6] Processing image 50ff584c-3b50-4395-af07-3e66769bc9f7 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 521.318126] env[62066]: DEBUG oslo_concurrency.lockutils [None req-70656e85-7c6a-408b-a415-8e3cff7746a0 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 521.318574] env[62066]: DEBUG oslo_concurrency.lockutils [None req-70656e85-7c6a-408b-a415-8e3cff7746a0 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Acquired lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 521.318694] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-70656e85-7c6a-408b-a415-8e3cff7746a0 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 521.318914] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-daa26aae-ec23-473e-b7c6-8e5955cc47cd {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 521.326049] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-70656e85-7c6a-408b-a415-8e3cff7746a0 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 521.326587] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-70656e85-7c6a-408b-a415-8e3cff7746a0 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 521.327233] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e9d3b618-9213-44f7-8c54-6124a3ed5f0a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 521.333155] env[62066]: DEBUG oslo_vmware.api [None req-70656e85-7c6a-408b-a415-8e3cff7746a0 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Waiting for the task: (returnval){ [ 521.333155] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]5291b76c-4219-3b3d-e532-69140ff22fea" [ 521.333155] env[62066]: _type = "Task" [ 521.333155] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 521.342363] env[62066]: DEBUG oslo_vmware.api [None req-70656e85-7c6a-408b-a415-8e3cff7746a0 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]5291b76c-4219-3b3d-e532-69140ff22fea, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 521.348450] env[62066]: ERROR nova.compute.manager [None req-0e23bf5c-5716-4eba-9510-a3597f398645 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 78745978-2782-4cbd-8fbd-fbdf247cc830, please check neutron logs for more information. [ 521.348450] env[62066]: ERROR nova.compute.manager Traceback (most recent call last): [ 521.348450] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 521.348450] env[62066]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 521.348450] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 521.348450] env[62066]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 521.348450] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 521.348450] env[62066]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 521.348450] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 521.348450] env[62066]: ERROR nova.compute.manager self.force_reraise() [ 521.348450] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 521.348450] env[62066]: ERROR nova.compute.manager raise self.value [ 521.348450] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 521.348450] env[62066]: ERROR nova.compute.manager updated_port = self._update_port( [ 521.348450] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 521.348450] env[62066]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 521.348911] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 521.348911] env[62066]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 521.348911] env[62066]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 78745978-2782-4cbd-8fbd-fbdf247cc830, please check neutron logs for more information. [ 521.348911] env[62066]: ERROR nova.compute.manager [ 521.348911] env[62066]: Traceback (most recent call last): [ 521.348911] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 521.348911] env[62066]: listener.cb(fileno) [ 521.348911] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 521.348911] env[62066]: result = function(*args, **kwargs) [ 521.348911] env[62066]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 521.348911] env[62066]: return func(*args, **kwargs) [ 521.348911] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 521.348911] env[62066]: raise e [ 521.348911] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 521.348911] env[62066]: nwinfo = self.network_api.allocate_for_instance( [ 521.348911] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 521.348911] env[62066]: created_port_ids = self._update_ports_for_instance( [ 521.348911] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 521.348911] env[62066]: with excutils.save_and_reraise_exception(): [ 521.348911] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 521.348911] env[62066]: self.force_reraise() [ 521.348911] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 521.348911] env[62066]: raise self.value [ 521.348911] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 521.348911] env[62066]: updated_port = self._update_port( [ 521.348911] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 521.348911] env[62066]: _ensure_no_port_binding_failure(port) [ 521.348911] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 521.348911] env[62066]: raise exception.PortBindingFailed(port_id=port['id']) [ 521.349843] env[62066]: nova.exception.PortBindingFailed: Binding failed for port 78745978-2782-4cbd-8fbd-fbdf247cc830, please check neutron logs for more information. [ 521.349843] env[62066]: Removing descriptor: 19 [ 521.349843] env[62066]: ERROR nova.compute.manager [None req-0e23bf5c-5716-4eba-9510-a3597f398645 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 78745978-2782-4cbd-8fbd-fbdf247cc830, please check neutron logs for more information. [ 521.349843] env[62066]: ERROR nova.compute.manager [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] Traceback (most recent call last): [ 521.349843] env[62066]: ERROR nova.compute.manager [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 521.349843] env[62066]: ERROR nova.compute.manager [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] yield resources [ 521.349843] env[62066]: ERROR nova.compute.manager [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 521.349843] env[62066]: ERROR nova.compute.manager [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] self.driver.spawn(context, instance, image_meta, [ 521.349843] env[62066]: ERROR nova.compute.manager [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 521.349843] env[62066]: ERROR nova.compute.manager [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] self._vmops.spawn(context, instance, image_meta, injected_files, [ 521.349843] env[62066]: ERROR nova.compute.manager [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 521.349843] env[62066]: ERROR nova.compute.manager [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] vm_ref = self.build_virtual_machine(instance, [ 521.350190] env[62066]: ERROR nova.compute.manager [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 521.350190] env[62066]: ERROR nova.compute.manager [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] vif_infos = vmwarevif.get_vif_info(self._session, [ 521.350190] env[62066]: ERROR nova.compute.manager [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 521.350190] env[62066]: ERROR nova.compute.manager [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] for vif in network_info: [ 521.350190] env[62066]: ERROR nova.compute.manager [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 521.350190] env[62066]: ERROR nova.compute.manager [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] return self._sync_wrapper(fn, *args, **kwargs) [ 521.350190] env[62066]: ERROR nova.compute.manager [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 521.350190] env[62066]: ERROR nova.compute.manager [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] self.wait() [ 521.350190] env[62066]: ERROR nova.compute.manager [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 521.350190] env[62066]: ERROR nova.compute.manager [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] self[:] = self._gt.wait() [ 521.350190] env[62066]: ERROR nova.compute.manager [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 521.350190] env[62066]: ERROR nova.compute.manager [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] return self._exit_event.wait() [ 521.350190] env[62066]: ERROR nova.compute.manager [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 521.350518] env[62066]: ERROR nova.compute.manager [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] result = hub.switch() [ 521.350518] env[62066]: ERROR nova.compute.manager [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 521.350518] env[62066]: ERROR nova.compute.manager [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] return self.greenlet.switch() [ 521.350518] env[62066]: ERROR nova.compute.manager [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 521.350518] env[62066]: ERROR nova.compute.manager [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] result = function(*args, **kwargs) [ 521.350518] env[62066]: ERROR nova.compute.manager [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 521.350518] env[62066]: ERROR nova.compute.manager [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] return func(*args, **kwargs) [ 521.350518] env[62066]: ERROR nova.compute.manager [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 521.350518] env[62066]: ERROR nova.compute.manager [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] raise e [ 521.350518] env[62066]: ERROR nova.compute.manager [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 521.350518] env[62066]: ERROR nova.compute.manager [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] nwinfo = self.network_api.allocate_for_instance( [ 521.350518] env[62066]: ERROR nova.compute.manager [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 521.350518] env[62066]: ERROR nova.compute.manager [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] created_port_ids = self._update_ports_for_instance( [ 521.350838] env[62066]: ERROR nova.compute.manager [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 521.350838] env[62066]: ERROR nova.compute.manager [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] with excutils.save_and_reraise_exception(): [ 521.350838] env[62066]: ERROR nova.compute.manager [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 521.350838] env[62066]: ERROR nova.compute.manager [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] self.force_reraise() [ 521.350838] env[62066]: ERROR nova.compute.manager [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 521.350838] env[62066]: ERROR nova.compute.manager [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] raise self.value [ 521.350838] env[62066]: ERROR nova.compute.manager [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 521.350838] env[62066]: ERROR nova.compute.manager [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] updated_port = self._update_port( [ 521.350838] env[62066]: ERROR nova.compute.manager [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 521.350838] env[62066]: ERROR nova.compute.manager [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] _ensure_no_port_binding_failure(port) [ 521.350838] env[62066]: ERROR nova.compute.manager [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 521.350838] env[62066]: ERROR nova.compute.manager [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] raise exception.PortBindingFailed(port_id=port['id']) [ 521.351787] env[62066]: ERROR nova.compute.manager [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] nova.exception.PortBindingFailed: Binding failed for port 78745978-2782-4cbd-8fbd-fbdf247cc830, please check neutron logs for more information. [ 521.351787] env[62066]: ERROR nova.compute.manager [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] [ 521.351787] env[62066]: INFO nova.compute.manager [None req-0e23bf5c-5716-4eba-9510-a3597f398645 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] Terminating instance [ 521.351914] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0e23bf5c-5716-4eba-9510-a3597f398645 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] Acquiring lock "refresh_cache-095d69fe-bbd3-4850-9dcd-b4c2290b8352" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 521.352120] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0e23bf5c-5716-4eba-9510-a3597f398645 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] Acquired lock "refresh_cache-095d69fe-bbd3-4850-9dcd-b4c2290b8352" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 521.352361] env[62066]: DEBUG nova.network.neutron [None req-0e23bf5c-5716-4eba-9510-a3597f398645 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 521.360458] env[62066]: DEBUG nova.network.neutron [None req-e2ac9da0-4290-4c63-a476-a96d29c6e38b tempest-ServerDiagnosticsNegativeTest-1631323339 tempest-ServerDiagnosticsNegativeTest-1631323339-project-member] [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 521.370433] env[62066]: DEBUG oslo_concurrency.lockutils [None req-dbc2121a-ef71-4ef4-8f94-3a34690382a8 tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.967s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 521.371157] env[62066]: ERROR nova.compute.manager [None req-dbc2121a-ef71-4ef4-8f94-3a34690382a8 tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 446f5348-3547-40f2-b27c-3762e9b219a7, please check neutron logs for more information. [ 521.371157] env[62066]: ERROR nova.compute.manager [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] Traceback (most recent call last): [ 521.371157] env[62066]: ERROR nova.compute.manager [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 521.371157] env[62066]: ERROR nova.compute.manager [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] self.driver.spawn(context, instance, image_meta, [ 521.371157] env[62066]: ERROR nova.compute.manager [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 521.371157] env[62066]: ERROR nova.compute.manager [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] self._vmops.spawn(context, instance, image_meta, injected_files, [ 521.371157] env[62066]: ERROR nova.compute.manager [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 521.371157] env[62066]: ERROR nova.compute.manager [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] vm_ref = self.build_virtual_machine(instance, [ 521.371157] env[62066]: ERROR nova.compute.manager [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 521.371157] env[62066]: ERROR nova.compute.manager [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] vif_infos = vmwarevif.get_vif_info(self._session, [ 521.371157] env[62066]: ERROR nova.compute.manager [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 521.371461] env[62066]: ERROR nova.compute.manager [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] for vif in network_info: [ 521.371461] env[62066]: ERROR nova.compute.manager [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 521.371461] env[62066]: ERROR nova.compute.manager [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] return self._sync_wrapper(fn, *args, **kwargs) [ 521.371461] env[62066]: ERROR nova.compute.manager [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 521.371461] env[62066]: ERROR nova.compute.manager [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] self.wait() [ 521.371461] env[62066]: ERROR nova.compute.manager [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 521.371461] env[62066]: ERROR nova.compute.manager [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] self[:] = self._gt.wait() [ 521.371461] env[62066]: ERROR nova.compute.manager [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 521.371461] env[62066]: ERROR nova.compute.manager [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] return self._exit_event.wait() [ 521.371461] env[62066]: ERROR nova.compute.manager [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 521.371461] env[62066]: ERROR nova.compute.manager [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] result = hub.switch() [ 521.371461] env[62066]: ERROR nova.compute.manager [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 521.371461] env[62066]: ERROR nova.compute.manager [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] return self.greenlet.switch() [ 521.371785] env[62066]: ERROR nova.compute.manager [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 521.371785] env[62066]: ERROR nova.compute.manager [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] result = function(*args, **kwargs) [ 521.371785] env[62066]: ERROR nova.compute.manager [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 521.371785] env[62066]: ERROR nova.compute.manager [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] return func(*args, **kwargs) [ 521.371785] env[62066]: ERROR nova.compute.manager [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 521.371785] env[62066]: ERROR nova.compute.manager [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] raise e [ 521.371785] env[62066]: ERROR nova.compute.manager [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 521.371785] env[62066]: ERROR nova.compute.manager [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] nwinfo = self.network_api.allocate_for_instance( [ 521.371785] env[62066]: ERROR nova.compute.manager [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 521.371785] env[62066]: ERROR nova.compute.manager [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] created_port_ids = self._update_ports_for_instance( [ 521.371785] env[62066]: ERROR nova.compute.manager [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 521.371785] env[62066]: ERROR nova.compute.manager [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] with excutils.save_and_reraise_exception(): [ 521.371785] env[62066]: ERROR nova.compute.manager [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 521.372137] env[62066]: ERROR nova.compute.manager [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] self.force_reraise() [ 521.372137] env[62066]: ERROR nova.compute.manager [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 521.372137] env[62066]: ERROR nova.compute.manager [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] raise self.value [ 521.372137] env[62066]: ERROR nova.compute.manager [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 521.372137] env[62066]: ERROR nova.compute.manager [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] updated_port = self._update_port( [ 521.372137] env[62066]: ERROR nova.compute.manager [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 521.372137] env[62066]: ERROR nova.compute.manager [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] _ensure_no_port_binding_failure(port) [ 521.372137] env[62066]: ERROR nova.compute.manager [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 521.372137] env[62066]: ERROR nova.compute.manager [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] raise exception.PortBindingFailed(port_id=port['id']) [ 521.372137] env[62066]: ERROR nova.compute.manager [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] nova.exception.PortBindingFailed: Binding failed for port 446f5348-3547-40f2-b27c-3762e9b219a7, please check neutron logs for more information. [ 521.372137] env[62066]: ERROR nova.compute.manager [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] [ 521.372404] env[62066]: DEBUG nova.compute.utils [None req-dbc2121a-ef71-4ef4-8f94-3a34690382a8 tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] Binding failed for port 446f5348-3547-40f2-b27c-3762e9b219a7, please check neutron logs for more information. {{(pid=62066) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 521.374280] env[62066]: DEBUG nova.compute.manager [None req-dbc2121a-ef71-4ef4-8f94-3a34690382a8 tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] Build of instance 9fa0bae6-065d-49ad-b076-0956f1cc3677 was re-scheduled: Binding failed for port 446f5348-3547-40f2-b27c-3762e9b219a7, please check neutron logs for more information. {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 521.375200] env[62066]: DEBUG nova.compute.manager [None req-dbc2121a-ef71-4ef4-8f94-3a34690382a8 tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] Unplugging VIFs for instance {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 521.375200] env[62066]: DEBUG oslo_concurrency.lockutils [None req-dbc2121a-ef71-4ef4-8f94-3a34690382a8 tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] Acquiring lock "refresh_cache-9fa0bae6-065d-49ad-b076-0956f1cc3677" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 521.375200] env[62066]: DEBUG oslo_concurrency.lockutils [None req-dbc2121a-ef71-4ef4-8f94-3a34690382a8 tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] Acquired lock "refresh_cache-9fa0bae6-065d-49ad-b076-0956f1cc3677" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 521.375343] env[62066]: DEBUG nova.network.neutron [None req-dbc2121a-ef71-4ef4-8f94-3a34690382a8 tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 521.376930] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 17.003s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 521.377161] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 521.377350] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62066) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 521.377650] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9eabd10d-5afe-4c06-abad-b512a51e6251 tempest-ServersAdminNegativeTestJSON-732915097 tempest-ServersAdminNegativeTestJSON-732915097-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 11.160s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 521.381468] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb1855ea-c658-45cb-b043-af6c940be411 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 521.390847] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3432f63a-98bd-46df-a9ae-b9ca5270b065 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 521.406331] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-884be156-8e93-410f-8250-3598540ff17e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 521.413750] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73ec6dd5-70de-4363-9380-d5a077a0d76e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 521.447639] env[62066]: DEBUG oslo_concurrency.lockutils [None req-df85dbc4-0217-427e-9c09-82d1a0e6d774 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] Releasing lock "refresh_cache-1eeb879b-6d1e-4746-98b5-12c30d674fa9" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 521.447939] env[62066]: DEBUG nova.compute.manager [None req-df85dbc4-0217-427e-9c09-82d1a0e6d774 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 521.448190] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-df85dbc4-0217-427e-9c09-82d1a0e6d774 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 521.448667] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181481MB free_disk=154GB free_vcpus=48 pci_devices=None {{(pid=62066) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 521.449247] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 521.449485] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-31fe2cbd-9919-477b-9a32-75178cc3de04 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 521.458260] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5149c856-39e9-426b-8c93-e9d3728bfd12 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 521.483231] env[62066]: WARNING nova.virt.vmwareapi.vmops [None req-df85dbc4-0217-427e-9c09-82d1a0e6d774 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 1eeb879b-6d1e-4746-98b5-12c30d674fa9 could not be found. [ 521.483466] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-df85dbc4-0217-427e-9c09-82d1a0e6d774 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 521.483644] env[62066]: INFO nova.compute.manager [None req-df85dbc4-0217-427e-9c09-82d1a0e6d774 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] Took 0.04 seconds to destroy the instance on the hypervisor. [ 521.483878] env[62066]: DEBUG oslo.service.loopingcall [None req-df85dbc4-0217-427e-9c09-82d1a0e6d774 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 521.484109] env[62066]: DEBUG nova.compute.manager [-] [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 521.484203] env[62066]: DEBUG nova.network.neutron [-] [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 521.512807] env[62066]: DEBUG nova.network.neutron [-] [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 521.851042] env[62066]: DEBUG oslo_vmware.api [None req-70656e85-7c6a-408b-a415-8e3cff7746a0 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]5291b76c-4219-3b3d-e532-69140ff22fea, 'name': SearchDatastore_Task, 'duration_secs': 0.00804} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 521.851042] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-88d90ebb-8bc5-4075-a369-66e5f38c3d8f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 521.858086] env[62066]: DEBUG oslo_vmware.api [None req-70656e85-7c6a-408b-a415-8e3cff7746a0 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Waiting for the task: (returnval){ [ 521.858086] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]525d2130-0582-da38-647b-309843b4dd32" [ 521.858086] env[62066]: _type = "Task" [ 521.858086] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 521.863836] env[62066]: INFO nova.compute.manager [None req-e2ac9da0-4290-4c63-a476-a96d29c6e38b tempest-ServerDiagnosticsNegativeTest-1631323339 tempest-ServerDiagnosticsNegativeTest-1631323339-project-member] [instance: 8b29fc32-d27d-4474-84bc-f727f90fde09] Took 1.06 seconds to deallocate network for instance. [ 521.901587] env[62066]: DEBUG oslo_vmware.api [None req-70656e85-7c6a-408b-a415-8e3cff7746a0 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]525d2130-0582-da38-647b-309843b4dd32, 'name': SearchDatastore_Task, 'duration_secs': 0.008759} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 521.901587] env[62066]: DEBUG oslo_concurrency.lockutils [None req-70656e85-7c6a-408b-a415-8e3cff7746a0 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Releasing lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 521.901587] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-70656e85-7c6a-408b-a415-8e3cff7746a0 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk to [datastore2] 56eee633-2c01-4df2-8d2e-58ad712942f6/56eee633-2c01-4df2-8d2e-58ad712942f6.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 521.901587] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8cf8322f-7009-4c6b-881c-aa2074624799 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 521.909539] env[62066]: DEBUG oslo_vmware.api [None req-70656e85-7c6a-408b-a415-8e3cff7746a0 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Waiting for the task: (returnval){ [ 521.909539] env[62066]: value = "task-1340623" [ 521.909539] env[62066]: _type = "Task" [ 521.909539] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 521.922198] env[62066]: DEBUG oslo_vmware.api [None req-70656e85-7c6a-408b-a415-8e3cff7746a0 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Task: {'id': task-1340623, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 521.987095] env[62066]: DEBUG nova.network.neutron [None req-dbc2121a-ef71-4ef4-8f94-3a34690382a8 tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 521.989655] env[62066]: DEBUG nova.network.neutron [None req-0e23bf5c-5716-4eba-9510-a3597f398645 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 522.018276] env[62066]: DEBUG nova.network.neutron [-] [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 522.048019] env[62066]: DEBUG oslo_concurrency.lockutils [None req-fc3cf2dd-2491-4541-a245-5eefc9a5c042 tempest-TenantUsagesTestJSON-970395509 tempest-TenantUsagesTestJSON-970395509-project-member] Acquiring lock "a187f73b-8fb1-42b3-9a07-24d16aa0f152" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 522.048019] env[62066]: DEBUG oslo_concurrency.lockutils [None req-fc3cf2dd-2491-4541-a245-5eefc9a5c042 tempest-TenantUsagesTestJSON-970395509 tempest-TenantUsagesTestJSON-970395509-project-member] Lock "a187f73b-8fb1-42b3-9a07-24d16aa0f152" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 522.178258] env[62066]: DEBUG nova.network.neutron [None req-0e23bf5c-5716-4eba-9510-a3597f398645 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 522.275422] env[62066]: DEBUG nova.network.neutron [None req-dbc2121a-ef71-4ef4-8f94-3a34690382a8 tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 522.358704] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-071a32bb-7578-4744-9a5b-b87a19f84ba1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.367471] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef2ce933-cc51-4abe-9645-a04858d00108 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.405475] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd3fd1b5-570c-4c90-9672-303f75949b89 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.415697] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba8cf65b-8b06-475b-9034-a0dc300a3e8c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.425344] env[62066]: DEBUG oslo_vmware.api [None req-70656e85-7c6a-408b-a415-8e3cff7746a0 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Task: {'id': task-1340623, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.484054} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 522.432769] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-70656e85-7c6a-408b-a415-8e3cff7746a0 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk to [datastore2] 56eee633-2c01-4df2-8d2e-58ad712942f6/56eee633-2c01-4df2-8d2e-58ad712942f6.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 522.433590] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-70656e85-7c6a-408b-a415-8e3cff7746a0 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] [instance: 56eee633-2c01-4df2-8d2e-58ad712942f6] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 522.433590] env[62066]: DEBUG nova.compute.provider_tree [None req-9eabd10d-5afe-4c06-abad-b512a51e6251 tempest-ServersAdminNegativeTestJSON-732915097 tempest-ServersAdminNegativeTestJSON-732915097-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 522.435475] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-cf9c51f0-91e9-42a1-a47d-1a2494222e03 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.440967] env[62066]: DEBUG oslo_vmware.api [None req-70656e85-7c6a-408b-a415-8e3cff7746a0 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Waiting for the task: (returnval){ [ 522.440967] env[62066]: value = "task-1340624" [ 522.440967] env[62066]: _type = "Task" [ 522.440967] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 522.449633] env[62066]: DEBUG oslo_vmware.api [None req-70656e85-7c6a-408b-a415-8e3cff7746a0 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Task: {'id': task-1340624, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 522.523091] env[62066]: INFO nova.compute.manager [-] [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] Took 1.04 seconds to deallocate network for instance. [ 522.525736] env[62066]: DEBUG nova.compute.claims [None req-df85dbc4-0217-427e-9c09-82d1a0e6d774 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] Aborting claim: {{(pid=62066) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 522.526101] env[62066]: DEBUG oslo_concurrency.lockutils [None req-df85dbc4-0217-427e-9c09-82d1a0e6d774 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 522.683324] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0e23bf5c-5716-4eba-9510-a3597f398645 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] Releasing lock "refresh_cache-095d69fe-bbd3-4850-9dcd-b4c2290b8352" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 522.683750] env[62066]: DEBUG nova.compute.manager [None req-0e23bf5c-5716-4eba-9510-a3597f398645 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 522.683993] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-0e23bf5c-5716-4eba-9510-a3597f398645 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 522.684299] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-cc261342-01e5-4bc7-bea5-bb11df7a1648 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.699595] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3be8f2c-28d6-4df0-a86b-fde649f85f97 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.723525] env[62066]: WARNING nova.virt.vmwareapi.vmops [None req-0e23bf5c-5716-4eba-9510-a3597f398645 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 095d69fe-bbd3-4850-9dcd-b4c2290b8352 could not be found. [ 522.724613] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-0e23bf5c-5716-4eba-9510-a3597f398645 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 522.724750] env[62066]: INFO nova.compute.manager [None req-0e23bf5c-5716-4eba-9510-a3597f398645 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] Took 0.04 seconds to destroy the instance on the hypervisor. [ 522.725018] env[62066]: DEBUG oslo.service.loopingcall [None req-0e23bf5c-5716-4eba-9510-a3597f398645 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 522.726338] env[62066]: DEBUG nova.compute.manager [-] [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 522.726402] env[62066]: DEBUG nova.network.neutron [-] [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 522.769155] env[62066]: DEBUG nova.network.neutron [-] [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 522.780242] env[62066]: DEBUG oslo_concurrency.lockutils [None req-dbc2121a-ef71-4ef4-8f94-3a34690382a8 tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] Releasing lock "refresh_cache-9fa0bae6-065d-49ad-b076-0956f1cc3677" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 522.780551] env[62066]: DEBUG nova.compute.manager [None req-dbc2121a-ef71-4ef4-8f94-3a34690382a8 tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 522.780779] env[62066]: DEBUG nova.compute.manager [None req-dbc2121a-ef71-4ef4-8f94-3a34690382a8 tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 522.780969] env[62066]: DEBUG nova.network.neutron [None req-dbc2121a-ef71-4ef4-8f94-3a34690382a8 tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 522.814035] env[62066]: DEBUG nova.network.neutron [None req-dbc2121a-ef71-4ef4-8f94-3a34690382a8 tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 522.820954] env[62066]: ERROR nova.compute.manager [None req-68481586-e4e9-4c46-8b51-aa7845aca824 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 5d597d92-a2ed-414a-babb-5eacc2e2e5c2, please check neutron logs for more information. [ 522.820954] env[62066]: ERROR nova.compute.manager Traceback (most recent call last): [ 522.820954] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 522.820954] env[62066]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 522.820954] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 522.820954] env[62066]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 522.820954] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 522.820954] env[62066]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 522.820954] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 522.820954] env[62066]: ERROR nova.compute.manager self.force_reraise() [ 522.820954] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 522.820954] env[62066]: ERROR nova.compute.manager raise self.value [ 522.820954] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 522.820954] env[62066]: ERROR nova.compute.manager updated_port = self._update_port( [ 522.820954] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 522.820954] env[62066]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 522.821443] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 522.821443] env[62066]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 522.821443] env[62066]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 5d597d92-a2ed-414a-babb-5eacc2e2e5c2, please check neutron logs for more information. [ 522.821443] env[62066]: ERROR nova.compute.manager [ 522.821443] env[62066]: Traceback (most recent call last): [ 522.821443] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 522.821443] env[62066]: listener.cb(fileno) [ 522.821443] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 522.821443] env[62066]: result = function(*args, **kwargs) [ 522.821443] env[62066]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 522.821443] env[62066]: return func(*args, **kwargs) [ 522.821443] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 522.821443] env[62066]: raise e [ 522.821443] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 522.821443] env[62066]: nwinfo = self.network_api.allocate_for_instance( [ 522.821443] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 522.821443] env[62066]: created_port_ids = self._update_ports_for_instance( [ 522.821443] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 522.821443] env[62066]: with excutils.save_and_reraise_exception(): [ 522.821443] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 522.821443] env[62066]: self.force_reraise() [ 522.821443] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 522.821443] env[62066]: raise self.value [ 522.821443] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 522.821443] env[62066]: updated_port = self._update_port( [ 522.821443] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 522.821443] env[62066]: _ensure_no_port_binding_failure(port) [ 522.821443] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 522.821443] env[62066]: raise exception.PortBindingFailed(port_id=port['id']) [ 522.822188] env[62066]: nova.exception.PortBindingFailed: Binding failed for port 5d597d92-a2ed-414a-babb-5eacc2e2e5c2, please check neutron logs for more information. [ 522.822188] env[62066]: Removing descriptor: 15 [ 522.822188] env[62066]: ERROR nova.compute.manager [None req-68481586-e4e9-4c46-8b51-aa7845aca824 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 5d597d92-a2ed-414a-babb-5eacc2e2e5c2, please check neutron logs for more information. [ 522.822188] env[62066]: ERROR nova.compute.manager [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] Traceback (most recent call last): [ 522.822188] env[62066]: ERROR nova.compute.manager [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 522.822188] env[62066]: ERROR nova.compute.manager [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] yield resources [ 522.822188] env[62066]: ERROR nova.compute.manager [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 522.822188] env[62066]: ERROR nova.compute.manager [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] self.driver.spawn(context, instance, image_meta, [ 522.822188] env[62066]: ERROR nova.compute.manager [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 522.822188] env[62066]: ERROR nova.compute.manager [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] self._vmops.spawn(context, instance, image_meta, injected_files, [ 522.822188] env[62066]: ERROR nova.compute.manager [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 522.822188] env[62066]: ERROR nova.compute.manager [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] vm_ref = self.build_virtual_machine(instance, [ 522.822514] env[62066]: ERROR nova.compute.manager [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 522.822514] env[62066]: ERROR nova.compute.manager [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] vif_infos = vmwarevif.get_vif_info(self._session, [ 522.822514] env[62066]: ERROR nova.compute.manager [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 522.822514] env[62066]: ERROR nova.compute.manager [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] for vif in network_info: [ 522.822514] env[62066]: ERROR nova.compute.manager [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 522.822514] env[62066]: ERROR nova.compute.manager [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] return self._sync_wrapper(fn, *args, **kwargs) [ 522.822514] env[62066]: ERROR nova.compute.manager [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 522.822514] env[62066]: ERROR nova.compute.manager [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] self.wait() [ 522.822514] env[62066]: ERROR nova.compute.manager [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 522.822514] env[62066]: ERROR nova.compute.manager [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] self[:] = self._gt.wait() [ 522.822514] env[62066]: ERROR nova.compute.manager [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 522.822514] env[62066]: ERROR nova.compute.manager [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] return self._exit_event.wait() [ 522.822514] env[62066]: ERROR nova.compute.manager [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 522.822913] env[62066]: ERROR nova.compute.manager [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] result = hub.switch() [ 522.822913] env[62066]: ERROR nova.compute.manager [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 522.822913] env[62066]: ERROR nova.compute.manager [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] return self.greenlet.switch() [ 522.822913] env[62066]: ERROR nova.compute.manager [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 522.822913] env[62066]: ERROR nova.compute.manager [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] result = function(*args, **kwargs) [ 522.822913] env[62066]: ERROR nova.compute.manager [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 522.822913] env[62066]: ERROR nova.compute.manager [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] return func(*args, **kwargs) [ 522.822913] env[62066]: ERROR nova.compute.manager [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 522.822913] env[62066]: ERROR nova.compute.manager [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] raise e [ 522.822913] env[62066]: ERROR nova.compute.manager [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 522.822913] env[62066]: ERROR nova.compute.manager [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] nwinfo = self.network_api.allocate_for_instance( [ 522.822913] env[62066]: ERROR nova.compute.manager [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 522.822913] env[62066]: ERROR nova.compute.manager [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] created_port_ids = self._update_ports_for_instance( [ 522.823325] env[62066]: ERROR nova.compute.manager [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 522.823325] env[62066]: ERROR nova.compute.manager [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] with excutils.save_and_reraise_exception(): [ 522.823325] env[62066]: ERROR nova.compute.manager [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 522.823325] env[62066]: ERROR nova.compute.manager [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] self.force_reraise() [ 522.823325] env[62066]: ERROR nova.compute.manager [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 522.823325] env[62066]: ERROR nova.compute.manager [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] raise self.value [ 522.823325] env[62066]: ERROR nova.compute.manager [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 522.823325] env[62066]: ERROR nova.compute.manager [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] updated_port = self._update_port( [ 522.823325] env[62066]: ERROR nova.compute.manager [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 522.823325] env[62066]: ERROR nova.compute.manager [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] _ensure_no_port_binding_failure(port) [ 522.823325] env[62066]: ERROR nova.compute.manager [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 522.823325] env[62066]: ERROR nova.compute.manager [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] raise exception.PortBindingFailed(port_id=port['id']) [ 522.823705] env[62066]: ERROR nova.compute.manager [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] nova.exception.PortBindingFailed: Binding failed for port 5d597d92-a2ed-414a-babb-5eacc2e2e5c2, please check neutron logs for more information. [ 522.823705] env[62066]: ERROR nova.compute.manager [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] [ 522.823705] env[62066]: INFO nova.compute.manager [None req-68481586-e4e9-4c46-8b51-aa7845aca824 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] Terminating instance [ 522.825083] env[62066]: DEBUG oslo_concurrency.lockutils [None req-68481586-e4e9-4c46-8b51-aa7845aca824 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] Acquiring lock "refresh_cache-d522ee4a-9bd9-402b-b3bd-c36f93366249" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 522.825213] env[62066]: DEBUG oslo_concurrency.lockutils [None req-68481586-e4e9-4c46-8b51-aa7845aca824 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] Acquired lock "refresh_cache-d522ee4a-9bd9-402b-b3bd-c36f93366249" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 522.825381] env[62066]: DEBUG nova.network.neutron [None req-68481586-e4e9-4c46-8b51-aa7845aca824 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 522.940497] env[62066]: DEBUG nova.scheduler.client.report [None req-9eabd10d-5afe-4c06-abad-b512a51e6251 tempest-ServersAdminNegativeTestJSON-732915097 tempest-ServersAdminNegativeTestJSON-732915097-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 522.944541] env[62066]: INFO nova.scheduler.client.report [None req-e2ac9da0-4290-4c63-a476-a96d29c6e38b tempest-ServerDiagnosticsNegativeTest-1631323339 tempest-ServerDiagnosticsNegativeTest-1631323339-project-member] Deleted allocations for instance 8b29fc32-d27d-4474-84bc-f727f90fde09 [ 522.960846] env[62066]: DEBUG oslo_vmware.api [None req-70656e85-7c6a-408b-a415-8e3cff7746a0 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Task: {'id': task-1340624, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.145472} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 522.961208] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-70656e85-7c6a-408b-a415-8e3cff7746a0 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] [instance: 56eee633-2c01-4df2-8d2e-58ad712942f6] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 522.961992] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b283ceb-e795-42c6-ad04-444f95942122 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.983572] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-70656e85-7c6a-408b-a415-8e3cff7746a0 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] [instance: 56eee633-2c01-4df2-8d2e-58ad712942f6] Reconfiguring VM instance instance-00000005 to attach disk [datastore2] 56eee633-2c01-4df2-8d2e-58ad712942f6/56eee633-2c01-4df2-8d2e-58ad712942f6.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 522.984559] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-88c3d275-0762-49b5-ad70-70339b98f770 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.007213] env[62066]: DEBUG oslo_vmware.api [None req-70656e85-7c6a-408b-a415-8e3cff7746a0 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Waiting for the task: (returnval){ [ 523.007213] env[62066]: value = "task-1340625" [ 523.007213] env[62066]: _type = "Task" [ 523.007213] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 523.014562] env[62066]: DEBUG oslo_vmware.api [None req-70656e85-7c6a-408b-a415-8e3cff7746a0 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Task: {'id': task-1340625, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 523.116052] env[62066]: DEBUG nova.compute.manager [req-38d91511-91ef-4438-a080-42bcb0ad12dc req-25c22b77-d92b-43e3-869e-06c082e662a8 service nova] [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] Received event network-vif-deleted-e279d463-9a92-4d4b-913e-d08e912e8b3e {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 523.116432] env[62066]: DEBUG nova.compute.manager [req-38d91511-91ef-4438-a080-42bcb0ad12dc req-25c22b77-d92b-43e3-869e-06c082e662a8 service nova] [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] Received event network-changed-78745978-2782-4cbd-8fbd-fbdf247cc830 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 523.116712] env[62066]: DEBUG nova.compute.manager [req-38d91511-91ef-4438-a080-42bcb0ad12dc req-25c22b77-d92b-43e3-869e-06c082e662a8 service nova] [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] Refreshing instance network info cache due to event network-changed-78745978-2782-4cbd-8fbd-fbdf247cc830. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 523.117317] env[62066]: DEBUG oslo_concurrency.lockutils [req-38d91511-91ef-4438-a080-42bcb0ad12dc req-25c22b77-d92b-43e3-869e-06c082e662a8 service nova] Acquiring lock "refresh_cache-095d69fe-bbd3-4850-9dcd-b4c2290b8352" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 523.121766] env[62066]: DEBUG oslo_concurrency.lockutils [req-38d91511-91ef-4438-a080-42bcb0ad12dc req-25c22b77-d92b-43e3-869e-06c082e662a8 service nova] Acquired lock "refresh_cache-095d69fe-bbd3-4850-9dcd-b4c2290b8352" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 523.121766] env[62066]: DEBUG nova.network.neutron [req-38d91511-91ef-4438-a080-42bcb0ad12dc req-25c22b77-d92b-43e3-869e-06c082e662a8 service nova] [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] Refreshing network info cache for port 78745978-2782-4cbd-8fbd-fbdf247cc830 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 523.272592] env[62066]: DEBUG nova.network.neutron [-] [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 523.320081] env[62066]: DEBUG nova.network.neutron [None req-dbc2121a-ef71-4ef4-8f94-3a34690382a8 tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 523.371372] env[62066]: DEBUG nova.network.neutron [None req-68481586-e4e9-4c46-8b51-aa7845aca824 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 523.451941] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9eabd10d-5afe-4c06-abad-b512a51e6251 tempest-ServersAdminNegativeTestJSON-732915097 tempest-ServersAdminNegativeTestJSON-732915097-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.074s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 523.452613] env[62066]: ERROR nova.compute.manager [None req-9eabd10d-5afe-4c06-abad-b512a51e6251 tempest-ServersAdminNegativeTestJSON-732915097 tempest-ServersAdminNegativeTestJSON-732915097-project-member] [instance: 21010917-f820-4163-848d-e4141c89b13d] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 358d9a83-0d84-4a7b-8e89-6754aadd6c84, please check neutron logs for more information. [ 523.452613] env[62066]: ERROR nova.compute.manager [instance: 21010917-f820-4163-848d-e4141c89b13d] Traceback (most recent call last): [ 523.452613] env[62066]: ERROR nova.compute.manager [instance: 21010917-f820-4163-848d-e4141c89b13d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 523.452613] env[62066]: ERROR nova.compute.manager [instance: 21010917-f820-4163-848d-e4141c89b13d] self.driver.spawn(context, instance, image_meta, [ 523.452613] env[62066]: ERROR nova.compute.manager [instance: 21010917-f820-4163-848d-e4141c89b13d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 523.452613] env[62066]: ERROR nova.compute.manager [instance: 21010917-f820-4163-848d-e4141c89b13d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 523.452613] env[62066]: ERROR nova.compute.manager [instance: 21010917-f820-4163-848d-e4141c89b13d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 523.452613] env[62066]: ERROR nova.compute.manager [instance: 21010917-f820-4163-848d-e4141c89b13d] vm_ref = self.build_virtual_machine(instance, [ 523.452613] env[62066]: ERROR nova.compute.manager [instance: 21010917-f820-4163-848d-e4141c89b13d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 523.452613] env[62066]: ERROR nova.compute.manager [instance: 21010917-f820-4163-848d-e4141c89b13d] vif_infos = vmwarevif.get_vif_info(self._session, [ 523.452613] env[62066]: ERROR nova.compute.manager [instance: 21010917-f820-4163-848d-e4141c89b13d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 523.452951] env[62066]: ERROR nova.compute.manager [instance: 21010917-f820-4163-848d-e4141c89b13d] for vif in network_info: [ 523.452951] env[62066]: ERROR nova.compute.manager [instance: 21010917-f820-4163-848d-e4141c89b13d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 523.452951] env[62066]: ERROR nova.compute.manager [instance: 21010917-f820-4163-848d-e4141c89b13d] return self._sync_wrapper(fn, *args, **kwargs) [ 523.452951] env[62066]: ERROR nova.compute.manager [instance: 21010917-f820-4163-848d-e4141c89b13d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 523.452951] env[62066]: ERROR nova.compute.manager [instance: 21010917-f820-4163-848d-e4141c89b13d] self.wait() [ 523.452951] env[62066]: ERROR nova.compute.manager [instance: 21010917-f820-4163-848d-e4141c89b13d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 523.452951] env[62066]: ERROR nova.compute.manager [instance: 21010917-f820-4163-848d-e4141c89b13d] self[:] = self._gt.wait() [ 523.452951] env[62066]: ERROR nova.compute.manager [instance: 21010917-f820-4163-848d-e4141c89b13d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 523.452951] env[62066]: ERROR nova.compute.manager [instance: 21010917-f820-4163-848d-e4141c89b13d] return self._exit_event.wait() [ 523.452951] env[62066]: ERROR nova.compute.manager [instance: 21010917-f820-4163-848d-e4141c89b13d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 523.452951] env[62066]: ERROR nova.compute.manager [instance: 21010917-f820-4163-848d-e4141c89b13d] result = hub.switch() [ 523.452951] env[62066]: ERROR nova.compute.manager [instance: 21010917-f820-4163-848d-e4141c89b13d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 523.452951] env[62066]: ERROR nova.compute.manager [instance: 21010917-f820-4163-848d-e4141c89b13d] return self.greenlet.switch() [ 523.453290] env[62066]: ERROR nova.compute.manager [instance: 21010917-f820-4163-848d-e4141c89b13d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 523.453290] env[62066]: ERROR nova.compute.manager [instance: 21010917-f820-4163-848d-e4141c89b13d] result = function(*args, **kwargs) [ 523.453290] env[62066]: ERROR nova.compute.manager [instance: 21010917-f820-4163-848d-e4141c89b13d] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 523.453290] env[62066]: ERROR nova.compute.manager [instance: 21010917-f820-4163-848d-e4141c89b13d] return func(*args, **kwargs) [ 523.453290] env[62066]: ERROR nova.compute.manager [instance: 21010917-f820-4163-848d-e4141c89b13d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 523.453290] env[62066]: ERROR nova.compute.manager [instance: 21010917-f820-4163-848d-e4141c89b13d] raise e [ 523.453290] env[62066]: ERROR nova.compute.manager [instance: 21010917-f820-4163-848d-e4141c89b13d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 523.453290] env[62066]: ERROR nova.compute.manager [instance: 21010917-f820-4163-848d-e4141c89b13d] nwinfo = self.network_api.allocate_for_instance( [ 523.453290] env[62066]: ERROR nova.compute.manager [instance: 21010917-f820-4163-848d-e4141c89b13d] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 523.453290] env[62066]: ERROR nova.compute.manager [instance: 21010917-f820-4163-848d-e4141c89b13d] created_port_ids = self._update_ports_for_instance( [ 523.453290] env[62066]: ERROR nova.compute.manager [instance: 21010917-f820-4163-848d-e4141c89b13d] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 523.453290] env[62066]: ERROR nova.compute.manager [instance: 21010917-f820-4163-848d-e4141c89b13d] with excutils.save_and_reraise_exception(): [ 523.453290] env[62066]: ERROR nova.compute.manager [instance: 21010917-f820-4163-848d-e4141c89b13d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 523.453728] env[62066]: ERROR nova.compute.manager [instance: 21010917-f820-4163-848d-e4141c89b13d] self.force_reraise() [ 523.453728] env[62066]: ERROR nova.compute.manager [instance: 21010917-f820-4163-848d-e4141c89b13d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 523.453728] env[62066]: ERROR nova.compute.manager [instance: 21010917-f820-4163-848d-e4141c89b13d] raise self.value [ 523.453728] env[62066]: ERROR nova.compute.manager [instance: 21010917-f820-4163-848d-e4141c89b13d] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 523.453728] env[62066]: ERROR nova.compute.manager [instance: 21010917-f820-4163-848d-e4141c89b13d] updated_port = self._update_port( [ 523.453728] env[62066]: ERROR nova.compute.manager [instance: 21010917-f820-4163-848d-e4141c89b13d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 523.453728] env[62066]: ERROR nova.compute.manager [instance: 21010917-f820-4163-848d-e4141c89b13d] _ensure_no_port_binding_failure(port) [ 523.453728] env[62066]: ERROR nova.compute.manager [instance: 21010917-f820-4163-848d-e4141c89b13d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 523.453728] env[62066]: ERROR nova.compute.manager [instance: 21010917-f820-4163-848d-e4141c89b13d] raise exception.PortBindingFailed(port_id=port['id']) [ 523.453728] env[62066]: ERROR nova.compute.manager [instance: 21010917-f820-4163-848d-e4141c89b13d] nova.exception.PortBindingFailed: Binding failed for port 358d9a83-0d84-4a7b-8e89-6754aadd6c84, please check neutron logs for more information. [ 523.453728] env[62066]: ERROR nova.compute.manager [instance: 21010917-f820-4163-848d-e4141c89b13d] [ 523.454744] env[62066]: DEBUG nova.compute.utils [None req-9eabd10d-5afe-4c06-abad-b512a51e6251 tempest-ServersAdminNegativeTestJSON-732915097 tempest-ServersAdminNegativeTestJSON-732915097-project-member] [instance: 21010917-f820-4163-848d-e4141c89b13d] Binding failed for port 358d9a83-0d84-4a7b-8e89-6754aadd6c84, please check neutron logs for more information. {{(pid=62066) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 523.456695] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9838e0c6-e0b2-45d5-9c5b-a0a0b9c60219 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.106s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 523.461344] env[62066]: INFO nova.compute.claims [None req-9838e0c6-e0b2-45d5-9c5b-a0a0b9c60219 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] [instance: 64f9698e-16f5-447a-914a-9e936e05d101] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 523.464264] env[62066]: DEBUG nova.compute.manager [None req-9eabd10d-5afe-4c06-abad-b512a51e6251 tempest-ServersAdminNegativeTestJSON-732915097 tempest-ServersAdminNegativeTestJSON-732915097-project-member] [instance: 21010917-f820-4163-848d-e4141c89b13d] Build of instance 21010917-f820-4163-848d-e4141c89b13d was re-scheduled: Binding failed for port 358d9a83-0d84-4a7b-8e89-6754aadd6c84, please check neutron logs for more information. {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 523.464765] env[62066]: DEBUG nova.compute.manager [None req-9eabd10d-5afe-4c06-abad-b512a51e6251 tempest-ServersAdminNegativeTestJSON-732915097 tempest-ServersAdminNegativeTestJSON-732915097-project-member] [instance: 21010917-f820-4163-848d-e4141c89b13d] Unplugging VIFs for instance {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 523.465015] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9eabd10d-5afe-4c06-abad-b512a51e6251 tempest-ServersAdminNegativeTestJSON-732915097 tempest-ServersAdminNegativeTestJSON-732915097-project-member] Acquiring lock "refresh_cache-21010917-f820-4163-848d-e4141c89b13d" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 523.465160] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9eabd10d-5afe-4c06-abad-b512a51e6251 tempest-ServersAdminNegativeTestJSON-732915097 tempest-ServersAdminNegativeTestJSON-732915097-project-member] Acquired lock "refresh_cache-21010917-f820-4163-848d-e4141c89b13d" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 523.465313] env[62066]: DEBUG nova.network.neutron [None req-9eabd10d-5afe-4c06-abad-b512a51e6251 tempest-ServersAdminNegativeTestJSON-732915097 tempest-ServersAdminNegativeTestJSON-732915097-project-member] [instance: 21010917-f820-4163-848d-e4141c89b13d] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 523.466497] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e2ac9da0-4290-4c63-a476-a96d29c6e38b tempest-ServerDiagnosticsNegativeTest-1631323339 tempest-ServerDiagnosticsNegativeTest-1631323339-project-member] Lock "8b29fc32-d27d-4474-84bc-f727f90fde09" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 31.079s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 523.518904] env[62066]: DEBUG oslo_vmware.api [None req-70656e85-7c6a-408b-a415-8e3cff7746a0 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Task: {'id': task-1340625, 'name': ReconfigVM_Task, 'duration_secs': 0.273951} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 523.519159] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-70656e85-7c6a-408b-a415-8e3cff7746a0 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] [instance: 56eee633-2c01-4df2-8d2e-58ad712942f6] Reconfigured VM instance instance-00000005 to attach disk [datastore2] 56eee633-2c01-4df2-8d2e-58ad712942f6/56eee633-2c01-4df2-8d2e-58ad712942f6.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 523.519744] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-fa7b57c8-a4a0-4064-9dd4-8ffc180d0837 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.526421] env[62066]: DEBUG oslo_vmware.api [None req-70656e85-7c6a-408b-a415-8e3cff7746a0 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Waiting for the task: (returnval){ [ 523.526421] env[62066]: value = "task-1340626" [ 523.526421] env[62066]: _type = "Task" [ 523.526421] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 523.539744] env[62066]: DEBUG oslo_vmware.api [None req-70656e85-7c6a-408b-a415-8e3cff7746a0 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Task: {'id': task-1340626, 'name': Rename_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 523.588198] env[62066]: DEBUG nova.network.neutron [None req-68481586-e4e9-4c46-8b51-aa7845aca824 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 523.634155] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e7221104-9243-4d76-b03e-5782ae8f55d4 tempest-FloatingIPsAssociationNegativeTestJSON-961065153 tempest-FloatingIPsAssociationNegativeTestJSON-961065153-project-member] Acquiring lock "eee6b9e6-2707-4a3b-925b-a7f50cdc32cb" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 523.634480] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e7221104-9243-4d76-b03e-5782ae8f55d4 tempest-FloatingIPsAssociationNegativeTestJSON-961065153 tempest-FloatingIPsAssociationNegativeTestJSON-961065153-project-member] Lock "eee6b9e6-2707-4a3b-925b-a7f50cdc32cb" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 523.650639] env[62066]: DEBUG nova.network.neutron [req-38d91511-91ef-4438-a080-42bcb0ad12dc req-25c22b77-d92b-43e3-869e-06c082e662a8 service nova] [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 523.758871] env[62066]: DEBUG nova.network.neutron [req-38d91511-91ef-4438-a080-42bcb0ad12dc req-25c22b77-d92b-43e3-869e-06c082e662a8 service nova] [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 523.775329] env[62066]: INFO nova.compute.manager [-] [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] Took 1.05 seconds to deallocate network for instance. [ 523.781714] env[62066]: DEBUG nova.compute.claims [None req-0e23bf5c-5716-4eba-9510-a3597f398645 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] Aborting claim: {{(pid=62066) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 523.781903] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0e23bf5c-5716-4eba-9510-a3597f398645 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 523.822542] env[62066]: INFO nova.compute.manager [None req-dbc2121a-ef71-4ef4-8f94-3a34690382a8 tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] [instance: 9fa0bae6-065d-49ad-b076-0956f1cc3677] Took 1.04 seconds to deallocate network for instance. [ 523.976252] env[62066]: DEBUG nova.compute.manager [None req-d274d4c0-1d30-44ad-b71d-9f542d291186 tempest-VolumesAssistedSnapshotsTest-1416743605 tempest-VolumesAssistedSnapshotsTest-1416743605-project-member] [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 524.003899] env[62066]: DEBUG nova.network.neutron [None req-9eabd10d-5afe-4c06-abad-b512a51e6251 tempest-ServersAdminNegativeTestJSON-732915097 tempest-ServersAdminNegativeTestJSON-732915097-project-member] [instance: 21010917-f820-4163-848d-e4141c89b13d] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 524.046010] env[62066]: DEBUG oslo_vmware.api [None req-70656e85-7c6a-408b-a415-8e3cff7746a0 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Task: {'id': task-1340626, 'name': Rename_Task, 'duration_secs': 0.139451} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 524.046312] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-70656e85-7c6a-408b-a415-8e3cff7746a0 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] [instance: 56eee633-2c01-4df2-8d2e-58ad712942f6] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 524.046984] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f4030330-9ed2-46bc-8505-4496203f16da {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.060088] env[62066]: DEBUG oslo_vmware.api [None req-70656e85-7c6a-408b-a415-8e3cff7746a0 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Waiting for the task: (returnval){ [ 524.060088] env[62066]: value = "task-1340627" [ 524.060088] env[62066]: _type = "Task" [ 524.060088] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 524.070032] env[62066]: DEBUG oslo_vmware.api [None req-70656e85-7c6a-408b-a415-8e3cff7746a0 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Task: {'id': task-1340627, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 524.091978] env[62066]: DEBUG oslo_concurrency.lockutils [None req-68481586-e4e9-4c46-8b51-aa7845aca824 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] Releasing lock "refresh_cache-d522ee4a-9bd9-402b-b3bd-c36f93366249" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 524.092733] env[62066]: DEBUG nova.compute.manager [None req-68481586-e4e9-4c46-8b51-aa7845aca824 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 524.093043] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-68481586-e4e9-4c46-8b51-aa7845aca824 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 524.094116] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-afa47297-76eb-41b2-b19e-4a9af4a2f695 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.097505] env[62066]: DEBUG nova.network.neutron [None req-9eabd10d-5afe-4c06-abad-b512a51e6251 tempest-ServersAdminNegativeTestJSON-732915097 tempest-ServersAdminNegativeTestJSON-732915097-project-member] [instance: 21010917-f820-4163-848d-e4141c89b13d] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 524.104214] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af0c187e-ed30-4d65-857e-b2dcf89ae8ca {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.128410] env[62066]: WARNING nova.virt.vmwareapi.vmops [None req-68481586-e4e9-4c46-8b51-aa7845aca824 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance d522ee4a-9bd9-402b-b3bd-c36f93366249 could not be found. [ 524.128649] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-68481586-e4e9-4c46-8b51-aa7845aca824 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 524.128826] env[62066]: INFO nova.compute.manager [None req-68481586-e4e9-4c46-8b51-aa7845aca824 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] Took 0.04 seconds to destroy the instance on the hypervisor. [ 524.129082] env[62066]: DEBUG oslo.service.loopingcall [None req-68481586-e4e9-4c46-8b51-aa7845aca824 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 524.129491] env[62066]: DEBUG nova.compute.manager [-] [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 524.129597] env[62066]: DEBUG nova.network.neutron [-] [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 524.157321] env[62066]: DEBUG nova.network.neutron [-] [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 524.264301] env[62066]: DEBUG oslo_concurrency.lockutils [req-38d91511-91ef-4438-a080-42bcb0ad12dc req-25c22b77-d92b-43e3-869e-06c082e662a8 service nova] Releasing lock "refresh_cache-095d69fe-bbd3-4850-9dcd-b4c2290b8352" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 524.518767] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d274d4c0-1d30-44ad-b71d-9f542d291186 tempest-VolumesAssistedSnapshotsTest-1416743605 tempest-VolumesAssistedSnapshotsTest-1416743605-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 524.575746] env[62066]: DEBUG oslo_vmware.api [None req-70656e85-7c6a-408b-a415-8e3cff7746a0 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Task: {'id': task-1340627, 'name': PowerOnVM_Task, 'duration_secs': 0.508832} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 524.577884] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-70656e85-7c6a-408b-a415-8e3cff7746a0 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] [instance: 56eee633-2c01-4df2-8d2e-58ad712942f6] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 524.578155] env[62066]: DEBUG nova.compute.manager [None req-70656e85-7c6a-408b-a415-8e3cff7746a0 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] [instance: 56eee633-2c01-4df2-8d2e-58ad712942f6] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 524.579270] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e81b8558-619c-4718-b19a-541622844532 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.599748] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9eabd10d-5afe-4c06-abad-b512a51e6251 tempest-ServersAdminNegativeTestJSON-732915097 tempest-ServersAdminNegativeTestJSON-732915097-project-member] Releasing lock "refresh_cache-21010917-f820-4163-848d-e4141c89b13d" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 524.599748] env[62066]: DEBUG nova.compute.manager [None req-9eabd10d-5afe-4c06-abad-b512a51e6251 tempest-ServersAdminNegativeTestJSON-732915097 tempest-ServersAdminNegativeTestJSON-732915097-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 524.599921] env[62066]: DEBUG nova.compute.manager [None req-9eabd10d-5afe-4c06-abad-b512a51e6251 tempest-ServersAdminNegativeTestJSON-732915097 tempest-ServersAdminNegativeTestJSON-732915097-project-member] [instance: 21010917-f820-4163-848d-e4141c89b13d] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 524.600113] env[62066]: DEBUG nova.network.neutron [None req-9eabd10d-5afe-4c06-abad-b512a51e6251 tempest-ServersAdminNegativeTestJSON-732915097 tempest-ServersAdminNegativeTestJSON-732915097-project-member] [instance: 21010917-f820-4163-848d-e4141c89b13d] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 524.625488] env[62066]: DEBUG nova.network.neutron [None req-9eabd10d-5afe-4c06-abad-b512a51e6251 tempest-ServersAdminNegativeTestJSON-732915097 tempest-ServersAdminNegativeTestJSON-732915097-project-member] [instance: 21010917-f820-4163-848d-e4141c89b13d] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 524.659905] env[62066]: DEBUG nova.network.neutron [-] [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 524.739256] env[62066]: ERROR nova.compute.manager [None req-99d923d5-ef6f-43f2-959c-4b252ac1d398 tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 04cc7ab7-caaa-4afc-afd7-b4230a6e2326, please check neutron logs for more information. [ 524.739256] env[62066]: ERROR nova.compute.manager Traceback (most recent call last): [ 524.739256] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 524.739256] env[62066]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 524.739256] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 524.739256] env[62066]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 524.739256] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 524.739256] env[62066]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 524.739256] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 524.739256] env[62066]: ERROR nova.compute.manager self.force_reraise() [ 524.739256] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 524.739256] env[62066]: ERROR nova.compute.manager raise self.value [ 524.739256] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 524.739256] env[62066]: ERROR nova.compute.manager updated_port = self._update_port( [ 524.739256] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 524.739256] env[62066]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 524.739704] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 524.739704] env[62066]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 524.739704] env[62066]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 04cc7ab7-caaa-4afc-afd7-b4230a6e2326, please check neutron logs for more information. [ 524.739704] env[62066]: ERROR nova.compute.manager [ 524.739704] env[62066]: Traceback (most recent call last): [ 524.739704] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 524.739704] env[62066]: listener.cb(fileno) [ 524.739704] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 524.739704] env[62066]: result = function(*args, **kwargs) [ 524.739704] env[62066]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 524.739704] env[62066]: return func(*args, **kwargs) [ 524.739704] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 524.739704] env[62066]: raise e [ 524.739704] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 524.739704] env[62066]: nwinfo = self.network_api.allocate_for_instance( [ 524.739704] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 524.739704] env[62066]: created_port_ids = self._update_ports_for_instance( [ 524.739704] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 524.739704] env[62066]: with excutils.save_and_reraise_exception(): [ 524.739704] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 524.739704] env[62066]: self.force_reraise() [ 524.739704] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 524.739704] env[62066]: raise self.value [ 524.739704] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 524.739704] env[62066]: updated_port = self._update_port( [ 524.739704] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 524.739704] env[62066]: _ensure_no_port_binding_failure(port) [ 524.739704] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 524.739704] env[62066]: raise exception.PortBindingFailed(port_id=port['id']) [ 524.740546] env[62066]: nova.exception.PortBindingFailed: Binding failed for port 04cc7ab7-caaa-4afc-afd7-b4230a6e2326, please check neutron logs for more information. [ 524.740546] env[62066]: Removing descriptor: 20 [ 524.740546] env[62066]: ERROR nova.compute.manager [None req-99d923d5-ef6f-43f2-959c-4b252ac1d398 tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 04cc7ab7-caaa-4afc-afd7-b4230a6e2326, please check neutron logs for more information. [ 524.740546] env[62066]: ERROR nova.compute.manager [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] Traceback (most recent call last): [ 524.740546] env[62066]: ERROR nova.compute.manager [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 524.740546] env[62066]: ERROR nova.compute.manager [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] yield resources [ 524.740546] env[62066]: ERROR nova.compute.manager [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 524.740546] env[62066]: ERROR nova.compute.manager [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] self.driver.spawn(context, instance, image_meta, [ 524.740546] env[62066]: ERROR nova.compute.manager [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 524.740546] env[62066]: ERROR nova.compute.manager [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 524.740546] env[62066]: ERROR nova.compute.manager [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 524.740546] env[62066]: ERROR nova.compute.manager [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] vm_ref = self.build_virtual_machine(instance, [ 524.741481] env[62066]: ERROR nova.compute.manager [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 524.741481] env[62066]: ERROR nova.compute.manager [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] vif_infos = vmwarevif.get_vif_info(self._session, [ 524.741481] env[62066]: ERROR nova.compute.manager [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 524.741481] env[62066]: ERROR nova.compute.manager [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] for vif in network_info: [ 524.741481] env[62066]: ERROR nova.compute.manager [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 524.741481] env[62066]: ERROR nova.compute.manager [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] return self._sync_wrapper(fn, *args, **kwargs) [ 524.741481] env[62066]: ERROR nova.compute.manager [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 524.741481] env[62066]: ERROR nova.compute.manager [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] self.wait() [ 524.741481] env[62066]: ERROR nova.compute.manager [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 524.741481] env[62066]: ERROR nova.compute.manager [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] self[:] = self._gt.wait() [ 524.741481] env[62066]: ERROR nova.compute.manager [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 524.741481] env[62066]: ERROR nova.compute.manager [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] return self._exit_event.wait() [ 524.741481] env[62066]: ERROR nova.compute.manager [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 524.741815] env[62066]: ERROR nova.compute.manager [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] result = hub.switch() [ 524.741815] env[62066]: ERROR nova.compute.manager [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 524.741815] env[62066]: ERROR nova.compute.manager [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] return self.greenlet.switch() [ 524.741815] env[62066]: ERROR nova.compute.manager [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 524.741815] env[62066]: ERROR nova.compute.manager [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] result = function(*args, **kwargs) [ 524.741815] env[62066]: ERROR nova.compute.manager [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 524.741815] env[62066]: ERROR nova.compute.manager [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] return func(*args, **kwargs) [ 524.741815] env[62066]: ERROR nova.compute.manager [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 524.741815] env[62066]: ERROR nova.compute.manager [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] raise e [ 524.741815] env[62066]: ERROR nova.compute.manager [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 524.741815] env[62066]: ERROR nova.compute.manager [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] nwinfo = self.network_api.allocate_for_instance( [ 524.741815] env[62066]: ERROR nova.compute.manager [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 524.741815] env[62066]: ERROR nova.compute.manager [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] created_port_ids = self._update_ports_for_instance( [ 524.743399] env[62066]: ERROR nova.compute.manager [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 524.743399] env[62066]: ERROR nova.compute.manager [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] with excutils.save_and_reraise_exception(): [ 524.743399] env[62066]: ERROR nova.compute.manager [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 524.743399] env[62066]: ERROR nova.compute.manager [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] self.force_reraise() [ 524.743399] env[62066]: ERROR nova.compute.manager [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 524.743399] env[62066]: ERROR nova.compute.manager [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] raise self.value [ 524.743399] env[62066]: ERROR nova.compute.manager [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 524.743399] env[62066]: ERROR nova.compute.manager [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] updated_port = self._update_port( [ 524.743399] env[62066]: ERROR nova.compute.manager [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 524.743399] env[62066]: ERROR nova.compute.manager [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] _ensure_no_port_binding_failure(port) [ 524.743399] env[62066]: ERROR nova.compute.manager [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 524.743399] env[62066]: ERROR nova.compute.manager [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] raise exception.PortBindingFailed(port_id=port['id']) [ 524.743912] env[62066]: ERROR nova.compute.manager [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] nova.exception.PortBindingFailed: Binding failed for port 04cc7ab7-caaa-4afc-afd7-b4230a6e2326, please check neutron logs for more information. [ 524.743912] env[62066]: ERROR nova.compute.manager [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] [ 524.743912] env[62066]: INFO nova.compute.manager [None req-99d923d5-ef6f-43f2-959c-4b252ac1d398 tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] Terminating instance [ 524.747225] env[62066]: DEBUG oslo_concurrency.lockutils [None req-99d923d5-ef6f-43f2-959c-4b252ac1d398 tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] Acquiring lock "refresh_cache-68b65162-23a6-464f-a2f9-1635bea8786f" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 524.747923] env[62066]: DEBUG oslo_concurrency.lockutils [None req-99d923d5-ef6f-43f2-959c-4b252ac1d398 tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] Acquired lock "refresh_cache-68b65162-23a6-464f-a2f9-1635bea8786f" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 524.749951] env[62066]: DEBUG nova.network.neutron [None req-99d923d5-ef6f-43f2-959c-4b252ac1d398 tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 524.880090] env[62066]: INFO nova.scheduler.client.report [None req-dbc2121a-ef71-4ef4-8f94-3a34690382a8 tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] Deleted allocations for instance 9fa0bae6-065d-49ad-b076-0956f1cc3677 [ 524.888458] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bea587a2-9536-4170-8efa-51c241e7d4f9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.898965] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db8a3f23-130a-4e24-8e3f-1d4c07fc19df {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.942815] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-313393ff-b3dd-42c3-8073-c8cf101001f1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.952686] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35dbb386-d4c8-408a-8c4a-cbc361a0ce3c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.973950] env[62066]: DEBUG nova.compute.provider_tree [None req-9838e0c6-e0b2-45d5-9c5b-a0a0b9c60219 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 525.103253] env[62066]: DEBUG oslo_concurrency.lockutils [None req-70656e85-7c6a-408b-a415-8e3cff7746a0 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 525.127579] env[62066]: DEBUG nova.network.neutron [None req-9eabd10d-5afe-4c06-abad-b512a51e6251 tempest-ServersAdminNegativeTestJSON-732915097 tempest-ServersAdminNegativeTestJSON-732915097-project-member] [instance: 21010917-f820-4163-848d-e4141c89b13d] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 525.163080] env[62066]: INFO nova.compute.manager [-] [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] Took 1.03 seconds to deallocate network for instance. [ 525.166122] env[62066]: DEBUG nova.compute.claims [None req-68481586-e4e9-4c46-8b51-aa7845aca824 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] Aborting claim: {{(pid=62066) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 525.166339] env[62066]: DEBUG oslo_concurrency.lockutils [None req-68481586-e4e9-4c46-8b51-aa7845aca824 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 525.248722] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Acquiring lock "5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 525.248856] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Lock "5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 525.291620] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Acquiring lock "72c2dbe0-78de-4a7d-98df-cf405a5f20e2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 525.291620] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Lock "72c2dbe0-78de-4a7d-98df-cf405a5f20e2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 525.294217] env[62066]: DEBUG nova.network.neutron [None req-99d923d5-ef6f-43f2-959c-4b252ac1d398 tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 525.395645] env[62066]: DEBUG oslo_concurrency.lockutils [None req-dbc2121a-ef71-4ef4-8f94-3a34690382a8 tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] Lock "9fa0bae6-065d-49ad-b076-0956f1cc3677" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 32.284s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 525.476442] env[62066]: DEBUG nova.scheduler.client.report [None req-9838e0c6-e0b2-45d5-9c5b-a0a0b9c60219 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 525.568249] env[62066]: DEBUG nova.network.neutron [None req-99d923d5-ef6f-43f2-959c-4b252ac1d398 tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 525.632145] env[62066]: INFO nova.compute.manager [None req-9eabd10d-5afe-4c06-abad-b512a51e6251 tempest-ServersAdminNegativeTestJSON-732915097 tempest-ServersAdminNegativeTestJSON-732915097-project-member] [instance: 21010917-f820-4163-848d-e4141c89b13d] Took 1.03 seconds to deallocate network for instance. [ 525.812494] env[62066]: DEBUG nova.compute.manager [req-ae6a8b0c-06c4-4b0b-bbdc-4190af9c12d3 req-8bae769c-091a-40c5-97fe-3d681dca4425 service nova] [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] Received event network-changed-04cc7ab7-caaa-4afc-afd7-b4230a6e2326 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 525.813023] env[62066]: DEBUG nova.compute.manager [req-ae6a8b0c-06c4-4b0b-bbdc-4190af9c12d3 req-8bae769c-091a-40c5-97fe-3d681dca4425 service nova] [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] Refreshing instance network info cache due to event network-changed-04cc7ab7-caaa-4afc-afd7-b4230a6e2326. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 525.813023] env[62066]: DEBUG oslo_concurrency.lockutils [req-ae6a8b0c-06c4-4b0b-bbdc-4190af9c12d3 req-8bae769c-091a-40c5-97fe-3d681dca4425 service nova] Acquiring lock "refresh_cache-68b65162-23a6-464f-a2f9-1635bea8786f" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 525.898776] env[62066]: DEBUG nova.compute.manager [None req-c4a0feae-1b12-4166-bd78-82a96fe8ac4c tempest-ServersWithSpecificFlavorTestJSON-184391817 tempest-ServersWithSpecificFlavorTestJSON-184391817-project-member] [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 525.983909] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9838e0c6-e0b2-45d5-9c5b-a0a0b9c60219 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.529s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 525.984427] env[62066]: DEBUG nova.compute.manager [None req-9838e0c6-e0b2-45d5-9c5b-a0a0b9c60219 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] [instance: 64f9698e-16f5-447a-914a-9e936e05d101] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 525.988390] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c743a8d-c557-4d89-8fe0-1cc65e783f93 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 10.043s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 526.071552] env[62066]: DEBUG oslo_concurrency.lockutils [None req-99d923d5-ef6f-43f2-959c-4b252ac1d398 tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] Releasing lock "refresh_cache-68b65162-23a6-464f-a2f9-1635bea8786f" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 526.072108] env[62066]: DEBUG nova.compute.manager [None req-99d923d5-ef6f-43f2-959c-4b252ac1d398 tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 526.072318] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-99d923d5-ef6f-43f2-959c-4b252ac1d398 tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 526.072932] env[62066]: DEBUG oslo_concurrency.lockutils [req-ae6a8b0c-06c4-4b0b-bbdc-4190af9c12d3 req-8bae769c-091a-40c5-97fe-3d681dca4425 service nova] Acquired lock "refresh_cache-68b65162-23a6-464f-a2f9-1635bea8786f" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 526.073124] env[62066]: DEBUG nova.network.neutron [req-ae6a8b0c-06c4-4b0b-bbdc-4190af9c12d3 req-8bae769c-091a-40c5-97fe-3d681dca4425 service nova] [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] Refreshing network info cache for port 04cc7ab7-caaa-4afc-afd7-b4230a6e2326 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 526.075501] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-563887f2-3c04-4ae8-b980-d76fdedb51ee {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.089847] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-419d877f-0659-4e45-9733-cda1cd057329 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.113335] env[62066]: WARNING nova.virt.vmwareapi.vmops [None req-99d923d5-ef6f-43f2-959c-4b252ac1d398 tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 68b65162-23a6-464f-a2f9-1635bea8786f could not be found. [ 526.113553] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-99d923d5-ef6f-43f2-959c-4b252ac1d398 tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 526.113888] env[62066]: INFO nova.compute.manager [None req-99d923d5-ef6f-43f2-959c-4b252ac1d398 tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] Took 0.04 seconds to destroy the instance on the hypervisor. [ 526.114031] env[62066]: DEBUG oslo.service.loopingcall [None req-99d923d5-ef6f-43f2-959c-4b252ac1d398 tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 526.114249] env[62066]: DEBUG nova.compute.manager [-] [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 526.114304] env[62066]: DEBUG nova.network.neutron [-] [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 526.132860] env[62066]: DEBUG nova.network.neutron [-] [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 526.333894] env[62066]: DEBUG nova.compute.manager [req-05588cef-afc1-470a-9955-ac0715084982 req-3da80cc3-ddc7-438a-b86b-19f3dffc474f service nova] [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] Received event network-vif-deleted-78745978-2782-4cbd-8fbd-fbdf247cc830 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 526.334116] env[62066]: DEBUG nova.compute.manager [req-05588cef-afc1-470a-9955-ac0715084982 req-3da80cc3-ddc7-438a-b86b-19f3dffc474f service nova] [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] Received event network-changed-5d597d92-a2ed-414a-babb-5eacc2e2e5c2 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 526.335209] env[62066]: DEBUG nova.compute.manager [req-05588cef-afc1-470a-9955-ac0715084982 req-3da80cc3-ddc7-438a-b86b-19f3dffc474f service nova] [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] Refreshing instance network info cache due to event network-changed-5d597d92-a2ed-414a-babb-5eacc2e2e5c2. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 526.335456] env[62066]: DEBUG oslo_concurrency.lockutils [req-05588cef-afc1-470a-9955-ac0715084982 req-3da80cc3-ddc7-438a-b86b-19f3dffc474f service nova] Acquiring lock "refresh_cache-d522ee4a-9bd9-402b-b3bd-c36f93366249" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 526.338204] env[62066]: DEBUG oslo_concurrency.lockutils [req-05588cef-afc1-470a-9955-ac0715084982 req-3da80cc3-ddc7-438a-b86b-19f3dffc474f service nova] Acquired lock "refresh_cache-d522ee4a-9bd9-402b-b3bd-c36f93366249" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 526.338204] env[62066]: DEBUG nova.network.neutron [req-05588cef-afc1-470a-9955-ac0715084982 req-3da80cc3-ddc7-438a-b86b-19f3dffc474f service nova] [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] Refreshing network info cache for port 5d597d92-a2ed-414a-babb-5eacc2e2e5c2 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 526.427049] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c4a0feae-1b12-4166-bd78-82a96fe8ac4c tempest-ServersWithSpecificFlavorTestJSON-184391817 tempest-ServersWithSpecificFlavorTestJSON-184391817-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 526.491427] env[62066]: DEBUG nova.compute.utils [None req-9838e0c6-e0b2-45d5-9c5b-a0a0b9c60219 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 526.496317] env[62066]: DEBUG nova.compute.manager [None req-9838e0c6-e0b2-45d5-9c5b-a0a0b9c60219 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] [instance: 64f9698e-16f5-447a-914a-9e936e05d101] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 526.496317] env[62066]: DEBUG nova.network.neutron [None req-9838e0c6-e0b2-45d5-9c5b-a0a0b9c60219 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] [instance: 64f9698e-16f5-447a-914a-9e936e05d101] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 526.590981] env[62066]: DEBUG nova.policy [None req-9838e0c6-e0b2-45d5-9c5b-a0a0b9c60219 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ef2bdd5efd6b4a4e8287c63007defdc4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a7fb7655514d47fa88c7fccb28665caa', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 526.608477] env[62066]: DEBUG nova.network.neutron [req-ae6a8b0c-06c4-4b0b-bbdc-4190af9c12d3 req-8bae769c-091a-40c5-97fe-3d681dca4425 service nova] [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 526.637255] env[62066]: DEBUG nova.network.neutron [-] [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 526.670063] env[62066]: INFO nova.scheduler.client.report [None req-9eabd10d-5afe-4c06-abad-b512a51e6251 tempest-ServersAdminNegativeTestJSON-732915097 tempest-ServersAdminNegativeTestJSON-732915097-project-member] Deleted allocations for instance 21010917-f820-4163-848d-e4141c89b13d [ 526.745412] env[62066]: DEBUG nova.network.neutron [req-ae6a8b0c-06c4-4b0b-bbdc-4190af9c12d3 req-8bae769c-091a-40c5-97fe-3d681dca4425 service nova] [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 526.873495] env[62066]: DEBUG nova.network.neutron [req-05588cef-afc1-470a-9955-ac0715084982 req-3da80cc3-ddc7-438a-b86b-19f3dffc474f service nova] [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 526.944594] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39b8eb4a-a251-4983-b461-2a1bc6cba54b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.954972] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4bfcdd4-c31e-472e-9398-a919c356fa00 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.988107] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-288b6a8f-be92-45cf-bf5f-e473a5baa444 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.996345] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-985b1e7a-fac5-4321-a5f1-bc0a8eee8e07 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 527.012531] env[62066]: DEBUG nova.compute.manager [None req-9838e0c6-e0b2-45d5-9c5b-a0a0b9c60219 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] [instance: 64f9698e-16f5-447a-914a-9e936e05d101] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 527.015354] env[62066]: DEBUG nova.compute.provider_tree [None req-5c743a8d-c557-4d89-8fe0-1cc65e783f93 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 527.120139] env[62066]: DEBUG nova.network.neutron [None req-9838e0c6-e0b2-45d5-9c5b-a0a0b9c60219 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] [instance: 64f9698e-16f5-447a-914a-9e936e05d101] Successfully created port: 8a431920-fe14-44e1-85f9-31b9af5b10c2 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 527.139856] env[62066]: INFO nova.compute.manager [-] [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] Took 1.03 seconds to deallocate network for instance. [ 527.146432] env[62066]: DEBUG nova.compute.claims [None req-99d923d5-ef6f-43f2-959c-4b252ac1d398 tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] Aborting claim: {{(pid=62066) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 527.146603] env[62066]: DEBUG oslo_concurrency.lockutils [None req-99d923d5-ef6f-43f2-959c-4b252ac1d398 tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 527.150562] env[62066]: DEBUG nova.network.neutron [req-05588cef-afc1-470a-9955-ac0715084982 req-3da80cc3-ddc7-438a-b86b-19f3dffc474f service nova] [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 527.191192] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9eabd10d-5afe-4c06-abad-b512a51e6251 tempest-ServersAdminNegativeTestJSON-732915097 tempest-ServersAdminNegativeTestJSON-732915097-project-member] Lock "21010917-f820-4163-848d-e4141c89b13d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 32.597s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 527.248014] env[62066]: DEBUG oslo_concurrency.lockutils [req-ae6a8b0c-06c4-4b0b-bbdc-4190af9c12d3 req-8bae769c-091a-40c5-97fe-3d681dca4425 service nova] Releasing lock "refresh_cache-68b65162-23a6-464f-a2f9-1635bea8786f" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 527.426861] env[62066]: DEBUG nova.compute.manager [None req-9c869e49-1bde-4e0b-8593-f38429a39437 tempest-ServerDiagnosticsV248Test-1738630200 tempest-ServerDiagnosticsV248Test-1738630200-project-admin] [instance: 4b34ff57-e137-4abc-8aed-76dd2ec8d313] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 527.427659] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd7225c0-4454-4ef2-8d0c-038e075ee355 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 527.436071] env[62066]: INFO nova.compute.manager [None req-9c869e49-1bde-4e0b-8593-f38429a39437 tempest-ServerDiagnosticsV248Test-1738630200 tempest-ServerDiagnosticsV248Test-1738630200-project-admin] [instance: 4b34ff57-e137-4abc-8aed-76dd2ec8d313] Retrieving diagnostics [ 527.436888] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef9a281d-1ee2-4a22-9262-636006d6687c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 527.478853] env[62066]: DEBUG oslo_concurrency.lockutils [None req-14085d2f-7e58-404a-a2d0-4487ac658e45 tempest-FloatingIPsAssociationTestJSON-608653087 tempest-FloatingIPsAssociationTestJSON-608653087-project-member] Acquiring lock "08d41411-7928-4379-9f2e-c6ce00843d82" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 527.479196] env[62066]: DEBUG oslo_concurrency.lockutils [None req-14085d2f-7e58-404a-a2d0-4487ac658e45 tempest-FloatingIPsAssociationTestJSON-608653087 tempest-FloatingIPsAssociationTestJSON-608653087-project-member] Lock "08d41411-7928-4379-9f2e-c6ce00843d82" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 527.521868] env[62066]: DEBUG nova.scheduler.client.report [None req-5c743a8d-c557-4d89-8fe0-1cc65e783f93 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 527.658290] env[62066]: DEBUG oslo_concurrency.lockutils [req-05588cef-afc1-470a-9955-ac0715084982 req-3da80cc3-ddc7-438a-b86b-19f3dffc474f service nova] Releasing lock "refresh_cache-d522ee4a-9bd9-402b-b3bd-c36f93366249" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 527.658807] env[62066]: DEBUG nova.compute.manager [req-05588cef-afc1-470a-9955-ac0715084982 req-3da80cc3-ddc7-438a-b86b-19f3dffc474f service nova] [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] Received event network-vif-deleted-5d597d92-a2ed-414a-babb-5eacc2e2e5c2 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 527.694869] env[62066]: DEBUG nova.compute.manager [None req-cb29a201-afba-4eb8-8336-fe27a681c8f8 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 528.027498] env[62066]: DEBUG nova.compute.manager [None req-9838e0c6-e0b2-45d5-9c5b-a0a0b9c60219 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] [instance: 64f9698e-16f5-447a-914a-9e936e05d101] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 528.030368] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c743a8d-c557-4d89-8fe0-1cc65e783f93 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.042s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 528.031325] env[62066]: ERROR nova.compute.manager [None req-5c743a8d-c557-4d89-8fe0-1cc65e783f93 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 33397b47-1617-4402-98e4-8e41bebcc4a5, please check neutron logs for more information. [ 528.031325] env[62066]: ERROR nova.compute.manager [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] Traceback (most recent call last): [ 528.031325] env[62066]: ERROR nova.compute.manager [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 528.031325] env[62066]: ERROR nova.compute.manager [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] self.driver.spawn(context, instance, image_meta, [ 528.031325] env[62066]: ERROR nova.compute.manager [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 528.031325] env[62066]: ERROR nova.compute.manager [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] self._vmops.spawn(context, instance, image_meta, injected_files, [ 528.031325] env[62066]: ERROR nova.compute.manager [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 528.031325] env[62066]: ERROR nova.compute.manager [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] vm_ref = self.build_virtual_machine(instance, [ 528.031325] env[62066]: ERROR nova.compute.manager [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 528.031325] env[62066]: ERROR nova.compute.manager [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] vif_infos = vmwarevif.get_vif_info(self._session, [ 528.031325] env[62066]: ERROR nova.compute.manager [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 528.031652] env[62066]: ERROR nova.compute.manager [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] for vif in network_info: [ 528.031652] env[62066]: ERROR nova.compute.manager [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 528.031652] env[62066]: ERROR nova.compute.manager [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] return self._sync_wrapper(fn, *args, **kwargs) [ 528.031652] env[62066]: ERROR nova.compute.manager [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 528.031652] env[62066]: ERROR nova.compute.manager [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] self.wait() [ 528.031652] env[62066]: ERROR nova.compute.manager [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 528.031652] env[62066]: ERROR nova.compute.manager [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] self[:] = self._gt.wait() [ 528.031652] env[62066]: ERROR nova.compute.manager [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 528.031652] env[62066]: ERROR nova.compute.manager [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] return self._exit_event.wait() [ 528.031652] env[62066]: ERROR nova.compute.manager [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 528.031652] env[62066]: ERROR nova.compute.manager [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] result = hub.switch() [ 528.031652] env[62066]: ERROR nova.compute.manager [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 528.031652] env[62066]: ERROR nova.compute.manager [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] return self.greenlet.switch() [ 528.031970] env[62066]: ERROR nova.compute.manager [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 528.031970] env[62066]: ERROR nova.compute.manager [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] result = function(*args, **kwargs) [ 528.031970] env[62066]: ERROR nova.compute.manager [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 528.031970] env[62066]: ERROR nova.compute.manager [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] return func(*args, **kwargs) [ 528.031970] env[62066]: ERROR nova.compute.manager [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 528.031970] env[62066]: ERROR nova.compute.manager [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] raise e [ 528.031970] env[62066]: ERROR nova.compute.manager [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 528.031970] env[62066]: ERROR nova.compute.manager [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] nwinfo = self.network_api.allocate_for_instance( [ 528.031970] env[62066]: ERROR nova.compute.manager [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 528.031970] env[62066]: ERROR nova.compute.manager [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] created_port_ids = self._update_ports_for_instance( [ 528.031970] env[62066]: ERROR nova.compute.manager [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 528.031970] env[62066]: ERROR nova.compute.manager [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] with excutils.save_and_reraise_exception(): [ 528.031970] env[62066]: ERROR nova.compute.manager [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 528.032330] env[62066]: ERROR nova.compute.manager [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] self.force_reraise() [ 528.032330] env[62066]: ERROR nova.compute.manager [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 528.032330] env[62066]: ERROR nova.compute.manager [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] raise self.value [ 528.032330] env[62066]: ERROR nova.compute.manager [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 528.032330] env[62066]: ERROR nova.compute.manager [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] updated_port = self._update_port( [ 528.032330] env[62066]: ERROR nova.compute.manager [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 528.032330] env[62066]: ERROR nova.compute.manager [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] _ensure_no_port_binding_failure(port) [ 528.032330] env[62066]: ERROR nova.compute.manager [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 528.032330] env[62066]: ERROR nova.compute.manager [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] raise exception.PortBindingFailed(port_id=port['id']) [ 528.032330] env[62066]: ERROR nova.compute.manager [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] nova.exception.PortBindingFailed: Binding failed for port 33397b47-1617-4402-98e4-8e41bebcc4a5, please check neutron logs for more information. [ 528.032330] env[62066]: ERROR nova.compute.manager [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] [ 528.032769] env[62066]: DEBUG nova.compute.utils [None req-5c743a8d-c557-4d89-8fe0-1cc65e783f93 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] Binding failed for port 33397b47-1617-4402-98e4-8e41bebcc4a5, please check neutron logs for more information. {{(pid=62066) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 528.033876] env[62066]: DEBUG oslo_concurrency.lockutils [None req-67ee84a7-af8a-43fc-bff6-48212ddfc8ae tempest-ServersNegativeTestMultiTenantJSON-1325431503 tempest-ServersNegativeTestMultiTenantJSON-1325431503-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.760s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 528.037214] env[62066]: INFO nova.compute.claims [None req-67ee84a7-af8a-43fc-bff6-48212ddfc8ae tempest-ServersNegativeTestMultiTenantJSON-1325431503 tempest-ServersNegativeTestMultiTenantJSON-1325431503-project-member] [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 528.041190] env[62066]: DEBUG nova.compute.manager [None req-5c743a8d-c557-4d89-8fe0-1cc65e783f93 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] Build of instance e8693f3d-ce6e-4c94-a719-4e6723c16de8 was re-scheduled: Binding failed for port 33397b47-1617-4402-98e4-8e41bebcc4a5, please check neutron logs for more information. {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 528.041190] env[62066]: DEBUG nova.compute.manager [None req-5c743a8d-c557-4d89-8fe0-1cc65e783f93 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] Unplugging VIFs for instance {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 528.041190] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c743a8d-c557-4d89-8fe0-1cc65e783f93 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] Acquiring lock "refresh_cache-e8693f3d-ce6e-4c94-a719-4e6723c16de8" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 528.041190] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c743a8d-c557-4d89-8fe0-1cc65e783f93 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] Acquired lock "refresh_cache-e8693f3d-ce6e-4c94-a719-4e6723c16de8" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 528.041476] env[62066]: DEBUG nova.network.neutron [None req-5c743a8d-c557-4d89-8fe0-1cc65e783f93 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 528.066043] env[62066]: DEBUG nova.virt.hardware [None req-9838e0c6-e0b2-45d5-9c5b-a0a0b9c60219 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-23T13:40:41Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-23T13:40:23Z,direct_url=,disk_format='vmdk',id=50ff584c-3b50-4395-af07-3e66769bc9f7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='eb52f7069a374c61ae946f052007c6d9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-23T13:40:24Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 528.070086] env[62066]: DEBUG nova.virt.hardware [None req-9838e0c6-e0b2-45d5-9c5b-a0a0b9c60219 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 528.070086] env[62066]: DEBUG nova.virt.hardware [None req-9838e0c6-e0b2-45d5-9c5b-a0a0b9c60219 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 528.070086] env[62066]: DEBUG nova.virt.hardware [None req-9838e0c6-e0b2-45d5-9c5b-a0a0b9c60219 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 528.070086] env[62066]: DEBUG nova.virt.hardware [None req-9838e0c6-e0b2-45d5-9c5b-a0a0b9c60219 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 528.070265] env[62066]: DEBUG nova.virt.hardware [None req-9838e0c6-e0b2-45d5-9c5b-a0a0b9c60219 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 528.070512] env[62066]: DEBUG nova.virt.hardware [None req-9838e0c6-e0b2-45d5-9c5b-a0a0b9c60219 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 528.070698] env[62066]: DEBUG nova.virt.hardware [None req-9838e0c6-e0b2-45d5-9c5b-a0a0b9c60219 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 528.070896] env[62066]: DEBUG nova.virt.hardware [None req-9838e0c6-e0b2-45d5-9c5b-a0a0b9c60219 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 528.071103] env[62066]: DEBUG nova.virt.hardware [None req-9838e0c6-e0b2-45d5-9c5b-a0a0b9c60219 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 528.071820] env[62066]: DEBUG nova.virt.hardware [None req-9838e0c6-e0b2-45d5-9c5b-a0a0b9c60219 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 528.076257] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cb11b3a-fcdb-4597-bf18-c971f8bf66c3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.087882] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-944ad562-6656-4d75-9f42-a1c72fc0c7d2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.227858] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cb29a201-afba-4eb8-8336-fe27a681c8f8 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 528.577734] env[62066]: DEBUG nova.network.neutron [None req-5c743a8d-c557-4d89-8fe0-1cc65e783f93 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 528.625501] env[62066]: INFO nova.compute.manager [None req-a16a38bd-97b1-4d18-bd95-9847ae1f8463 tempest-ServersAdmin275Test-1729403672 tempest-ServersAdmin275Test-1729403672-project-admin] [instance: 56eee633-2c01-4df2-8d2e-58ad712942f6] Rebuilding instance [ 528.681440] env[62066]: DEBUG nova.compute.manager [None req-a16a38bd-97b1-4d18-bd95-9847ae1f8463 tempest-ServersAdmin275Test-1729403672 tempest-ServersAdmin275Test-1729403672-project-admin] [instance: 56eee633-2c01-4df2-8d2e-58ad712942f6] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 528.684274] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c02e7842-bb39-4793-abc0-2c5b2df2454f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.715541] env[62066]: DEBUG nova.network.neutron [None req-5c743a8d-c557-4d89-8fe0-1cc65e783f93 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 529.199167] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-a16a38bd-97b1-4d18-bd95-9847ae1f8463 tempest-ServersAdmin275Test-1729403672 tempest-ServersAdmin275Test-1729403672-project-admin] [instance: 56eee633-2c01-4df2-8d2e-58ad712942f6] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 529.201716] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-db4e8f05-8155-4fc4-8317-afc91d18e519 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.210988] env[62066]: DEBUG oslo_vmware.api [None req-a16a38bd-97b1-4d18-bd95-9847ae1f8463 tempest-ServersAdmin275Test-1729403672 tempest-ServersAdmin275Test-1729403672-project-admin] Waiting for the task: (returnval){ [ 529.210988] env[62066]: value = "task-1340628" [ 529.210988] env[62066]: _type = "Task" [ 529.210988] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 529.219811] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c743a8d-c557-4d89-8fe0-1cc65e783f93 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] Releasing lock "refresh_cache-e8693f3d-ce6e-4c94-a719-4e6723c16de8" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 529.220055] env[62066]: DEBUG nova.compute.manager [None req-5c743a8d-c557-4d89-8fe0-1cc65e783f93 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 529.220265] env[62066]: DEBUG nova.compute.manager [None req-5c743a8d-c557-4d89-8fe0-1cc65e783f93 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 529.220473] env[62066]: DEBUG nova.network.neutron [None req-5c743a8d-c557-4d89-8fe0-1cc65e783f93 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 529.232293] env[62066]: DEBUG oslo_vmware.api [None req-a16a38bd-97b1-4d18-bd95-9847ae1f8463 tempest-ServersAdmin275Test-1729403672 tempest-ServersAdmin275Test-1729403672-project-admin] Task: {'id': task-1340628, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 529.260638] env[62066]: DEBUG nova.network.neutron [None req-5c743a8d-c557-4d89-8fe0-1cc65e783f93 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 529.489375] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00f183c3-b1b2-4e5d-b893-356238454dfd {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.497090] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db99968d-b763-4cf5-ad63-e9b5e14c7ec8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.528378] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bffa15b7-54bf-49a2-9d34-f61578748c0b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.536882] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1eb2f0d4-0ee7-4a35-b988-d478ee319b29 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.556275] env[62066]: DEBUG nova.compute.provider_tree [None req-67ee84a7-af8a-43fc-bff6-48212ddfc8ae tempest-ServersNegativeTestMultiTenantJSON-1325431503 tempest-ServersNegativeTestMultiTenantJSON-1325431503-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 529.725251] env[62066]: DEBUG oslo_vmware.api [None req-a16a38bd-97b1-4d18-bd95-9847ae1f8463 tempest-ServersAdmin275Test-1729403672 tempest-ServersAdmin275Test-1729403672-project-admin] Task: {'id': task-1340628, 'name': PowerOffVM_Task, 'duration_secs': 0.134923} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 529.725251] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-a16a38bd-97b1-4d18-bd95-9847ae1f8463 tempest-ServersAdmin275Test-1729403672 tempest-ServersAdmin275Test-1729403672-project-admin] [instance: 56eee633-2c01-4df2-8d2e-58ad712942f6] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 529.725251] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-a16a38bd-97b1-4d18-bd95-9847ae1f8463 tempest-ServersAdmin275Test-1729403672 tempest-ServersAdmin275Test-1729403672-project-admin] [instance: 56eee633-2c01-4df2-8d2e-58ad712942f6] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 529.726594] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df19e3e9-acaa-437e-a423-f6d58a9aa48b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.736272] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-a16a38bd-97b1-4d18-bd95-9847ae1f8463 tempest-ServersAdmin275Test-1729403672 tempest-ServersAdmin275Test-1729403672-project-admin] [instance: 56eee633-2c01-4df2-8d2e-58ad712942f6] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 529.736554] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e39f66b8-3a5e-4f71-a42a-e06b2080a12c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.763514] env[62066]: DEBUG nova.network.neutron [None req-5c743a8d-c557-4d89-8fe0-1cc65e783f93 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 529.766821] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-a16a38bd-97b1-4d18-bd95-9847ae1f8463 tempest-ServersAdmin275Test-1729403672 tempest-ServersAdmin275Test-1729403672-project-admin] [instance: 56eee633-2c01-4df2-8d2e-58ad712942f6] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 529.767097] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-a16a38bd-97b1-4d18-bd95-9847ae1f8463 tempest-ServersAdmin275Test-1729403672 tempest-ServersAdmin275Test-1729403672-project-admin] [instance: 56eee633-2c01-4df2-8d2e-58ad712942f6] Deleting contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 529.767567] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-a16a38bd-97b1-4d18-bd95-9847ae1f8463 tempest-ServersAdmin275Test-1729403672 tempest-ServersAdmin275Test-1729403672-project-admin] Deleting the datastore file [datastore2] 56eee633-2c01-4df2-8d2e-58ad712942f6 {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 529.767827] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c941da7b-d401-476a-9413-948846339392 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.776801] env[62066]: DEBUG oslo_vmware.api [None req-a16a38bd-97b1-4d18-bd95-9847ae1f8463 tempest-ServersAdmin275Test-1729403672 tempest-ServersAdmin275Test-1729403672-project-admin] Waiting for the task: (returnval){ [ 529.776801] env[62066]: value = "task-1340630" [ 529.776801] env[62066]: _type = "Task" [ 529.776801] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 529.787728] env[62066]: DEBUG oslo_vmware.api [None req-a16a38bd-97b1-4d18-bd95-9847ae1f8463 tempest-ServersAdmin275Test-1729403672 tempest-ServersAdmin275Test-1729403672-project-admin] Task: {'id': task-1340630, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 529.824817] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a9ebb4d2-ff8d-4333-a9df-86618ea2d54c tempest-ImagesOneServerTestJSON-1591819949 tempest-ImagesOneServerTestJSON-1591819949-project-member] Acquiring lock "623300f7-54d2-4b0a-b356-a1ae585682cb" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 529.825078] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a9ebb4d2-ff8d-4333-a9df-86618ea2d54c tempest-ImagesOneServerTestJSON-1591819949 tempest-ImagesOneServerTestJSON-1591819949-project-member] Lock "623300f7-54d2-4b0a-b356-a1ae585682cb" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 529.869957] env[62066]: ERROR nova.compute.manager [None req-9838e0c6-e0b2-45d5-9c5b-a0a0b9c60219 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 8a431920-fe14-44e1-85f9-31b9af5b10c2, please check neutron logs for more information. [ 529.869957] env[62066]: ERROR nova.compute.manager Traceback (most recent call last): [ 529.869957] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 529.869957] env[62066]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 529.869957] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 529.869957] env[62066]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 529.869957] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 529.869957] env[62066]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 529.869957] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 529.869957] env[62066]: ERROR nova.compute.manager self.force_reraise() [ 529.869957] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 529.869957] env[62066]: ERROR nova.compute.manager raise self.value [ 529.869957] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 529.869957] env[62066]: ERROR nova.compute.manager updated_port = self._update_port( [ 529.869957] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 529.869957] env[62066]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 529.870462] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 529.870462] env[62066]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 529.870462] env[62066]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 8a431920-fe14-44e1-85f9-31b9af5b10c2, please check neutron logs for more information. [ 529.870462] env[62066]: ERROR nova.compute.manager [ 529.870462] env[62066]: Traceback (most recent call last): [ 529.870462] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 529.870462] env[62066]: listener.cb(fileno) [ 529.870462] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 529.870462] env[62066]: result = function(*args, **kwargs) [ 529.870462] env[62066]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 529.870462] env[62066]: return func(*args, **kwargs) [ 529.870462] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 529.870462] env[62066]: raise e [ 529.870462] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 529.870462] env[62066]: nwinfo = self.network_api.allocate_for_instance( [ 529.870462] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 529.870462] env[62066]: created_port_ids = self._update_ports_for_instance( [ 529.870462] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 529.870462] env[62066]: with excutils.save_and_reraise_exception(): [ 529.870462] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 529.870462] env[62066]: self.force_reraise() [ 529.870462] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 529.870462] env[62066]: raise self.value [ 529.870462] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 529.870462] env[62066]: updated_port = self._update_port( [ 529.870462] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 529.870462] env[62066]: _ensure_no_port_binding_failure(port) [ 529.870462] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 529.870462] env[62066]: raise exception.PortBindingFailed(port_id=port['id']) [ 529.871256] env[62066]: nova.exception.PortBindingFailed: Binding failed for port 8a431920-fe14-44e1-85f9-31b9af5b10c2, please check neutron logs for more information. [ 529.871256] env[62066]: Removing descriptor: 20 [ 529.871728] env[62066]: ERROR nova.compute.manager [None req-9838e0c6-e0b2-45d5-9c5b-a0a0b9c60219 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] [instance: 64f9698e-16f5-447a-914a-9e936e05d101] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 8a431920-fe14-44e1-85f9-31b9af5b10c2, please check neutron logs for more information. [ 529.871728] env[62066]: ERROR nova.compute.manager [instance: 64f9698e-16f5-447a-914a-9e936e05d101] Traceback (most recent call last): [ 529.871728] env[62066]: ERROR nova.compute.manager [instance: 64f9698e-16f5-447a-914a-9e936e05d101] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 529.871728] env[62066]: ERROR nova.compute.manager [instance: 64f9698e-16f5-447a-914a-9e936e05d101] yield resources [ 529.871728] env[62066]: ERROR nova.compute.manager [instance: 64f9698e-16f5-447a-914a-9e936e05d101] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 529.871728] env[62066]: ERROR nova.compute.manager [instance: 64f9698e-16f5-447a-914a-9e936e05d101] self.driver.spawn(context, instance, image_meta, [ 529.871728] env[62066]: ERROR nova.compute.manager [instance: 64f9698e-16f5-447a-914a-9e936e05d101] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 529.871728] env[62066]: ERROR nova.compute.manager [instance: 64f9698e-16f5-447a-914a-9e936e05d101] self._vmops.spawn(context, instance, image_meta, injected_files, [ 529.871728] env[62066]: ERROR nova.compute.manager [instance: 64f9698e-16f5-447a-914a-9e936e05d101] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 529.871728] env[62066]: ERROR nova.compute.manager [instance: 64f9698e-16f5-447a-914a-9e936e05d101] vm_ref = self.build_virtual_machine(instance, [ 529.871728] env[62066]: ERROR nova.compute.manager [instance: 64f9698e-16f5-447a-914a-9e936e05d101] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 529.872146] env[62066]: ERROR nova.compute.manager [instance: 64f9698e-16f5-447a-914a-9e936e05d101] vif_infos = vmwarevif.get_vif_info(self._session, [ 529.872146] env[62066]: ERROR nova.compute.manager [instance: 64f9698e-16f5-447a-914a-9e936e05d101] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 529.872146] env[62066]: ERROR nova.compute.manager [instance: 64f9698e-16f5-447a-914a-9e936e05d101] for vif in network_info: [ 529.872146] env[62066]: ERROR nova.compute.manager [instance: 64f9698e-16f5-447a-914a-9e936e05d101] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 529.872146] env[62066]: ERROR nova.compute.manager [instance: 64f9698e-16f5-447a-914a-9e936e05d101] return self._sync_wrapper(fn, *args, **kwargs) [ 529.872146] env[62066]: ERROR nova.compute.manager [instance: 64f9698e-16f5-447a-914a-9e936e05d101] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 529.872146] env[62066]: ERROR nova.compute.manager [instance: 64f9698e-16f5-447a-914a-9e936e05d101] self.wait() [ 529.872146] env[62066]: ERROR nova.compute.manager [instance: 64f9698e-16f5-447a-914a-9e936e05d101] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 529.872146] env[62066]: ERROR nova.compute.manager [instance: 64f9698e-16f5-447a-914a-9e936e05d101] self[:] = self._gt.wait() [ 529.872146] env[62066]: ERROR nova.compute.manager [instance: 64f9698e-16f5-447a-914a-9e936e05d101] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 529.872146] env[62066]: ERROR nova.compute.manager [instance: 64f9698e-16f5-447a-914a-9e936e05d101] return self._exit_event.wait() [ 529.872146] env[62066]: ERROR nova.compute.manager [instance: 64f9698e-16f5-447a-914a-9e936e05d101] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 529.872146] env[62066]: ERROR nova.compute.manager [instance: 64f9698e-16f5-447a-914a-9e936e05d101] result = hub.switch() [ 529.873521] env[62066]: ERROR nova.compute.manager [instance: 64f9698e-16f5-447a-914a-9e936e05d101] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 529.873521] env[62066]: ERROR nova.compute.manager [instance: 64f9698e-16f5-447a-914a-9e936e05d101] return self.greenlet.switch() [ 529.873521] env[62066]: ERROR nova.compute.manager [instance: 64f9698e-16f5-447a-914a-9e936e05d101] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 529.873521] env[62066]: ERROR nova.compute.manager [instance: 64f9698e-16f5-447a-914a-9e936e05d101] result = function(*args, **kwargs) [ 529.873521] env[62066]: ERROR nova.compute.manager [instance: 64f9698e-16f5-447a-914a-9e936e05d101] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 529.873521] env[62066]: ERROR nova.compute.manager [instance: 64f9698e-16f5-447a-914a-9e936e05d101] return func(*args, **kwargs) [ 529.873521] env[62066]: ERROR nova.compute.manager [instance: 64f9698e-16f5-447a-914a-9e936e05d101] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 529.873521] env[62066]: ERROR nova.compute.manager [instance: 64f9698e-16f5-447a-914a-9e936e05d101] raise e [ 529.873521] env[62066]: ERROR nova.compute.manager [instance: 64f9698e-16f5-447a-914a-9e936e05d101] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 529.873521] env[62066]: ERROR nova.compute.manager [instance: 64f9698e-16f5-447a-914a-9e936e05d101] nwinfo = self.network_api.allocate_for_instance( [ 529.873521] env[62066]: ERROR nova.compute.manager [instance: 64f9698e-16f5-447a-914a-9e936e05d101] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 529.873521] env[62066]: ERROR nova.compute.manager [instance: 64f9698e-16f5-447a-914a-9e936e05d101] created_port_ids = self._update_ports_for_instance( [ 529.873521] env[62066]: ERROR nova.compute.manager [instance: 64f9698e-16f5-447a-914a-9e936e05d101] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 529.874289] env[62066]: ERROR nova.compute.manager [instance: 64f9698e-16f5-447a-914a-9e936e05d101] with excutils.save_and_reraise_exception(): [ 529.874289] env[62066]: ERROR nova.compute.manager [instance: 64f9698e-16f5-447a-914a-9e936e05d101] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 529.874289] env[62066]: ERROR nova.compute.manager [instance: 64f9698e-16f5-447a-914a-9e936e05d101] self.force_reraise() [ 529.874289] env[62066]: ERROR nova.compute.manager [instance: 64f9698e-16f5-447a-914a-9e936e05d101] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 529.874289] env[62066]: ERROR nova.compute.manager [instance: 64f9698e-16f5-447a-914a-9e936e05d101] raise self.value [ 529.874289] env[62066]: ERROR nova.compute.manager [instance: 64f9698e-16f5-447a-914a-9e936e05d101] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 529.874289] env[62066]: ERROR nova.compute.manager [instance: 64f9698e-16f5-447a-914a-9e936e05d101] updated_port = self._update_port( [ 529.874289] env[62066]: ERROR nova.compute.manager [instance: 64f9698e-16f5-447a-914a-9e936e05d101] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 529.874289] env[62066]: ERROR nova.compute.manager [instance: 64f9698e-16f5-447a-914a-9e936e05d101] _ensure_no_port_binding_failure(port) [ 529.874289] env[62066]: ERROR nova.compute.manager [instance: 64f9698e-16f5-447a-914a-9e936e05d101] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 529.874289] env[62066]: ERROR nova.compute.manager [instance: 64f9698e-16f5-447a-914a-9e936e05d101] raise exception.PortBindingFailed(port_id=port['id']) [ 529.874289] env[62066]: ERROR nova.compute.manager [instance: 64f9698e-16f5-447a-914a-9e936e05d101] nova.exception.PortBindingFailed: Binding failed for port 8a431920-fe14-44e1-85f9-31b9af5b10c2, please check neutron logs for more information. [ 529.874289] env[62066]: ERROR nova.compute.manager [instance: 64f9698e-16f5-447a-914a-9e936e05d101] [ 529.874637] env[62066]: INFO nova.compute.manager [None req-9838e0c6-e0b2-45d5-9c5b-a0a0b9c60219 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] [instance: 64f9698e-16f5-447a-914a-9e936e05d101] Terminating instance [ 529.874637] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9838e0c6-e0b2-45d5-9c5b-a0a0b9c60219 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] Acquiring lock "refresh_cache-64f9698e-16f5-447a-914a-9e936e05d101" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 529.874711] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9838e0c6-e0b2-45d5-9c5b-a0a0b9c60219 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] Acquired lock "refresh_cache-64f9698e-16f5-447a-914a-9e936e05d101" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 529.875210] env[62066]: DEBUG nova.network.neutron [None req-9838e0c6-e0b2-45d5-9c5b-a0a0b9c60219 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] [instance: 64f9698e-16f5-447a-914a-9e936e05d101] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 529.908366] env[62066]: DEBUG nova.compute.manager [req-4e1e6dee-0269-48be-904c-739f4025e46f req-dc9b91e9-691e-47ac-ab29-52533f4444b6 service nova] [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] Received event network-vif-deleted-04cc7ab7-caaa-4afc-afd7-b4230a6e2326 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 530.055655] env[62066]: DEBUG nova.scheduler.client.report [None req-67ee84a7-af8a-43fc-bff6-48212ddfc8ae tempest-ServersNegativeTestMultiTenantJSON-1325431503 tempest-ServersNegativeTestMultiTenantJSON-1325431503-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 530.269576] env[62066]: INFO nova.compute.manager [None req-5c743a8d-c557-4d89-8fe0-1cc65e783f93 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] [instance: e8693f3d-ce6e-4c94-a719-4e6723c16de8] Took 1.05 seconds to deallocate network for instance. [ 530.288787] env[62066]: DEBUG oslo_vmware.api [None req-a16a38bd-97b1-4d18-bd95-9847ae1f8463 tempest-ServersAdmin275Test-1729403672 tempest-ServersAdmin275Test-1729403672-project-admin] Task: {'id': task-1340630, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.101624} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 530.289722] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-a16a38bd-97b1-4d18-bd95-9847ae1f8463 tempest-ServersAdmin275Test-1729403672 tempest-ServersAdmin275Test-1729403672-project-admin] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 530.290111] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-a16a38bd-97b1-4d18-bd95-9847ae1f8463 tempest-ServersAdmin275Test-1729403672 tempest-ServersAdmin275Test-1729403672-project-admin] [instance: 56eee633-2c01-4df2-8d2e-58ad712942f6] Deleted contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 530.290474] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-a16a38bd-97b1-4d18-bd95-9847ae1f8463 tempest-ServersAdmin275Test-1729403672 tempest-ServersAdmin275Test-1729403672-project-admin] [instance: 56eee633-2c01-4df2-8d2e-58ad712942f6] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 530.398580] env[62066]: DEBUG nova.network.neutron [None req-9838e0c6-e0b2-45d5-9c5b-a0a0b9c60219 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] [instance: 64f9698e-16f5-447a-914a-9e936e05d101] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 530.540304] env[62066]: DEBUG oslo_concurrency.lockutils [None req-659c7f6d-78fb-4b73-9024-3866f9052a37 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] Acquiring lock "4b34ff57-e137-4abc-8aed-76dd2ec8d313" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 530.540597] env[62066]: DEBUG oslo_concurrency.lockutils [None req-659c7f6d-78fb-4b73-9024-3866f9052a37 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] Lock "4b34ff57-e137-4abc-8aed-76dd2ec8d313" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 530.540818] env[62066]: DEBUG oslo_concurrency.lockutils [None req-659c7f6d-78fb-4b73-9024-3866f9052a37 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] Acquiring lock "4b34ff57-e137-4abc-8aed-76dd2ec8d313-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 530.541036] env[62066]: DEBUG oslo_concurrency.lockutils [None req-659c7f6d-78fb-4b73-9024-3866f9052a37 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] Lock "4b34ff57-e137-4abc-8aed-76dd2ec8d313-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 530.541355] env[62066]: DEBUG oslo_concurrency.lockutils [None req-659c7f6d-78fb-4b73-9024-3866f9052a37 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] Lock "4b34ff57-e137-4abc-8aed-76dd2ec8d313-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 530.544788] env[62066]: INFO nova.compute.manager [None req-659c7f6d-78fb-4b73-9024-3866f9052a37 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] [instance: 4b34ff57-e137-4abc-8aed-76dd2ec8d313] Terminating instance [ 530.546666] env[62066]: DEBUG oslo_concurrency.lockutils [None req-659c7f6d-78fb-4b73-9024-3866f9052a37 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] Acquiring lock "refresh_cache-4b34ff57-e137-4abc-8aed-76dd2ec8d313" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 530.546828] env[62066]: DEBUG oslo_concurrency.lockutils [None req-659c7f6d-78fb-4b73-9024-3866f9052a37 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] Acquired lock "refresh_cache-4b34ff57-e137-4abc-8aed-76dd2ec8d313" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 530.546991] env[62066]: DEBUG nova.network.neutron [None req-659c7f6d-78fb-4b73-9024-3866f9052a37 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] [instance: 4b34ff57-e137-4abc-8aed-76dd2ec8d313] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 530.556741] env[62066]: DEBUG nova.network.neutron [None req-9838e0c6-e0b2-45d5-9c5b-a0a0b9c60219 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] [instance: 64f9698e-16f5-447a-914a-9e936e05d101] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 530.562398] env[62066]: DEBUG oslo_concurrency.lockutils [None req-67ee84a7-af8a-43fc-bff6-48212ddfc8ae tempest-ServersNegativeTestMultiTenantJSON-1325431503 tempest-ServersNegativeTestMultiTenantJSON-1325431503-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.529s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 530.562862] env[62066]: DEBUG nova.compute.manager [None req-67ee84a7-af8a-43fc-bff6-48212ddfc8ae tempest-ServersNegativeTestMultiTenantJSON-1325431503 tempest-ServersNegativeTestMultiTenantJSON-1325431503-project-member] [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 530.565151] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 9.116s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 531.063810] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9838e0c6-e0b2-45d5-9c5b-a0a0b9c60219 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] Releasing lock "refresh_cache-64f9698e-16f5-447a-914a-9e936e05d101" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 531.063810] env[62066]: DEBUG nova.compute.manager [None req-9838e0c6-e0b2-45d5-9c5b-a0a0b9c60219 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] [instance: 64f9698e-16f5-447a-914a-9e936e05d101] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 531.063810] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-9838e0c6-e0b2-45d5-9c5b-a0a0b9c60219 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] [instance: 64f9698e-16f5-447a-914a-9e936e05d101] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 531.065742] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-fdeb7493-0860-4a2b-b6d0-8fc4a612d930 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.069326] env[62066]: DEBUG nova.compute.utils [None req-67ee84a7-af8a-43fc-bff6-48212ddfc8ae tempest-ServersNegativeTestMultiTenantJSON-1325431503 tempest-ServersNegativeTestMultiTenantJSON-1325431503-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 531.079609] env[62066]: DEBUG nova.compute.manager [None req-67ee84a7-af8a-43fc-bff6-48212ddfc8ae tempest-ServersNegativeTestMultiTenantJSON-1325431503 tempest-ServersNegativeTestMultiTenantJSON-1325431503-project-member] [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 531.079609] env[62066]: DEBUG nova.network.neutron [None req-67ee84a7-af8a-43fc-bff6-48212ddfc8ae tempest-ServersNegativeTestMultiTenantJSON-1325431503 tempest-ServersNegativeTestMultiTenantJSON-1325431503-project-member] [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 531.081113] env[62066]: DEBUG nova.network.neutron [None req-659c7f6d-78fb-4b73-9024-3866f9052a37 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] [instance: 4b34ff57-e137-4abc-8aed-76dd2ec8d313] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 531.088018] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea4e1f83-bedb-4746-8896-9807352e8cac {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.112212] env[62066]: WARNING nova.virt.vmwareapi.vmops [None req-9838e0c6-e0b2-45d5-9c5b-a0a0b9c60219 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] [instance: 64f9698e-16f5-447a-914a-9e936e05d101] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 64f9698e-16f5-447a-914a-9e936e05d101 could not be found. [ 531.112453] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-9838e0c6-e0b2-45d5-9c5b-a0a0b9c60219 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] [instance: 64f9698e-16f5-447a-914a-9e936e05d101] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 531.112645] env[62066]: INFO nova.compute.manager [None req-9838e0c6-e0b2-45d5-9c5b-a0a0b9c60219 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] [instance: 64f9698e-16f5-447a-914a-9e936e05d101] Took 0.05 seconds to destroy the instance on the hypervisor. [ 531.112886] env[62066]: DEBUG oslo.service.loopingcall [None req-9838e0c6-e0b2-45d5-9c5b-a0a0b9c60219 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 531.113126] env[62066]: DEBUG nova.compute.manager [-] [instance: 64f9698e-16f5-447a-914a-9e936e05d101] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 531.113260] env[62066]: DEBUG nova.network.neutron [-] [instance: 64f9698e-16f5-447a-914a-9e936e05d101] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 531.142285] env[62066]: DEBUG nova.network.neutron [-] [instance: 64f9698e-16f5-447a-914a-9e936e05d101] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 531.193468] env[62066]: DEBUG nova.policy [None req-67ee84a7-af8a-43fc-bff6-48212ddfc8ae tempest-ServersNegativeTestMultiTenantJSON-1325431503 tempest-ServersNegativeTestMultiTenantJSON-1325431503-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c310e3d424ef4e189341d281c67ba71f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '952aba5d63eb4c548447780336a87b4f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 531.308761] env[62066]: INFO nova.scheduler.client.report [None req-5c743a8d-c557-4d89-8fe0-1cc65e783f93 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] Deleted allocations for instance e8693f3d-ce6e-4c94-a719-4e6723c16de8 [ 531.332384] env[62066]: DEBUG nova.network.neutron [None req-659c7f6d-78fb-4b73-9024-3866f9052a37 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] [instance: 4b34ff57-e137-4abc-8aed-76dd2ec8d313] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 531.339216] env[62066]: DEBUG nova.virt.hardware [None req-a16a38bd-97b1-4d18-bd95-9847ae1f8463 tempest-ServersAdmin275Test-1729403672 tempest-ServersAdmin275Test-1729403672-project-admin] Getting desirable topologies for flavor Flavor(created_at=2024-10-23T13:40:41Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-23T13:40:23Z,direct_url=,disk_format='vmdk',id=50ff584c-3b50-4395-af07-3e66769bc9f7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='eb52f7069a374c61ae946f052007c6d9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-23T13:40:24Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 531.339661] env[62066]: DEBUG nova.virt.hardware [None req-a16a38bd-97b1-4d18-bd95-9847ae1f8463 tempest-ServersAdmin275Test-1729403672 tempest-ServersAdmin275Test-1729403672-project-admin] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 531.340042] env[62066]: DEBUG nova.virt.hardware [None req-a16a38bd-97b1-4d18-bd95-9847ae1f8463 tempest-ServersAdmin275Test-1729403672 tempest-ServersAdmin275Test-1729403672-project-admin] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 531.340707] env[62066]: DEBUG nova.virt.hardware [None req-a16a38bd-97b1-4d18-bd95-9847ae1f8463 tempest-ServersAdmin275Test-1729403672 tempest-ServersAdmin275Test-1729403672-project-admin] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 531.340707] env[62066]: DEBUG nova.virt.hardware [None req-a16a38bd-97b1-4d18-bd95-9847ae1f8463 tempest-ServersAdmin275Test-1729403672 tempest-ServersAdmin275Test-1729403672-project-admin] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 531.340831] env[62066]: DEBUG nova.virt.hardware [None req-a16a38bd-97b1-4d18-bd95-9847ae1f8463 tempest-ServersAdmin275Test-1729403672 tempest-ServersAdmin275Test-1729403672-project-admin] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 531.341277] env[62066]: DEBUG nova.virt.hardware [None req-a16a38bd-97b1-4d18-bd95-9847ae1f8463 tempest-ServersAdmin275Test-1729403672 tempest-ServersAdmin275Test-1729403672-project-admin] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 531.341428] env[62066]: DEBUG nova.virt.hardware [None req-a16a38bd-97b1-4d18-bd95-9847ae1f8463 tempest-ServersAdmin275Test-1729403672 tempest-ServersAdmin275Test-1729403672-project-admin] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 531.341783] env[62066]: DEBUG nova.virt.hardware [None req-a16a38bd-97b1-4d18-bd95-9847ae1f8463 tempest-ServersAdmin275Test-1729403672 tempest-ServersAdmin275Test-1729403672-project-admin] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 531.341783] env[62066]: DEBUG nova.virt.hardware [None req-a16a38bd-97b1-4d18-bd95-9847ae1f8463 tempest-ServersAdmin275Test-1729403672 tempest-ServersAdmin275Test-1729403672-project-admin] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 531.342059] env[62066]: DEBUG nova.virt.hardware [None req-a16a38bd-97b1-4d18-bd95-9847ae1f8463 tempest-ServersAdmin275Test-1729403672 tempest-ServersAdmin275Test-1729403672-project-admin] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 531.342909] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-052039fd-ee36-4dda-b22b-2b4a15b89c06 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.353312] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3a67ffe-7437-4d39-95c8-b7fd522dad81 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.369704] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-a16a38bd-97b1-4d18-bd95-9847ae1f8463 tempest-ServersAdmin275Test-1729403672 tempest-ServersAdmin275Test-1729403672-project-admin] [instance: 56eee633-2c01-4df2-8d2e-58ad712942f6] Instance VIF info [] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 531.376316] env[62066]: DEBUG oslo.service.loopingcall [None req-a16a38bd-97b1-4d18-bd95-9847ae1f8463 tempest-ServersAdmin275Test-1729403672 tempest-ServersAdmin275Test-1729403672-project-admin] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 531.377274] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 56eee633-2c01-4df2-8d2e-58ad712942f6] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 531.377274] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-67b5581e-e7b0-44df-9366-fcc31910c95f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.393839] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 531.393839] env[62066]: value = "task-1340631" [ 531.393839] env[62066]: _type = "Task" [ 531.393839] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 531.402333] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1340631, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 531.410877] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9371bde4-46e3-4272-a1fe-5125bccbd2c7 tempest-ServersTestJSON-680983284 tempest-ServersTestJSON-680983284-project-member] Acquiring lock "0040f46d-9a60-4fa1-8a00-c08022e17df0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 531.411171] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9371bde4-46e3-4272-a1fe-5125bccbd2c7 tempest-ServersTestJSON-680983284 tempest-ServersTestJSON-680983284-project-member] Lock "0040f46d-9a60-4fa1-8a00-c08022e17df0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 531.572967] env[62066]: DEBUG nova.compute.manager [None req-67ee84a7-af8a-43fc-bff6-48212ddfc8ae tempest-ServersNegativeTestMultiTenantJSON-1325431503 tempest-ServersNegativeTestMultiTenantJSON-1325431503-project-member] [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 531.599674] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Instance 4b34ff57-e137-4abc-8aed-76dd2ec8d313 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 531.599818] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Instance 56eee633-2c01-4df2-8d2e-58ad712942f6 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 531.599953] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Instance 1eeb879b-6d1e-4746-98b5-12c30d674fa9 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 531.600061] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Instance 095d69fe-bbd3-4850-9dcd-b4c2290b8352 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 531.600226] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Instance d522ee4a-9bd9-402b-b3bd-c36f93366249 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 531.600301] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Instance 68b65162-23a6-464f-a2f9-1635bea8786f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 531.600452] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Instance 64f9698e-16f5-447a-914a-9e936e05d101 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 531.600560] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Instance 95bb27b9-e5a3-418f-a078-ae22ebf40013 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 531.647957] env[62066]: DEBUG nova.network.neutron [-] [instance: 64f9698e-16f5-447a-914a-9e936e05d101] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 531.785055] env[62066]: DEBUG nova.network.neutron [None req-67ee84a7-af8a-43fc-bff6-48212ddfc8ae tempest-ServersNegativeTestMultiTenantJSON-1325431503 tempest-ServersNegativeTestMultiTenantJSON-1325431503-project-member] [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] Successfully created port: 88f48a5f-c406-4d4f-8c70-39da5145d0f6 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 531.828856] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5c743a8d-c557-4d89-8fe0-1cc65e783f93 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] Lock "e8693f3d-ce6e-4c94-a719-4e6723c16de8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 35.322s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 531.836421] env[62066]: DEBUG oslo_concurrency.lockutils [None req-659c7f6d-78fb-4b73-9024-3866f9052a37 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] Releasing lock "refresh_cache-4b34ff57-e137-4abc-8aed-76dd2ec8d313" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 531.836831] env[62066]: DEBUG nova.compute.manager [None req-659c7f6d-78fb-4b73-9024-3866f9052a37 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] [instance: 4b34ff57-e137-4abc-8aed-76dd2ec8d313] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 531.837054] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-659c7f6d-78fb-4b73-9024-3866f9052a37 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] [instance: 4b34ff57-e137-4abc-8aed-76dd2ec8d313] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 531.838780] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78bceb24-8d34-4969-8243-7e42854b1809 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.848943] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-659c7f6d-78fb-4b73-9024-3866f9052a37 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] [instance: 4b34ff57-e137-4abc-8aed-76dd2ec8d313] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 531.849301] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-79976447-162d-48f7-83b3-a6199a2dbc59 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.857120] env[62066]: DEBUG oslo_vmware.api [None req-659c7f6d-78fb-4b73-9024-3866f9052a37 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] Waiting for the task: (returnval){ [ 531.857120] env[62066]: value = "task-1340632" [ 531.857120] env[62066]: _type = "Task" [ 531.857120] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 531.869508] env[62066]: DEBUG oslo_vmware.api [None req-659c7f6d-78fb-4b73-9024-3866f9052a37 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] Task: {'id': task-1340632, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 531.903916] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1340631, 'name': CreateVM_Task, 'duration_secs': 0.350072} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 531.904373] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 56eee633-2c01-4df2-8d2e-58ad712942f6] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 531.904848] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a16a38bd-97b1-4d18-bd95-9847ae1f8463 tempest-ServersAdmin275Test-1729403672 tempest-ServersAdmin275Test-1729403672-project-admin] Acquiring lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 531.905077] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a16a38bd-97b1-4d18-bd95-9847ae1f8463 tempest-ServersAdmin275Test-1729403672 tempest-ServersAdmin275Test-1729403672-project-admin] Acquired lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 531.905829] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a16a38bd-97b1-4d18-bd95-9847ae1f8463 tempest-ServersAdmin275Test-1729403672 tempest-ServersAdmin275Test-1729403672-project-admin] Acquired external semaphore "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 531.905829] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f7811c9f-e699-45b8-b0f9-58e7f3846340 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.910885] env[62066]: DEBUG oslo_vmware.api [None req-a16a38bd-97b1-4d18-bd95-9847ae1f8463 tempest-ServersAdmin275Test-1729403672 tempest-ServersAdmin275Test-1729403672-project-admin] Waiting for the task: (returnval){ [ 531.910885] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]5228de4d-5595-a8be-beba-de43ad3f34ad" [ 531.910885] env[62066]: _type = "Task" [ 531.910885] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 531.920574] env[62066]: DEBUG oslo_vmware.api [None req-a16a38bd-97b1-4d18-bd95-9847ae1f8463 tempest-ServersAdmin275Test-1729403672 tempest-ServersAdmin275Test-1729403672-project-admin] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]5228de4d-5595-a8be-beba-de43ad3f34ad, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 532.075114] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7abf52e5-ee7c-4f92-a757-ffa8ca735a57 tempest-ServerActionsTestJSON-164481335 tempest-ServerActionsTestJSON-164481335-project-member] Acquiring lock "bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 532.075431] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7abf52e5-ee7c-4f92-a757-ffa8ca735a57 tempest-ServerActionsTestJSON-164481335 tempest-ServerActionsTestJSON-164481335-project-member] Lock "bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 532.104482] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Instance 1fde7ae0-c88c-4d7f-b657-355c20e2cb10 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 532.150101] env[62066]: INFO nova.compute.manager [-] [instance: 64f9698e-16f5-447a-914a-9e936e05d101] Took 1.04 seconds to deallocate network for instance. [ 532.152486] env[62066]: DEBUG nova.compute.claims [None req-9838e0c6-e0b2-45d5-9c5b-a0a0b9c60219 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] [instance: 64f9698e-16f5-447a-914a-9e936e05d101] Aborting claim: {{(pid=62066) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 532.152663] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9838e0c6-e0b2-45d5-9c5b-a0a0b9c60219 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 532.334950] env[62066]: DEBUG nova.compute.manager [None req-65bf05da-7d56-4074-8d13-63fde71b9619 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 532.368527] env[62066]: DEBUG oslo_vmware.api [None req-659c7f6d-78fb-4b73-9024-3866f9052a37 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] Task: {'id': task-1340632, 'name': PowerOffVM_Task, 'duration_secs': 0.128675} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 532.369367] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-659c7f6d-78fb-4b73-9024-3866f9052a37 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] [instance: 4b34ff57-e137-4abc-8aed-76dd2ec8d313] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 532.369572] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-659c7f6d-78fb-4b73-9024-3866f9052a37 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] [instance: 4b34ff57-e137-4abc-8aed-76dd2ec8d313] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 532.369827] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e926c1ed-a5f9-44b8-9e0a-01111eb07435 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 532.399839] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-659c7f6d-78fb-4b73-9024-3866f9052a37 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] [instance: 4b34ff57-e137-4abc-8aed-76dd2ec8d313] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 532.401522] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-659c7f6d-78fb-4b73-9024-3866f9052a37 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] [instance: 4b34ff57-e137-4abc-8aed-76dd2ec8d313] Deleting contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 532.401522] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-659c7f6d-78fb-4b73-9024-3866f9052a37 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] Deleting the datastore file [datastore2] 4b34ff57-e137-4abc-8aed-76dd2ec8d313 {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 532.401522] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-55abf9e1-7e3d-48af-a482-22ed44f74fbd {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 532.409338] env[62066]: DEBUG oslo_vmware.api [None req-659c7f6d-78fb-4b73-9024-3866f9052a37 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] Waiting for the task: (returnval){ [ 532.409338] env[62066]: value = "task-1340634" [ 532.409338] env[62066]: _type = "Task" [ 532.409338] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 532.421075] env[62066]: DEBUG oslo_vmware.api [None req-659c7f6d-78fb-4b73-9024-3866f9052a37 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] Task: {'id': task-1340634, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 532.424586] env[62066]: DEBUG oslo_vmware.api [None req-a16a38bd-97b1-4d18-bd95-9847ae1f8463 tempest-ServersAdmin275Test-1729403672 tempest-ServersAdmin275Test-1729403672-project-admin] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]5228de4d-5595-a8be-beba-de43ad3f34ad, 'name': SearchDatastore_Task, 'duration_secs': 0.009402} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 532.424766] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a16a38bd-97b1-4d18-bd95-9847ae1f8463 tempest-ServersAdmin275Test-1729403672 tempest-ServersAdmin275Test-1729403672-project-admin] Releasing lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 532.424991] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-a16a38bd-97b1-4d18-bd95-9847ae1f8463 tempest-ServersAdmin275Test-1729403672 tempest-ServersAdmin275Test-1729403672-project-admin] [instance: 56eee633-2c01-4df2-8d2e-58ad712942f6] Processing image 50ff584c-3b50-4395-af07-3e66769bc9f7 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 532.425450] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a16a38bd-97b1-4d18-bd95-9847ae1f8463 tempest-ServersAdmin275Test-1729403672 tempest-ServersAdmin275Test-1729403672-project-admin] Acquiring lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 532.425616] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a16a38bd-97b1-4d18-bd95-9847ae1f8463 tempest-ServersAdmin275Test-1729403672 tempest-ServersAdmin275Test-1729403672-project-admin] Acquired lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 532.425798] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-a16a38bd-97b1-4d18-bd95-9847ae1f8463 tempest-ServersAdmin275Test-1729403672 tempest-ServersAdmin275Test-1729403672-project-admin] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 532.426050] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3a5788a3-988a-4a08-ad89-f564c1d7b554 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 532.433821] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-a16a38bd-97b1-4d18-bd95-9847ae1f8463 tempest-ServersAdmin275Test-1729403672 tempest-ServersAdmin275Test-1729403672-project-admin] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 532.434011] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-a16a38bd-97b1-4d18-bd95-9847ae1f8463 tempest-ServersAdmin275Test-1729403672 tempest-ServersAdmin275Test-1729403672-project-admin] Folder [datastore2] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 532.434700] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3ff2ade2-8600-4595-83cf-e9ec5934b747 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 532.440322] env[62066]: DEBUG oslo_vmware.api [None req-a16a38bd-97b1-4d18-bd95-9847ae1f8463 tempest-ServersAdmin275Test-1729403672 tempest-ServersAdmin275Test-1729403672-project-admin] Waiting for the task: (returnval){ [ 532.440322] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]52eff6d2-62da-cf67-f43a-83b719237d74" [ 532.440322] env[62066]: _type = "Task" [ 532.440322] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 532.448625] env[62066]: DEBUG oslo_vmware.api [None req-a16a38bd-97b1-4d18-bd95-9847ae1f8463 tempest-ServersAdmin275Test-1729403672 tempest-ServersAdmin275Test-1729403672-project-admin] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52eff6d2-62da-cf67-f43a-83b719237d74, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 532.583825] env[62066]: DEBUG oslo_concurrency.lockutils [None req-29878964-568e-4c92-80e6-64ee3e8e11bb tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] Acquiring lock "0e53c28c-a440-4e0e-9a4d-ff1d52557752" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 532.583958] env[62066]: DEBUG oslo_concurrency.lockutils [None req-29878964-568e-4c92-80e6-64ee3e8e11bb tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] Lock "0e53c28c-a440-4e0e-9a4d-ff1d52557752" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 532.590024] env[62066]: DEBUG nova.compute.manager [None req-67ee84a7-af8a-43fc-bff6-48212ddfc8ae tempest-ServersNegativeTestMultiTenantJSON-1325431503 tempest-ServersNegativeTestMultiTenantJSON-1325431503-project-member] [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 532.610193] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Instance 4c67e82d-b51b-4820-af81-e6eefc7bcb36 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 532.630646] env[62066]: DEBUG nova.virt.hardware [None req-67ee84a7-af8a-43fc-bff6-48212ddfc8ae tempest-ServersNegativeTestMultiTenantJSON-1325431503 tempest-ServersNegativeTestMultiTenantJSON-1325431503-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-23T13:40:41Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-23T13:40:23Z,direct_url=,disk_format='vmdk',id=50ff584c-3b50-4395-af07-3e66769bc9f7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='eb52f7069a374c61ae946f052007c6d9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-23T13:40:24Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 532.631024] env[62066]: DEBUG nova.virt.hardware [None req-67ee84a7-af8a-43fc-bff6-48212ddfc8ae tempest-ServersNegativeTestMultiTenantJSON-1325431503 tempest-ServersNegativeTestMultiTenantJSON-1325431503-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 532.631024] env[62066]: DEBUG nova.virt.hardware [None req-67ee84a7-af8a-43fc-bff6-48212ddfc8ae tempest-ServersNegativeTestMultiTenantJSON-1325431503 tempest-ServersNegativeTestMultiTenantJSON-1325431503-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 532.631197] env[62066]: DEBUG nova.virt.hardware [None req-67ee84a7-af8a-43fc-bff6-48212ddfc8ae tempest-ServersNegativeTestMultiTenantJSON-1325431503 tempest-ServersNegativeTestMultiTenantJSON-1325431503-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 532.632129] env[62066]: DEBUG nova.virt.hardware [None req-67ee84a7-af8a-43fc-bff6-48212ddfc8ae tempest-ServersNegativeTestMultiTenantJSON-1325431503 tempest-ServersNegativeTestMultiTenantJSON-1325431503-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 532.632129] env[62066]: DEBUG nova.virt.hardware [None req-67ee84a7-af8a-43fc-bff6-48212ddfc8ae tempest-ServersNegativeTestMultiTenantJSON-1325431503 tempest-ServersNegativeTestMultiTenantJSON-1325431503-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 532.632129] env[62066]: DEBUG nova.virt.hardware [None req-67ee84a7-af8a-43fc-bff6-48212ddfc8ae tempest-ServersNegativeTestMultiTenantJSON-1325431503 tempest-ServersNegativeTestMultiTenantJSON-1325431503-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 532.632129] env[62066]: DEBUG nova.virt.hardware [None req-67ee84a7-af8a-43fc-bff6-48212ddfc8ae tempest-ServersNegativeTestMultiTenantJSON-1325431503 tempest-ServersNegativeTestMultiTenantJSON-1325431503-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 532.632129] env[62066]: DEBUG nova.virt.hardware [None req-67ee84a7-af8a-43fc-bff6-48212ddfc8ae tempest-ServersNegativeTestMultiTenantJSON-1325431503 tempest-ServersNegativeTestMultiTenantJSON-1325431503-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 532.632474] env[62066]: DEBUG nova.virt.hardware [None req-67ee84a7-af8a-43fc-bff6-48212ddfc8ae tempest-ServersNegativeTestMultiTenantJSON-1325431503 tempest-ServersNegativeTestMultiTenantJSON-1325431503-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 532.632474] env[62066]: DEBUG nova.virt.hardware [None req-67ee84a7-af8a-43fc-bff6-48212ddfc8ae tempest-ServersNegativeTestMultiTenantJSON-1325431503 tempest-ServersNegativeTestMultiTenantJSON-1325431503-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 532.633178] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcf5d776-eb34-4976-8501-3d0b3970bde0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 532.647347] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-019221c1-86ef-4f12-95fc-ece15c7b8dcc {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 532.863309] env[62066]: DEBUG oslo_concurrency.lockutils [None req-65bf05da-7d56-4074-8d13-63fde71b9619 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 532.925021] env[62066]: DEBUG oslo_vmware.api [None req-659c7f6d-78fb-4b73-9024-3866f9052a37 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] Task: {'id': task-1340634, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.089965} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 532.925021] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-659c7f6d-78fb-4b73-9024-3866f9052a37 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 532.925021] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-659c7f6d-78fb-4b73-9024-3866f9052a37 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] [instance: 4b34ff57-e137-4abc-8aed-76dd2ec8d313] Deleted contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 532.925021] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-659c7f6d-78fb-4b73-9024-3866f9052a37 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] [instance: 4b34ff57-e137-4abc-8aed-76dd2ec8d313] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 532.925021] env[62066]: INFO nova.compute.manager [None req-659c7f6d-78fb-4b73-9024-3866f9052a37 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] [instance: 4b34ff57-e137-4abc-8aed-76dd2ec8d313] Took 1.09 seconds to destroy the instance on the hypervisor. [ 532.925365] env[62066]: DEBUG oslo.service.loopingcall [None req-659c7f6d-78fb-4b73-9024-3866f9052a37 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 532.925365] env[62066]: DEBUG nova.compute.manager [-] [instance: 4b34ff57-e137-4abc-8aed-76dd2ec8d313] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 532.925365] env[62066]: DEBUG nova.network.neutron [-] [instance: 4b34ff57-e137-4abc-8aed-76dd2ec8d313] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 532.952834] env[62066]: DEBUG oslo_vmware.api [None req-a16a38bd-97b1-4d18-bd95-9847ae1f8463 tempest-ServersAdmin275Test-1729403672 tempest-ServersAdmin275Test-1729403672-project-admin] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52eff6d2-62da-cf67-f43a-83b719237d74, 'name': SearchDatastore_Task, 'duration_secs': 0.00781} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 532.956770] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-479cb2b2-9371-49a1-81c8-ab0ff6f8e120 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 532.959766] env[62066]: DEBUG nova.network.neutron [-] [instance: 4b34ff57-e137-4abc-8aed-76dd2ec8d313] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 532.966500] env[62066]: DEBUG oslo_vmware.api [None req-a16a38bd-97b1-4d18-bd95-9847ae1f8463 tempest-ServersAdmin275Test-1729403672 tempest-ServersAdmin275Test-1729403672-project-admin] Waiting for the task: (returnval){ [ 532.966500] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]5242f428-383f-669e-1b1a-20cdcfff6ac7" [ 532.966500] env[62066]: _type = "Task" [ 532.966500] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 532.984123] env[62066]: DEBUG oslo_vmware.api [None req-a16a38bd-97b1-4d18-bd95-9847ae1f8463 tempest-ServersAdmin275Test-1729403672 tempest-ServersAdmin275Test-1729403672-project-admin] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]5242f428-383f-669e-1b1a-20cdcfff6ac7, 'name': SearchDatastore_Task, 'duration_secs': 0.008877} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 532.984123] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a16a38bd-97b1-4d18-bd95-9847ae1f8463 tempest-ServersAdmin275Test-1729403672 tempest-ServersAdmin275Test-1729403672-project-admin] Releasing lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 532.984123] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-a16a38bd-97b1-4d18-bd95-9847ae1f8463 tempest-ServersAdmin275Test-1729403672 tempest-ServersAdmin275Test-1729403672-project-admin] Copying Virtual Disk [datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk to [datastore2] 56eee633-2c01-4df2-8d2e-58ad712942f6/56eee633-2c01-4df2-8d2e-58ad712942f6.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 532.985169] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0d1736b8-3284-4531-8888-ec7567e53219 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 532.993747] env[62066]: DEBUG oslo_vmware.api [None req-a16a38bd-97b1-4d18-bd95-9847ae1f8463 tempest-ServersAdmin275Test-1729403672 tempest-ServersAdmin275Test-1729403672-project-admin] Waiting for the task: (returnval){ [ 532.993747] env[62066]: value = "task-1340635" [ 532.993747] env[62066]: _type = "Task" [ 532.993747] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 533.004163] env[62066]: DEBUG oslo_vmware.api [None req-a16a38bd-97b1-4d18-bd95-9847ae1f8463 tempest-ServersAdmin275Test-1729403672 tempest-ServersAdmin275Test-1729403672-project-admin] Task: {'id': task-1340635, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 533.114659] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Instance 69d99063-a6a7-4832-a864-dc9fd6024ea8 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 533.262320] env[62066]: DEBUG nova.compute.manager [req-876b52ae-9666-4923-8772-077d143cbad4 req-76a383a0-5369-43c5-b766-61fecd9bf25c service nova] [instance: 64f9698e-16f5-447a-914a-9e936e05d101] Received event network-changed-8a431920-fe14-44e1-85f9-31b9af5b10c2 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 533.262320] env[62066]: DEBUG nova.compute.manager [req-876b52ae-9666-4923-8772-077d143cbad4 req-76a383a0-5369-43c5-b766-61fecd9bf25c service nova] [instance: 64f9698e-16f5-447a-914a-9e936e05d101] Refreshing instance network info cache due to event network-changed-8a431920-fe14-44e1-85f9-31b9af5b10c2. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 533.262504] env[62066]: DEBUG oslo_concurrency.lockutils [req-876b52ae-9666-4923-8772-077d143cbad4 req-76a383a0-5369-43c5-b766-61fecd9bf25c service nova] Acquiring lock "refresh_cache-64f9698e-16f5-447a-914a-9e936e05d101" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 533.264108] env[62066]: DEBUG oslo_concurrency.lockutils [req-876b52ae-9666-4923-8772-077d143cbad4 req-76a383a0-5369-43c5-b766-61fecd9bf25c service nova] Acquired lock "refresh_cache-64f9698e-16f5-447a-914a-9e936e05d101" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 533.264108] env[62066]: DEBUG nova.network.neutron [req-876b52ae-9666-4923-8772-077d143cbad4 req-76a383a0-5369-43c5-b766-61fecd9bf25c service nova] [instance: 64f9698e-16f5-447a-914a-9e936e05d101] Refreshing network info cache for port 8a431920-fe14-44e1-85f9-31b9af5b10c2 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 533.464814] env[62066]: DEBUG nova.network.neutron [-] [instance: 4b34ff57-e137-4abc-8aed-76dd2ec8d313] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 533.488762] env[62066]: ERROR nova.compute.manager [None req-67ee84a7-af8a-43fc-bff6-48212ddfc8ae tempest-ServersNegativeTestMultiTenantJSON-1325431503 tempest-ServersNegativeTestMultiTenantJSON-1325431503-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 88f48a5f-c406-4d4f-8c70-39da5145d0f6, please check neutron logs for more information. [ 533.488762] env[62066]: ERROR nova.compute.manager Traceback (most recent call last): [ 533.488762] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 533.488762] env[62066]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 533.488762] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 533.488762] env[62066]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 533.488762] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 533.488762] env[62066]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 533.488762] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 533.488762] env[62066]: ERROR nova.compute.manager self.force_reraise() [ 533.488762] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 533.488762] env[62066]: ERROR nova.compute.manager raise self.value [ 533.488762] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 533.488762] env[62066]: ERROR nova.compute.manager updated_port = self._update_port( [ 533.488762] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 533.488762] env[62066]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 533.489392] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 533.489392] env[62066]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 533.489392] env[62066]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 88f48a5f-c406-4d4f-8c70-39da5145d0f6, please check neutron logs for more information. [ 533.489392] env[62066]: ERROR nova.compute.manager [ 533.489392] env[62066]: Traceback (most recent call last): [ 533.489392] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 533.489392] env[62066]: listener.cb(fileno) [ 533.489392] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 533.489392] env[62066]: result = function(*args, **kwargs) [ 533.489392] env[62066]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 533.489392] env[62066]: return func(*args, **kwargs) [ 533.489392] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 533.489392] env[62066]: raise e [ 533.489392] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 533.489392] env[62066]: nwinfo = self.network_api.allocate_for_instance( [ 533.489392] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 533.489392] env[62066]: created_port_ids = self._update_ports_for_instance( [ 533.489392] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 533.489392] env[62066]: with excutils.save_and_reraise_exception(): [ 533.489392] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 533.489392] env[62066]: self.force_reraise() [ 533.489392] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 533.489392] env[62066]: raise self.value [ 533.489392] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 533.489392] env[62066]: updated_port = self._update_port( [ 533.489392] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 533.489392] env[62066]: _ensure_no_port_binding_failure(port) [ 533.489392] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 533.489392] env[62066]: raise exception.PortBindingFailed(port_id=port['id']) [ 533.490560] env[62066]: nova.exception.PortBindingFailed: Binding failed for port 88f48a5f-c406-4d4f-8c70-39da5145d0f6, please check neutron logs for more information. [ 533.490560] env[62066]: Removing descriptor: 15 [ 533.490560] env[62066]: ERROR nova.compute.manager [None req-67ee84a7-af8a-43fc-bff6-48212ddfc8ae tempest-ServersNegativeTestMultiTenantJSON-1325431503 tempest-ServersNegativeTestMultiTenantJSON-1325431503-project-member] [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 88f48a5f-c406-4d4f-8c70-39da5145d0f6, please check neutron logs for more information. [ 533.490560] env[62066]: ERROR nova.compute.manager [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] Traceback (most recent call last): [ 533.490560] env[62066]: ERROR nova.compute.manager [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 533.490560] env[62066]: ERROR nova.compute.manager [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] yield resources [ 533.490560] env[62066]: ERROR nova.compute.manager [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 533.490560] env[62066]: ERROR nova.compute.manager [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] self.driver.spawn(context, instance, image_meta, [ 533.490560] env[62066]: ERROR nova.compute.manager [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 533.490560] env[62066]: ERROR nova.compute.manager [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] self._vmops.spawn(context, instance, image_meta, injected_files, [ 533.490560] env[62066]: ERROR nova.compute.manager [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 533.490560] env[62066]: ERROR nova.compute.manager [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] vm_ref = self.build_virtual_machine(instance, [ 533.490977] env[62066]: ERROR nova.compute.manager [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 533.490977] env[62066]: ERROR nova.compute.manager [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] vif_infos = vmwarevif.get_vif_info(self._session, [ 533.490977] env[62066]: ERROR nova.compute.manager [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 533.490977] env[62066]: ERROR nova.compute.manager [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] for vif in network_info: [ 533.490977] env[62066]: ERROR nova.compute.manager [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 533.490977] env[62066]: ERROR nova.compute.manager [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] return self._sync_wrapper(fn, *args, **kwargs) [ 533.490977] env[62066]: ERROR nova.compute.manager [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 533.490977] env[62066]: ERROR nova.compute.manager [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] self.wait() [ 533.490977] env[62066]: ERROR nova.compute.manager [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 533.490977] env[62066]: ERROR nova.compute.manager [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] self[:] = self._gt.wait() [ 533.490977] env[62066]: ERROR nova.compute.manager [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 533.490977] env[62066]: ERROR nova.compute.manager [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] return self._exit_event.wait() [ 533.490977] env[62066]: ERROR nova.compute.manager [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 533.491454] env[62066]: ERROR nova.compute.manager [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] result = hub.switch() [ 533.491454] env[62066]: ERROR nova.compute.manager [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 533.491454] env[62066]: ERROR nova.compute.manager [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] return self.greenlet.switch() [ 533.491454] env[62066]: ERROR nova.compute.manager [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 533.491454] env[62066]: ERROR nova.compute.manager [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] result = function(*args, **kwargs) [ 533.491454] env[62066]: ERROR nova.compute.manager [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 533.491454] env[62066]: ERROR nova.compute.manager [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] return func(*args, **kwargs) [ 533.491454] env[62066]: ERROR nova.compute.manager [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 533.491454] env[62066]: ERROR nova.compute.manager [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] raise e [ 533.491454] env[62066]: ERROR nova.compute.manager [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 533.491454] env[62066]: ERROR nova.compute.manager [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] nwinfo = self.network_api.allocate_for_instance( [ 533.491454] env[62066]: ERROR nova.compute.manager [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 533.491454] env[62066]: ERROR nova.compute.manager [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] created_port_ids = self._update_ports_for_instance( [ 533.491850] env[62066]: ERROR nova.compute.manager [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 533.491850] env[62066]: ERROR nova.compute.manager [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] with excutils.save_and_reraise_exception(): [ 533.491850] env[62066]: ERROR nova.compute.manager [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 533.491850] env[62066]: ERROR nova.compute.manager [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] self.force_reraise() [ 533.491850] env[62066]: ERROR nova.compute.manager [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 533.491850] env[62066]: ERROR nova.compute.manager [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] raise self.value [ 533.491850] env[62066]: ERROR nova.compute.manager [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 533.491850] env[62066]: ERROR nova.compute.manager [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] updated_port = self._update_port( [ 533.491850] env[62066]: ERROR nova.compute.manager [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 533.491850] env[62066]: ERROR nova.compute.manager [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] _ensure_no_port_binding_failure(port) [ 533.491850] env[62066]: ERROR nova.compute.manager [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 533.491850] env[62066]: ERROR nova.compute.manager [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] raise exception.PortBindingFailed(port_id=port['id']) [ 533.495383] env[62066]: ERROR nova.compute.manager [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] nova.exception.PortBindingFailed: Binding failed for port 88f48a5f-c406-4d4f-8c70-39da5145d0f6, please check neutron logs for more information. [ 533.495383] env[62066]: ERROR nova.compute.manager [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] [ 533.495383] env[62066]: INFO nova.compute.manager [None req-67ee84a7-af8a-43fc-bff6-48212ddfc8ae tempest-ServersNegativeTestMultiTenantJSON-1325431503 tempest-ServersNegativeTestMultiTenantJSON-1325431503-project-member] [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] Terminating instance [ 533.499507] env[62066]: DEBUG oslo_concurrency.lockutils [None req-67ee84a7-af8a-43fc-bff6-48212ddfc8ae tempest-ServersNegativeTestMultiTenantJSON-1325431503 tempest-ServersNegativeTestMultiTenantJSON-1325431503-project-member] Acquiring lock "refresh_cache-95bb27b9-e5a3-418f-a078-ae22ebf40013" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 533.499845] env[62066]: DEBUG oslo_concurrency.lockutils [None req-67ee84a7-af8a-43fc-bff6-48212ddfc8ae tempest-ServersNegativeTestMultiTenantJSON-1325431503 tempest-ServersNegativeTestMultiTenantJSON-1325431503-project-member] Acquired lock "refresh_cache-95bb27b9-e5a3-418f-a078-ae22ebf40013" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 533.499958] env[62066]: DEBUG nova.network.neutron [None req-67ee84a7-af8a-43fc-bff6-48212ddfc8ae tempest-ServersNegativeTestMultiTenantJSON-1325431503 tempest-ServersNegativeTestMultiTenantJSON-1325431503-project-member] [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 533.508528] env[62066]: DEBUG oslo_vmware.api [None req-a16a38bd-97b1-4d18-bd95-9847ae1f8463 tempest-ServersAdmin275Test-1729403672 tempest-ServersAdmin275Test-1729403672-project-admin] Task: {'id': task-1340635, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 533.621385] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Instance eacdd472-5bb0-48bc-8603-9e1d08b41f55 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 533.745922] env[62066]: DEBUG nova.compute.manager [req-990a21ec-0bba-4b6c-af59-1ee5317726e0 req-2e6823cc-84c0-4ae4-86d6-2874017c2904 service nova] [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] Received event network-changed-88f48a5f-c406-4d4f-8c70-39da5145d0f6 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 533.746147] env[62066]: DEBUG nova.compute.manager [req-990a21ec-0bba-4b6c-af59-1ee5317726e0 req-2e6823cc-84c0-4ae4-86d6-2874017c2904 service nova] [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] Refreshing instance network info cache due to event network-changed-88f48a5f-c406-4d4f-8c70-39da5145d0f6. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 533.746345] env[62066]: DEBUG oslo_concurrency.lockutils [req-990a21ec-0bba-4b6c-af59-1ee5317726e0 req-2e6823cc-84c0-4ae4-86d6-2874017c2904 service nova] Acquiring lock "refresh_cache-95bb27b9-e5a3-418f-a078-ae22ebf40013" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 533.799191] env[62066]: DEBUG nova.network.neutron [req-876b52ae-9666-4923-8772-077d143cbad4 req-76a383a0-5369-43c5-b766-61fecd9bf25c service nova] [instance: 64f9698e-16f5-447a-914a-9e936e05d101] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 533.969086] env[62066]: INFO nova.compute.manager [-] [instance: 4b34ff57-e137-4abc-8aed-76dd2ec8d313] Took 1.04 seconds to deallocate network for instance. [ 534.012082] env[62066]: DEBUG oslo_vmware.api [None req-a16a38bd-97b1-4d18-bd95-9847ae1f8463 tempest-ServersAdmin275Test-1729403672 tempest-ServersAdmin275Test-1729403672-project-admin] Task: {'id': task-1340635, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.530831} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 534.015395] env[62066]: DEBUG nova.network.neutron [req-876b52ae-9666-4923-8772-077d143cbad4 req-76a383a0-5369-43c5-b766-61fecd9bf25c service nova] [instance: 64f9698e-16f5-447a-914a-9e936e05d101] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 534.018020] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-a16a38bd-97b1-4d18-bd95-9847ae1f8463 tempest-ServersAdmin275Test-1729403672 tempest-ServersAdmin275Test-1729403672-project-admin] Copied Virtual Disk [datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk to [datastore2] 56eee633-2c01-4df2-8d2e-58ad712942f6/56eee633-2c01-4df2-8d2e-58ad712942f6.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 534.018020] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-a16a38bd-97b1-4d18-bd95-9847ae1f8463 tempest-ServersAdmin275Test-1729403672 tempest-ServersAdmin275Test-1729403672-project-admin] [instance: 56eee633-2c01-4df2-8d2e-58ad712942f6] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 534.018020] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f0ae9ef7-b970-4eb4-b2c1-471279e23907 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.025170] env[62066]: DEBUG oslo_vmware.api [None req-a16a38bd-97b1-4d18-bd95-9847ae1f8463 tempest-ServersAdmin275Test-1729403672 tempest-ServersAdmin275Test-1729403672-project-admin] Waiting for the task: (returnval){ [ 534.025170] env[62066]: value = "task-1340636" [ 534.025170] env[62066]: _type = "Task" [ 534.025170] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 534.040025] env[62066]: DEBUG oslo_vmware.api [None req-a16a38bd-97b1-4d18-bd95-9847ae1f8463 tempest-ServersAdmin275Test-1729403672 tempest-ServersAdmin275Test-1729403672-project-admin] Task: {'id': task-1340636, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 534.040025] env[62066]: DEBUG nova.network.neutron [None req-67ee84a7-af8a-43fc-bff6-48212ddfc8ae tempest-ServersNegativeTestMultiTenantJSON-1325431503 tempest-ServersNegativeTestMultiTenantJSON-1325431503-project-member] [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 534.129421] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Instance 454a790d-3cb2-4ef2-995f-97e0eb1e2469 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 534.170880] env[62066]: DEBUG nova.network.neutron [None req-67ee84a7-af8a-43fc-bff6-48212ddfc8ae tempest-ServersNegativeTestMultiTenantJSON-1325431503 tempest-ServersNegativeTestMultiTenantJSON-1325431503-project-member] [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 534.478756] env[62066]: DEBUG oslo_concurrency.lockutils [None req-659c7f6d-78fb-4b73-9024-3866f9052a37 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 534.518263] env[62066]: DEBUG oslo_concurrency.lockutils [req-876b52ae-9666-4923-8772-077d143cbad4 req-76a383a0-5369-43c5-b766-61fecd9bf25c service nova] Releasing lock "refresh_cache-64f9698e-16f5-447a-914a-9e936e05d101" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 534.518263] env[62066]: DEBUG nova.compute.manager [req-876b52ae-9666-4923-8772-077d143cbad4 req-76a383a0-5369-43c5-b766-61fecd9bf25c service nova] [instance: 64f9698e-16f5-447a-914a-9e936e05d101] Received event network-vif-deleted-8a431920-fe14-44e1-85f9-31b9af5b10c2 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 534.539396] env[62066]: DEBUG oslo_vmware.api [None req-a16a38bd-97b1-4d18-bd95-9847ae1f8463 tempest-ServersAdmin275Test-1729403672 tempest-ServersAdmin275Test-1729403672-project-admin] Task: {'id': task-1340636, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068221} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 534.539753] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-a16a38bd-97b1-4d18-bd95-9847ae1f8463 tempest-ServersAdmin275Test-1729403672 tempest-ServersAdmin275Test-1729403672-project-admin] [instance: 56eee633-2c01-4df2-8d2e-58ad712942f6] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 534.540426] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df6430cc-a350-4151-bcad-42331ad95530 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.563429] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-a16a38bd-97b1-4d18-bd95-9847ae1f8463 tempest-ServersAdmin275Test-1729403672 tempest-ServersAdmin275Test-1729403672-project-admin] [instance: 56eee633-2c01-4df2-8d2e-58ad712942f6] Reconfiguring VM instance instance-00000005 to attach disk [datastore2] 56eee633-2c01-4df2-8d2e-58ad712942f6/56eee633-2c01-4df2-8d2e-58ad712942f6.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 534.563709] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-06e52b85-74ec-4cb6-946f-92a074a457ae {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.584216] env[62066]: DEBUG oslo_vmware.api [None req-a16a38bd-97b1-4d18-bd95-9847ae1f8463 tempest-ServersAdmin275Test-1729403672 tempest-ServersAdmin275Test-1729403672-project-admin] Waiting for the task: (returnval){ [ 534.584216] env[62066]: value = "task-1340637" [ 534.584216] env[62066]: _type = "Task" [ 534.584216] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 534.592920] env[62066]: DEBUG oslo_vmware.api [None req-a16a38bd-97b1-4d18-bd95-9847ae1f8463 tempest-ServersAdmin275Test-1729403672 tempest-ServersAdmin275Test-1729403672-project-admin] Task: {'id': task-1340637, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 534.635532] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Instance 0a08bc46-0eea-4802-b0be-eb24f3507f73 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 534.677443] env[62066]: DEBUG oslo_concurrency.lockutils [None req-67ee84a7-af8a-43fc-bff6-48212ddfc8ae tempest-ServersNegativeTestMultiTenantJSON-1325431503 tempest-ServersNegativeTestMultiTenantJSON-1325431503-project-member] Releasing lock "refresh_cache-95bb27b9-e5a3-418f-a078-ae22ebf40013" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 534.677443] env[62066]: DEBUG nova.compute.manager [None req-67ee84a7-af8a-43fc-bff6-48212ddfc8ae tempest-ServersNegativeTestMultiTenantJSON-1325431503 tempest-ServersNegativeTestMultiTenantJSON-1325431503-project-member] [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 534.677776] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-67ee84a7-af8a-43fc-bff6-48212ddfc8ae tempest-ServersNegativeTestMultiTenantJSON-1325431503 tempest-ServersNegativeTestMultiTenantJSON-1325431503-project-member] [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 534.677918] env[62066]: DEBUG oslo_concurrency.lockutils [req-990a21ec-0bba-4b6c-af59-1ee5317726e0 req-2e6823cc-84c0-4ae4-86d6-2874017c2904 service nova] Acquired lock "refresh_cache-95bb27b9-e5a3-418f-a078-ae22ebf40013" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 534.678116] env[62066]: DEBUG nova.network.neutron [req-990a21ec-0bba-4b6c-af59-1ee5317726e0 req-2e6823cc-84c0-4ae4-86d6-2874017c2904 service nova] [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] Refreshing network info cache for port 88f48a5f-c406-4d4f-8c70-39da5145d0f6 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 534.683463] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-fe246bac-6cf2-47d0-934f-f7a1cdf844e1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.698117] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a574238b-653c-412e-ab08-922ec06975c9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.725557] env[62066]: WARNING nova.virt.vmwareapi.vmops [None req-67ee84a7-af8a-43fc-bff6-48212ddfc8ae tempest-ServersNegativeTestMultiTenantJSON-1325431503 tempest-ServersNegativeTestMultiTenantJSON-1325431503-project-member] [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 95bb27b9-e5a3-418f-a078-ae22ebf40013 could not be found. [ 534.725749] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-67ee84a7-af8a-43fc-bff6-48212ddfc8ae tempest-ServersNegativeTestMultiTenantJSON-1325431503 tempest-ServersNegativeTestMultiTenantJSON-1325431503-project-member] [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 534.725934] env[62066]: INFO nova.compute.manager [None req-67ee84a7-af8a-43fc-bff6-48212ddfc8ae tempest-ServersNegativeTestMultiTenantJSON-1325431503 tempest-ServersNegativeTestMultiTenantJSON-1325431503-project-member] [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] Took 0.05 seconds to destroy the instance on the hypervisor. [ 534.726226] env[62066]: DEBUG oslo.service.loopingcall [None req-67ee84a7-af8a-43fc-bff6-48212ddfc8ae tempest-ServersNegativeTestMultiTenantJSON-1325431503 tempest-ServersNegativeTestMultiTenantJSON-1325431503-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 534.726453] env[62066]: DEBUG nova.compute.manager [-] [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 534.726586] env[62066]: DEBUG nova.network.neutron [-] [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 534.768741] env[62066]: DEBUG nova.network.neutron [-] [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 535.030066] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9fdb5839-7f19-4e7f-8faa-852358a443a9 tempest-ServersListShow296Test-1792957209 tempest-ServersListShow296Test-1792957209-project-member] Acquiring lock "b68fe0ac-732c-448c-8bae-2dcd0ce2dc73" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 535.030066] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9fdb5839-7f19-4e7f-8faa-852358a443a9 tempest-ServersListShow296Test-1792957209 tempest-ServersListShow296Test-1792957209-project-member] Lock "b68fe0ac-732c-448c-8bae-2dcd0ce2dc73" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 535.098116] env[62066]: DEBUG oslo_vmware.api [None req-a16a38bd-97b1-4d18-bd95-9847ae1f8463 tempest-ServersAdmin275Test-1729403672 tempest-ServersAdmin275Test-1729403672-project-admin] Task: {'id': task-1340637, 'name': ReconfigVM_Task, 'duration_secs': 0.284582} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 535.098116] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-a16a38bd-97b1-4d18-bd95-9847ae1f8463 tempest-ServersAdmin275Test-1729403672 tempest-ServersAdmin275Test-1729403672-project-admin] [instance: 56eee633-2c01-4df2-8d2e-58ad712942f6] Reconfigured VM instance instance-00000005 to attach disk [datastore2] 56eee633-2c01-4df2-8d2e-58ad712942f6/56eee633-2c01-4df2-8d2e-58ad712942f6.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 535.099489] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a5ee8160-53e3-4f5f-8716-68678902253b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.106990] env[62066]: DEBUG oslo_vmware.api [None req-a16a38bd-97b1-4d18-bd95-9847ae1f8463 tempest-ServersAdmin275Test-1729403672 tempest-ServersAdmin275Test-1729403672-project-admin] Waiting for the task: (returnval){ [ 535.106990] env[62066]: value = "task-1340638" [ 535.106990] env[62066]: _type = "Task" [ 535.106990] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 535.115730] env[62066]: DEBUG oslo_vmware.api [None req-a16a38bd-97b1-4d18-bd95-9847ae1f8463 tempest-ServersAdmin275Test-1729403672 tempest-ServersAdmin275Test-1729403672-project-admin] Task: {'id': task-1340638, 'name': Rename_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 535.139720] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Instance 39cd204a-49c2-4309-9030-555056ab125e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 535.206480] env[62066]: DEBUG nova.network.neutron [req-990a21ec-0bba-4b6c-af59-1ee5317726e0 req-2e6823cc-84c0-4ae4-86d6-2874017c2904 service nova] [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 535.275226] env[62066]: DEBUG nova.network.neutron [-] [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 535.307061] env[62066]: DEBUG nova.network.neutron [req-990a21ec-0bba-4b6c-af59-1ee5317726e0 req-2e6823cc-84c0-4ae4-86d6-2874017c2904 service nova] [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 535.316183] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5e085536-daab-46ff-8eb8-69c1da687135 tempest-ServersAaction247Test-1482016775 tempest-ServersAaction247Test-1482016775-project-member] Acquiring lock "5d54c3fd-b239-4965-b7a3-5909e8de8bc0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 535.316478] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5e085536-daab-46ff-8eb8-69c1da687135 tempest-ServersAaction247Test-1482016775 tempest-ServersAaction247Test-1482016775-project-member] Lock "5d54c3fd-b239-4965-b7a3-5909e8de8bc0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 535.618219] env[62066]: DEBUG oslo_vmware.api [None req-a16a38bd-97b1-4d18-bd95-9847ae1f8463 tempest-ServersAdmin275Test-1729403672 tempest-ServersAdmin275Test-1729403672-project-admin] Task: {'id': task-1340638, 'name': Rename_Task, 'duration_secs': 0.149996} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 535.619045] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-a16a38bd-97b1-4d18-bd95-9847ae1f8463 tempest-ServersAdmin275Test-1729403672 tempest-ServersAdmin275Test-1729403672-project-admin] [instance: 56eee633-2c01-4df2-8d2e-58ad712942f6] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 535.619045] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-be6a9b9a-1a7e-46d0-ac49-8dc9e6639521 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.626093] env[62066]: DEBUG oslo_vmware.api [None req-a16a38bd-97b1-4d18-bd95-9847ae1f8463 tempest-ServersAdmin275Test-1729403672 tempest-ServersAdmin275Test-1729403672-project-admin] Waiting for the task: (returnval){ [ 535.626093] env[62066]: value = "task-1340639" [ 535.626093] env[62066]: _type = "Task" [ 535.626093] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 535.636132] env[62066]: DEBUG oslo_vmware.api [None req-a16a38bd-97b1-4d18-bd95-9847ae1f8463 tempest-ServersAdmin275Test-1729403672 tempest-ServersAdmin275Test-1729403672-project-admin] Task: {'id': task-1340639, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 535.645658] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Instance 194588b4-fe40-4286-8036-874a7c410327 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 535.783020] env[62066]: INFO nova.compute.manager [-] [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] Took 1.06 seconds to deallocate network for instance. [ 535.786570] env[62066]: DEBUG nova.compute.claims [None req-67ee84a7-af8a-43fc-bff6-48212ddfc8ae tempest-ServersNegativeTestMultiTenantJSON-1325431503 tempest-ServersNegativeTestMultiTenantJSON-1325431503-project-member] [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] Aborting claim: {{(pid=62066) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 535.786847] env[62066]: DEBUG oslo_concurrency.lockutils [None req-67ee84a7-af8a-43fc-bff6-48212ddfc8ae tempest-ServersNegativeTestMultiTenantJSON-1325431503 tempest-ServersNegativeTestMultiTenantJSON-1325431503-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 535.810011] env[62066]: DEBUG oslo_concurrency.lockutils [req-990a21ec-0bba-4b6c-af59-1ee5317726e0 req-2e6823cc-84c0-4ae4-86d6-2874017c2904 service nova] Releasing lock "refresh_cache-95bb27b9-e5a3-418f-a078-ae22ebf40013" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 535.913055] env[62066]: DEBUG oslo_concurrency.lockutils [None req-2c107b45-3cbf-4be9-af5a-796fbdb9b5fe tempest-ImagesNegativeTestJSON-881899573 tempest-ImagesNegativeTestJSON-881899573-project-member] Acquiring lock "cb319720-29ab-4ff2-a71e-f77ae8c85735" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 535.913055] env[62066]: DEBUG oslo_concurrency.lockutils [None req-2c107b45-3cbf-4be9-af5a-796fbdb9b5fe tempest-ImagesNegativeTestJSON-881899573 tempest-ImagesNegativeTestJSON-881899573-project-member] Lock "cb319720-29ab-4ff2-a71e-f77ae8c85735" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 536.137536] env[62066]: DEBUG oslo_vmware.api [None req-a16a38bd-97b1-4d18-bd95-9847ae1f8463 tempest-ServersAdmin275Test-1729403672 tempest-ServersAdmin275Test-1729403672-project-admin] Task: {'id': task-1340639, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 536.150391] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Instance a187f73b-8fb1-42b3-9a07-24d16aa0f152 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 536.249392] env[62066]: DEBUG nova.compute.manager [req-93ae4e90-09d8-4127-abd2-d7ad9df71c6f req-788a60bd-fea2-47c6-862b-72798b2ec7a1 service nova] [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] Received event network-vif-deleted-88f48a5f-c406-4d4f-8c70-39da5145d0f6 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 536.639267] env[62066]: DEBUG oslo_vmware.api [None req-a16a38bd-97b1-4d18-bd95-9847ae1f8463 tempest-ServersAdmin275Test-1729403672 tempest-ServersAdmin275Test-1729403672-project-admin] Task: {'id': task-1340639, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 536.653927] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Instance eee6b9e6-2707-4a3b-925b-a7f50cdc32cb has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 537.138182] env[62066]: DEBUG oslo_vmware.api [None req-a16a38bd-97b1-4d18-bd95-9847ae1f8463 tempest-ServersAdmin275Test-1729403672 tempest-ServersAdmin275Test-1729403672-project-admin] Task: {'id': task-1340639, 'name': PowerOnVM_Task, 'duration_secs': 1.022328} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 537.138501] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-a16a38bd-97b1-4d18-bd95-9847ae1f8463 tempest-ServersAdmin275Test-1729403672 tempest-ServersAdmin275Test-1729403672-project-admin] [instance: 56eee633-2c01-4df2-8d2e-58ad712942f6] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 537.138704] env[62066]: DEBUG nova.compute.manager [None req-a16a38bd-97b1-4d18-bd95-9847ae1f8463 tempest-ServersAdmin275Test-1729403672 tempest-ServersAdmin275Test-1729403672-project-admin] [instance: 56eee633-2c01-4df2-8d2e-58ad712942f6] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 537.139520] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92294a24-3dc6-40e3-9f64-4db25534973b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.156889] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Instance 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 537.660060] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Instance 72c2dbe0-78de-4a7d-98df-cf405a5f20e2 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 537.662059] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a16a38bd-97b1-4d18-bd95-9847ae1f8463 tempest-ServersAdmin275Test-1729403672 tempest-ServersAdmin275Test-1729403672-project-admin] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 538.163966] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Instance 08d41411-7928-4379-9f2e-c6ce00843d82 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 538.667781] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Instance 623300f7-54d2-4b0a-b356-a1ae585682cb has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 538.907480] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ef1737fc-6565-4a6f-bd10-02fc3b07f848 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Acquiring lock "97e51e6c-9a3a-4b68-b737-d48090a22b1b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 538.908257] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ef1737fc-6565-4a6f-bd10-02fc3b07f848 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Lock "97e51e6c-9a3a-4b68-b737-d48090a22b1b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 539.173638] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Instance 0040f46d-9a60-4fa1-8a00-c08022e17df0 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 539.409217] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b105d8cd-4f01-4e61-a779-8b209763af2b tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Acquiring lock "56eee633-2c01-4df2-8d2e-58ad712942f6" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 539.409471] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b105d8cd-4f01-4e61-a779-8b209763af2b tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Lock "56eee633-2c01-4df2-8d2e-58ad712942f6" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 539.409687] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b105d8cd-4f01-4e61-a779-8b209763af2b tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Acquiring lock "56eee633-2c01-4df2-8d2e-58ad712942f6-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 539.409872] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b105d8cd-4f01-4e61-a779-8b209763af2b tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Lock "56eee633-2c01-4df2-8d2e-58ad712942f6-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 539.410054] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b105d8cd-4f01-4e61-a779-8b209763af2b tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Lock "56eee633-2c01-4df2-8d2e-58ad712942f6-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 539.415228] env[62066]: INFO nova.compute.manager [None req-b105d8cd-4f01-4e61-a779-8b209763af2b tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] [instance: 56eee633-2c01-4df2-8d2e-58ad712942f6] Terminating instance [ 539.418405] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b105d8cd-4f01-4e61-a779-8b209763af2b tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Acquiring lock "refresh_cache-56eee633-2c01-4df2-8d2e-58ad712942f6" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 539.418581] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b105d8cd-4f01-4e61-a779-8b209763af2b tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Acquired lock "refresh_cache-56eee633-2c01-4df2-8d2e-58ad712942f6" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 539.418729] env[62066]: DEBUG nova.network.neutron [None req-b105d8cd-4f01-4e61-a779-8b209763af2b tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] [instance: 56eee633-2c01-4df2-8d2e-58ad712942f6] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 539.677343] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Instance bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 539.677343] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Total usable vcpus: 48, total allocated vcpus: 8 {{(pid=62066) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 539.677543] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2048MB phys_disk=200GB used_disk=8GB total_vcpus=48 used_vcpus=8 pci_stats=[] {{(pid=62066) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 539.984070] env[62066]: DEBUG nova.network.neutron [None req-b105d8cd-4f01-4e61-a779-8b209763af2b tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] [instance: 56eee633-2c01-4df2-8d2e-58ad712942f6] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 540.144901] env[62066]: DEBUG nova.network.neutron [None req-b105d8cd-4f01-4e61-a779-8b209763af2b tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] [instance: 56eee633-2c01-4df2-8d2e-58ad712942f6] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 540.162784] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10372515-492a-476d-be70-152881db53c4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.173683] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0d17060-ed41-453e-9d6f-3d9706aa840f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.206607] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ca59c98-999f-40e3-9910-1d8b1fc1d7b3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.214039] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0039e4a0-3b32-4adf-a7fb-1e9905ca7db0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.229659] env[62066]: DEBUG nova.compute.provider_tree [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 540.649575] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b105d8cd-4f01-4e61-a779-8b209763af2b tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Releasing lock "refresh_cache-56eee633-2c01-4df2-8d2e-58ad712942f6" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 540.650156] env[62066]: DEBUG nova.compute.manager [None req-b105d8cd-4f01-4e61-a779-8b209763af2b tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] [instance: 56eee633-2c01-4df2-8d2e-58ad712942f6] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 540.650156] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-b105d8cd-4f01-4e61-a779-8b209763af2b tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] [instance: 56eee633-2c01-4df2-8d2e-58ad712942f6] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 540.650957] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-534010e9-82c4-44d4-8e51-55ddc47a291b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.661730] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-b105d8cd-4f01-4e61-a779-8b209763af2b tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] [instance: 56eee633-2c01-4df2-8d2e-58ad712942f6] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 540.661938] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-677dd78e-20c4-404d-949b-7ea9ced6dc1e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.668243] env[62066]: DEBUG oslo_vmware.api [None req-b105d8cd-4f01-4e61-a779-8b209763af2b tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Waiting for the task: (returnval){ [ 540.668243] env[62066]: value = "task-1340640" [ 540.668243] env[62066]: _type = "Task" [ 540.668243] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 540.677033] env[62066]: DEBUG oslo_vmware.api [None req-b105d8cd-4f01-4e61-a779-8b209763af2b tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Task: {'id': task-1340640, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 540.739609] env[62066]: DEBUG nova.scheduler.client.report [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 541.182722] env[62066]: DEBUG oslo_vmware.api [None req-b105d8cd-4f01-4e61-a779-8b209763af2b tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Task: {'id': task-1340640, 'name': PowerOffVM_Task, 'duration_secs': 0.131081} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 541.183043] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-b105d8cd-4f01-4e61-a779-8b209763af2b tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] [instance: 56eee633-2c01-4df2-8d2e-58ad712942f6] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 541.183043] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-b105d8cd-4f01-4e61-a779-8b209763af2b tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] [instance: 56eee633-2c01-4df2-8d2e-58ad712942f6] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 541.183769] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c2be03e3-d67f-470a-979f-fd41eea9635c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.211129] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-b105d8cd-4f01-4e61-a779-8b209763af2b tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] [instance: 56eee633-2c01-4df2-8d2e-58ad712942f6] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 541.211350] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-b105d8cd-4f01-4e61-a779-8b209763af2b tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] [instance: 56eee633-2c01-4df2-8d2e-58ad712942f6] Deleting contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 541.211527] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-b105d8cd-4f01-4e61-a779-8b209763af2b tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Deleting the datastore file [datastore2] 56eee633-2c01-4df2-8d2e-58ad712942f6 {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 541.211857] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0ec8721d-3376-4dad-9604-a145b171013f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.219803] env[62066]: DEBUG oslo_vmware.api [None req-b105d8cd-4f01-4e61-a779-8b209763af2b tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Waiting for the task: (returnval){ [ 541.219803] env[62066]: value = "task-1340642" [ 541.219803] env[62066]: _type = "Task" [ 541.219803] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 541.229687] env[62066]: DEBUG oslo_vmware.api [None req-b105d8cd-4f01-4e61-a779-8b209763af2b tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Task: {'id': task-1340642, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 541.246644] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62066) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 541.246907] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 10.682s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 541.247206] env[62066]: DEBUG oslo_concurrency.lockutils [None req-df85dbc4-0217-427e-9c09-82d1a0e6d774 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 18.721s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 541.732753] env[62066]: DEBUG oslo_vmware.api [None req-b105d8cd-4f01-4e61-a779-8b209763af2b tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Task: {'id': task-1340642, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.095763} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 541.733815] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-b105d8cd-4f01-4e61-a779-8b209763af2b tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 541.734560] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-b105d8cd-4f01-4e61-a779-8b209763af2b tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] [instance: 56eee633-2c01-4df2-8d2e-58ad712942f6] Deleted contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 541.737594] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-b105d8cd-4f01-4e61-a779-8b209763af2b tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] [instance: 56eee633-2c01-4df2-8d2e-58ad712942f6] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 541.737594] env[62066]: INFO nova.compute.manager [None req-b105d8cd-4f01-4e61-a779-8b209763af2b tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] [instance: 56eee633-2c01-4df2-8d2e-58ad712942f6] Took 1.09 seconds to destroy the instance on the hypervisor. [ 541.737594] env[62066]: DEBUG oslo.service.loopingcall [None req-b105d8cd-4f01-4e61-a779-8b209763af2b tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 541.737594] env[62066]: DEBUG nova.compute.manager [-] [instance: 56eee633-2c01-4df2-8d2e-58ad712942f6] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 541.737594] env[62066]: DEBUG nova.network.neutron [-] [instance: 56eee633-2c01-4df2-8d2e-58ad712942f6] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 541.781559] env[62066]: DEBUG nova.network.neutron [-] [instance: 56eee633-2c01-4df2-8d2e-58ad712942f6] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 542.216124] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3be010bc-dcf1-43e7-b26f-7c1fc1d3a3c3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.224195] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-492a3c44-8606-422c-b97c-c5f013412474 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.258155] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ffd0f6a-0e62-43f3-b30b-b4e6086980bd {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.265956] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5586d8f-4dc2-4091-b050-df0e76038a0a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.279733] env[62066]: DEBUG nova.compute.provider_tree [None req-df85dbc4-0217-427e-9c09-82d1a0e6d774 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 542.284646] env[62066]: DEBUG nova.network.neutron [-] [instance: 56eee633-2c01-4df2-8d2e-58ad712942f6] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 542.783121] env[62066]: DEBUG nova.scheduler.client.report [None req-df85dbc4-0217-427e-9c09-82d1a0e6d774 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 542.788405] env[62066]: INFO nova.compute.manager [-] [instance: 56eee633-2c01-4df2-8d2e-58ad712942f6] Took 1.05 seconds to deallocate network for instance. [ 543.289713] env[62066]: DEBUG oslo_concurrency.lockutils [None req-df85dbc4-0217-427e-9c09-82d1a0e6d774 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.042s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 543.290840] env[62066]: ERROR nova.compute.manager [None req-df85dbc4-0217-427e-9c09-82d1a0e6d774 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port e279d463-9a92-4d4b-913e-d08e912e8b3e, please check neutron logs for more information. [ 543.290840] env[62066]: ERROR nova.compute.manager [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] Traceback (most recent call last): [ 543.290840] env[62066]: ERROR nova.compute.manager [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 543.290840] env[62066]: ERROR nova.compute.manager [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] self.driver.spawn(context, instance, image_meta, [ 543.290840] env[62066]: ERROR nova.compute.manager [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 543.290840] env[62066]: ERROR nova.compute.manager [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 543.290840] env[62066]: ERROR nova.compute.manager [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 543.290840] env[62066]: ERROR nova.compute.manager [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] vm_ref = self.build_virtual_machine(instance, [ 543.290840] env[62066]: ERROR nova.compute.manager [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 543.290840] env[62066]: ERROR nova.compute.manager [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] vif_infos = vmwarevif.get_vif_info(self._session, [ 543.290840] env[62066]: ERROR nova.compute.manager [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 543.291206] env[62066]: ERROR nova.compute.manager [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] for vif in network_info: [ 543.291206] env[62066]: ERROR nova.compute.manager [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 543.291206] env[62066]: ERROR nova.compute.manager [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] return self._sync_wrapper(fn, *args, **kwargs) [ 543.291206] env[62066]: ERROR nova.compute.manager [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 543.291206] env[62066]: ERROR nova.compute.manager [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] self.wait() [ 543.291206] env[62066]: ERROR nova.compute.manager [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 543.291206] env[62066]: ERROR nova.compute.manager [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] self[:] = self._gt.wait() [ 543.291206] env[62066]: ERROR nova.compute.manager [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 543.291206] env[62066]: ERROR nova.compute.manager [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] return self._exit_event.wait() [ 543.291206] env[62066]: ERROR nova.compute.manager [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 543.291206] env[62066]: ERROR nova.compute.manager [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] result = hub.switch() [ 543.291206] env[62066]: ERROR nova.compute.manager [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 543.291206] env[62066]: ERROR nova.compute.manager [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] return self.greenlet.switch() [ 543.291560] env[62066]: ERROR nova.compute.manager [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 543.291560] env[62066]: ERROR nova.compute.manager [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] result = function(*args, **kwargs) [ 543.291560] env[62066]: ERROR nova.compute.manager [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 543.291560] env[62066]: ERROR nova.compute.manager [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] return func(*args, **kwargs) [ 543.291560] env[62066]: ERROR nova.compute.manager [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 543.291560] env[62066]: ERROR nova.compute.manager [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] raise e [ 543.291560] env[62066]: ERROR nova.compute.manager [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 543.291560] env[62066]: ERROR nova.compute.manager [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] nwinfo = self.network_api.allocate_for_instance( [ 543.291560] env[62066]: ERROR nova.compute.manager [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 543.291560] env[62066]: ERROR nova.compute.manager [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] created_port_ids = self._update_ports_for_instance( [ 543.291560] env[62066]: ERROR nova.compute.manager [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 543.291560] env[62066]: ERROR nova.compute.manager [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] with excutils.save_and_reraise_exception(): [ 543.291560] env[62066]: ERROR nova.compute.manager [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 543.291922] env[62066]: ERROR nova.compute.manager [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] self.force_reraise() [ 543.291922] env[62066]: ERROR nova.compute.manager [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 543.291922] env[62066]: ERROR nova.compute.manager [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] raise self.value [ 543.291922] env[62066]: ERROR nova.compute.manager [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 543.291922] env[62066]: ERROR nova.compute.manager [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] updated_port = self._update_port( [ 543.291922] env[62066]: ERROR nova.compute.manager [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 543.291922] env[62066]: ERROR nova.compute.manager [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] _ensure_no_port_binding_failure(port) [ 543.291922] env[62066]: ERROR nova.compute.manager [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 543.291922] env[62066]: ERROR nova.compute.manager [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] raise exception.PortBindingFailed(port_id=port['id']) [ 543.291922] env[62066]: ERROR nova.compute.manager [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] nova.exception.PortBindingFailed: Binding failed for port e279d463-9a92-4d4b-913e-d08e912e8b3e, please check neutron logs for more information. [ 543.291922] env[62066]: ERROR nova.compute.manager [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] [ 543.292238] env[62066]: DEBUG nova.compute.utils [None req-df85dbc4-0217-427e-9c09-82d1a0e6d774 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] Binding failed for port e279d463-9a92-4d4b-913e-d08e912e8b3e, please check neutron logs for more information. {{(pid=62066) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 543.292847] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0e23bf5c-5716-4eba-9510-a3597f398645 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 19.511s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 543.299022] env[62066]: DEBUG nova.compute.manager [None req-df85dbc4-0217-427e-9c09-82d1a0e6d774 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] Build of instance 1eeb879b-6d1e-4746-98b5-12c30d674fa9 was re-scheduled: Binding failed for port e279d463-9a92-4d4b-913e-d08e912e8b3e, please check neutron logs for more information. {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 543.299022] env[62066]: DEBUG nova.compute.manager [None req-df85dbc4-0217-427e-9c09-82d1a0e6d774 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] Unplugging VIFs for instance {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 543.299022] env[62066]: DEBUG oslo_concurrency.lockutils [None req-df85dbc4-0217-427e-9c09-82d1a0e6d774 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] Acquiring lock "refresh_cache-1eeb879b-6d1e-4746-98b5-12c30d674fa9" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 543.299022] env[62066]: DEBUG oslo_concurrency.lockutils [None req-df85dbc4-0217-427e-9c09-82d1a0e6d774 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] Acquired lock "refresh_cache-1eeb879b-6d1e-4746-98b5-12c30d674fa9" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 543.299339] env[62066]: DEBUG nova.network.neutron [None req-df85dbc4-0217-427e-9c09-82d1a0e6d774 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 543.304206] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b105d8cd-4f01-4e61-a779-8b209763af2b tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 543.831671] env[62066]: DEBUG nova.network.neutron [None req-df85dbc4-0217-427e-9c09-82d1a0e6d774 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 543.996554] env[62066]: DEBUG nova.network.neutron [None req-df85dbc4-0217-427e-9c09-82d1a0e6d774 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 544.057606] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4d395966-f335-497d-b8a0-23e22d9b79f7 tempest-ServerActionsTestOtherA-1520864800 tempest-ServerActionsTestOtherA-1520864800-project-member] Acquiring lock "56984993-a8b1-464e-b20c-79ffe67d2491" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 544.057606] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4d395966-f335-497d-b8a0-23e22d9b79f7 tempest-ServerActionsTestOtherA-1520864800 tempest-ServerActionsTestOtherA-1520864800-project-member] Lock "56984993-a8b1-464e-b20c-79ffe67d2491" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 544.238853] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7340d79-8840-46bd-9c9f-d2a9e14abcda {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.247805] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a7333e3-bdfa-43da-8c75-58b9b6e7cb12 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.286336] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1aca0da-4032-42e7-9e1c-bd15fdf1f0b7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.291976] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbc63b49-3f4d-4a17-9e4a-1e478281edc0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.305742] env[62066]: DEBUG nova.compute.provider_tree [None req-0e23bf5c-5716-4eba-9510-a3597f398645 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 544.501851] env[62066]: DEBUG oslo_concurrency.lockutils [None req-df85dbc4-0217-427e-9c09-82d1a0e6d774 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] Releasing lock "refresh_cache-1eeb879b-6d1e-4746-98b5-12c30d674fa9" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 544.501851] env[62066]: DEBUG nova.compute.manager [None req-df85dbc4-0217-427e-9c09-82d1a0e6d774 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 544.501851] env[62066]: DEBUG nova.compute.manager [None req-df85dbc4-0217-427e-9c09-82d1a0e6d774 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 544.501851] env[62066]: DEBUG nova.network.neutron [None req-df85dbc4-0217-427e-9c09-82d1a0e6d774 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 544.524323] env[62066]: DEBUG nova.network.neutron [None req-df85dbc4-0217-427e-9c09-82d1a0e6d774 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 544.809514] env[62066]: DEBUG nova.scheduler.client.report [None req-0e23bf5c-5716-4eba-9510-a3597f398645 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 545.028032] env[62066]: DEBUG nova.network.neutron [None req-df85dbc4-0217-427e-9c09-82d1a0e6d774 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 545.317908] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0e23bf5c-5716-4eba-9510-a3597f398645 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.025s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 545.318565] env[62066]: ERROR nova.compute.manager [None req-0e23bf5c-5716-4eba-9510-a3597f398645 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 78745978-2782-4cbd-8fbd-fbdf247cc830, please check neutron logs for more information. [ 545.318565] env[62066]: ERROR nova.compute.manager [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] Traceback (most recent call last): [ 545.318565] env[62066]: ERROR nova.compute.manager [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 545.318565] env[62066]: ERROR nova.compute.manager [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] self.driver.spawn(context, instance, image_meta, [ 545.318565] env[62066]: ERROR nova.compute.manager [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 545.318565] env[62066]: ERROR nova.compute.manager [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] self._vmops.spawn(context, instance, image_meta, injected_files, [ 545.318565] env[62066]: ERROR nova.compute.manager [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 545.318565] env[62066]: ERROR nova.compute.manager [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] vm_ref = self.build_virtual_machine(instance, [ 545.318565] env[62066]: ERROR nova.compute.manager [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 545.318565] env[62066]: ERROR nova.compute.manager [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] vif_infos = vmwarevif.get_vif_info(self._session, [ 545.318565] env[62066]: ERROR nova.compute.manager [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 545.318958] env[62066]: ERROR nova.compute.manager [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] for vif in network_info: [ 545.318958] env[62066]: ERROR nova.compute.manager [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 545.318958] env[62066]: ERROR nova.compute.manager [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] return self._sync_wrapper(fn, *args, **kwargs) [ 545.318958] env[62066]: ERROR nova.compute.manager [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 545.318958] env[62066]: ERROR nova.compute.manager [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] self.wait() [ 545.318958] env[62066]: ERROR nova.compute.manager [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 545.318958] env[62066]: ERROR nova.compute.manager [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] self[:] = self._gt.wait() [ 545.318958] env[62066]: ERROR nova.compute.manager [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 545.318958] env[62066]: ERROR nova.compute.manager [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] return self._exit_event.wait() [ 545.318958] env[62066]: ERROR nova.compute.manager [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 545.318958] env[62066]: ERROR nova.compute.manager [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] result = hub.switch() [ 545.318958] env[62066]: ERROR nova.compute.manager [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 545.318958] env[62066]: ERROR nova.compute.manager [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] return self.greenlet.switch() [ 545.319456] env[62066]: ERROR nova.compute.manager [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 545.319456] env[62066]: ERROR nova.compute.manager [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] result = function(*args, **kwargs) [ 545.319456] env[62066]: ERROR nova.compute.manager [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 545.319456] env[62066]: ERROR nova.compute.manager [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] return func(*args, **kwargs) [ 545.319456] env[62066]: ERROR nova.compute.manager [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 545.319456] env[62066]: ERROR nova.compute.manager [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] raise e [ 545.319456] env[62066]: ERROR nova.compute.manager [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 545.319456] env[62066]: ERROR nova.compute.manager [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] nwinfo = self.network_api.allocate_for_instance( [ 545.319456] env[62066]: ERROR nova.compute.manager [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 545.319456] env[62066]: ERROR nova.compute.manager [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] created_port_ids = self._update_ports_for_instance( [ 545.319456] env[62066]: ERROR nova.compute.manager [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 545.319456] env[62066]: ERROR nova.compute.manager [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] with excutils.save_and_reraise_exception(): [ 545.319456] env[62066]: ERROR nova.compute.manager [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 545.319871] env[62066]: ERROR nova.compute.manager [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] self.force_reraise() [ 545.319871] env[62066]: ERROR nova.compute.manager [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 545.319871] env[62066]: ERROR nova.compute.manager [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] raise self.value [ 545.319871] env[62066]: ERROR nova.compute.manager [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 545.319871] env[62066]: ERROR nova.compute.manager [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] updated_port = self._update_port( [ 545.319871] env[62066]: ERROR nova.compute.manager [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 545.319871] env[62066]: ERROR nova.compute.manager [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] _ensure_no_port_binding_failure(port) [ 545.319871] env[62066]: ERROR nova.compute.manager [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 545.319871] env[62066]: ERROR nova.compute.manager [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] raise exception.PortBindingFailed(port_id=port['id']) [ 545.319871] env[62066]: ERROR nova.compute.manager [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] nova.exception.PortBindingFailed: Binding failed for port 78745978-2782-4cbd-8fbd-fbdf247cc830, please check neutron logs for more information. [ 545.319871] env[62066]: ERROR nova.compute.manager [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] [ 545.320235] env[62066]: DEBUG nova.compute.utils [None req-0e23bf5c-5716-4eba-9510-a3597f398645 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] Binding failed for port 78745978-2782-4cbd-8fbd-fbdf247cc830, please check neutron logs for more information. {{(pid=62066) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 545.321578] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d274d4c0-1d30-44ad-b71d-9f542d291186 tempest-VolumesAssistedSnapshotsTest-1416743605 tempest-VolumesAssistedSnapshotsTest-1416743605-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.803s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 545.323300] env[62066]: INFO nova.compute.claims [None req-d274d4c0-1d30-44ad-b71d-9f542d291186 tempest-VolumesAssistedSnapshotsTest-1416743605 tempest-VolumesAssistedSnapshotsTest-1416743605-project-member] [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 545.330019] env[62066]: DEBUG nova.compute.manager [None req-0e23bf5c-5716-4eba-9510-a3597f398645 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] Build of instance 095d69fe-bbd3-4850-9dcd-b4c2290b8352 was re-scheduled: Binding failed for port 78745978-2782-4cbd-8fbd-fbdf247cc830, please check neutron logs for more information. {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 545.330019] env[62066]: DEBUG nova.compute.manager [None req-0e23bf5c-5716-4eba-9510-a3597f398645 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] Unplugging VIFs for instance {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 545.330019] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0e23bf5c-5716-4eba-9510-a3597f398645 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] Acquiring lock "refresh_cache-095d69fe-bbd3-4850-9dcd-b4c2290b8352" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 545.330019] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0e23bf5c-5716-4eba-9510-a3597f398645 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] Acquired lock "refresh_cache-095d69fe-bbd3-4850-9dcd-b4c2290b8352" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 545.330358] env[62066]: DEBUG nova.network.neutron [None req-0e23bf5c-5716-4eba-9510-a3597f398645 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 545.530499] env[62066]: INFO nova.compute.manager [None req-df85dbc4-0217-427e-9c09-82d1a0e6d774 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] [instance: 1eeb879b-6d1e-4746-98b5-12c30d674fa9] Took 1.03 seconds to deallocate network for instance. [ 545.562744] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cf333fac-d104-4122-9317-eab87c4499f0 tempest-ServerDiagnosticsTest-1737887255 tempest-ServerDiagnosticsTest-1737887255-project-member] Acquiring lock "2cf18a04-6bf9-4f47-a920-2c568207057d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 545.563036] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cf333fac-d104-4122-9317-eab87c4499f0 tempest-ServerDiagnosticsTest-1737887255 tempest-ServerDiagnosticsTest-1737887255-project-member] Lock "2cf18a04-6bf9-4f47-a920-2c568207057d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 545.852063] env[62066]: DEBUG nova.network.neutron [None req-0e23bf5c-5716-4eba-9510-a3597f398645 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 545.927609] env[62066]: DEBUG nova.network.neutron [None req-0e23bf5c-5716-4eba-9510-a3597f398645 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 546.295709] env[62066]: DEBUG oslo_concurrency.lockutils [None req-67a0ccea-4cde-4de9-9aa4-1bb67e83772e tempest-ServersNegativeTestJSON-789906455 tempest-ServersNegativeTestJSON-789906455-project-member] Acquiring lock "5bbdf138-0aaf-42d1-b5a1-4f7ebc650393" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 546.296425] env[62066]: DEBUG oslo_concurrency.lockutils [None req-67a0ccea-4cde-4de9-9aa4-1bb67e83772e tempest-ServersNegativeTestJSON-789906455 tempest-ServersNegativeTestJSON-789906455-project-member] Lock "5bbdf138-0aaf-42d1-b5a1-4f7ebc650393" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 546.433235] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0e23bf5c-5716-4eba-9510-a3597f398645 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] Releasing lock "refresh_cache-095d69fe-bbd3-4850-9dcd-b4c2290b8352" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 546.433235] env[62066]: DEBUG nova.compute.manager [None req-0e23bf5c-5716-4eba-9510-a3597f398645 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 546.433235] env[62066]: DEBUG nova.compute.manager [None req-0e23bf5c-5716-4eba-9510-a3597f398645 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 546.433235] env[62066]: DEBUG nova.network.neutron [None req-0e23bf5c-5716-4eba-9510-a3597f398645 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 546.466453] env[62066]: DEBUG nova.network.neutron [None req-0e23bf5c-5716-4eba-9510-a3597f398645 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 546.565782] env[62066]: INFO nova.scheduler.client.report [None req-df85dbc4-0217-427e-9c09-82d1a0e6d774 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] Deleted allocations for instance 1eeb879b-6d1e-4746-98b5-12c30d674fa9 [ 546.937414] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b341c94a-a0bd-4a9e-9047-a863d8544522 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.948722] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be93876f-24a2-4130-ba7e-e305265da067 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.987208] env[62066]: DEBUG nova.network.neutron [None req-0e23bf5c-5716-4eba-9510-a3597f398645 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 546.988546] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30f8d63e-8021-4afe-8fc5-b32a550373d2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.996295] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36da7366-caa3-4f04-85e4-a77ed486c593 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.010716] env[62066]: DEBUG nova.compute.provider_tree [None req-d274d4c0-1d30-44ad-b71d-9f542d291186 tempest-VolumesAssistedSnapshotsTest-1416743605 tempest-VolumesAssistedSnapshotsTest-1416743605-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 547.080601] env[62066]: DEBUG oslo_concurrency.lockutils [None req-df85dbc4-0217-427e-9c09-82d1a0e6d774 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] Lock "1eeb879b-6d1e-4746-98b5-12c30d674fa9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 50.576s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 547.489712] env[62066]: INFO nova.compute.manager [None req-0e23bf5c-5716-4eba-9510-a3597f398645 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] [instance: 095d69fe-bbd3-4850-9dcd-b4c2290b8352] Took 1.06 seconds to deallocate network for instance. [ 547.514523] env[62066]: DEBUG nova.scheduler.client.report [None req-d274d4c0-1d30-44ad-b71d-9f542d291186 tempest-VolumesAssistedSnapshotsTest-1416743605 tempest-VolumesAssistedSnapshotsTest-1416743605-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 547.539140] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b96d7111-6e60-4fd2-a38a-3863242412fb tempest-ServerActionsTestOtherB-426172427 tempest-ServerActionsTestOtherB-426172427-project-member] Acquiring lock "e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 547.539240] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b96d7111-6e60-4fd2-a38a-3863242412fb tempest-ServerActionsTestOtherB-426172427 tempest-ServerActionsTestOtherB-426172427-project-member] Lock "e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 547.586167] env[62066]: DEBUG nova.compute.manager [None req-4ae0c9b4-adca-4b03-8ac7-409a4ee035f6 tempest-InstanceActionsTestJSON-1008103477 tempest-InstanceActionsTestJSON-1008103477-project-member] [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 548.020598] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d274d4c0-1d30-44ad-b71d-9f542d291186 tempest-VolumesAssistedSnapshotsTest-1416743605 tempest-VolumesAssistedSnapshotsTest-1416743605-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.699s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 548.021015] env[62066]: DEBUG nova.compute.manager [None req-d274d4c0-1d30-44ad-b71d-9f542d291186 tempest-VolumesAssistedSnapshotsTest-1416743605 tempest-VolumesAssistedSnapshotsTest-1416743605-project-member] [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 548.024341] env[62066]: DEBUG oslo_concurrency.lockutils [None req-70656e85-7c6a-408b-a415-8e3cff7746a0 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 22.921s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 548.024580] env[62066]: DEBUG nova.objects.instance [None req-70656e85-7c6a-408b-a415-8e3cff7746a0 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] [instance: 56eee633-2c01-4df2-8d2e-58ad712942f6] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62066) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 548.111697] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4ae0c9b4-adca-4b03-8ac7-409a4ee035f6 tempest-InstanceActionsTestJSON-1008103477 tempest-InstanceActionsTestJSON-1008103477-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 548.524380] env[62066]: INFO nova.scheduler.client.report [None req-0e23bf5c-5716-4eba-9510-a3597f398645 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] Deleted allocations for instance 095d69fe-bbd3-4850-9dcd-b4c2290b8352 [ 548.531754] env[62066]: DEBUG nova.compute.utils [None req-d274d4c0-1d30-44ad-b71d-9f542d291186 tempest-VolumesAssistedSnapshotsTest-1416743605 tempest-VolumesAssistedSnapshotsTest-1416743605-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 548.538371] env[62066]: DEBUG nova.compute.manager [None req-d274d4c0-1d30-44ad-b71d-9f542d291186 tempest-VolumesAssistedSnapshotsTest-1416743605 tempest-VolumesAssistedSnapshotsTest-1416743605-project-member] [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 548.539031] env[62066]: DEBUG nova.network.neutron [None req-d274d4c0-1d30-44ad-b71d-9f542d291186 tempest-VolumesAssistedSnapshotsTest-1416743605 tempest-VolumesAssistedSnapshotsTest-1416743605-project-member] [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 548.612809] env[62066]: DEBUG nova.policy [None req-d274d4c0-1d30-44ad-b71d-9f542d291186 tempest-VolumesAssistedSnapshotsTest-1416743605 tempest-VolumesAssistedSnapshotsTest-1416743605-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '691db1fba48d4a3c8c263dbdbf986aad', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e4cbb7498972432ea832c243e1493f2d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 549.039724] env[62066]: DEBUG nova.compute.manager [None req-d274d4c0-1d30-44ad-b71d-9f542d291186 tempest-VolumesAssistedSnapshotsTest-1416743605 tempest-VolumesAssistedSnapshotsTest-1416743605-project-member] [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 549.048026] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0e23bf5c-5716-4eba-9510-a3597f398645 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] Lock "095d69fe-bbd3-4850-9dcd-b4c2290b8352" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 51.709s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 549.048026] env[62066]: DEBUG oslo_concurrency.lockutils [None req-70656e85-7c6a-408b-a415-8e3cff7746a0 tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.022s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 549.048026] env[62066]: DEBUG oslo_concurrency.lockutils [None req-68481586-e4e9-4c46-8b51-aa7845aca824 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 23.881s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 549.092718] env[62066]: DEBUG nova.network.neutron [None req-d274d4c0-1d30-44ad-b71d-9f542d291186 tempest-VolumesAssistedSnapshotsTest-1416743605 tempest-VolumesAssistedSnapshotsTest-1416743605-project-member] [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] Successfully created port: 4bb5d522-c456-4fd5-babb-0d4606b2df97 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 549.555749] env[62066]: DEBUG nova.compute.manager [None req-cba923da-ec24-4fd6-9695-23f810adaa32 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 550.052125] env[62066]: DEBUG nova.compute.manager [None req-d274d4c0-1d30-44ad-b71d-9f542d291186 tempest-VolumesAssistedSnapshotsTest-1416743605 tempest-VolumesAssistedSnapshotsTest-1416743605-project-member] [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 550.086012] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c3bbd5c-ceb7-420b-a533-6a82ebb27faf {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.089776] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cba923da-ec24-4fd6-9695-23f810adaa32 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 550.094987] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-647f96b4-ac1e-4db6-8722-6bb9e7de3eb2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.100082] env[62066]: DEBUG nova.virt.hardware [None req-d274d4c0-1d30-44ad-b71d-9f542d291186 tempest-VolumesAssistedSnapshotsTest-1416743605 tempest-VolumesAssistedSnapshotsTest-1416743605-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-23T13:40:41Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-23T13:40:23Z,direct_url=,disk_format='vmdk',id=50ff584c-3b50-4395-af07-3e66769bc9f7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='eb52f7069a374c61ae946f052007c6d9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-23T13:40:24Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 550.100329] env[62066]: DEBUG nova.virt.hardware [None req-d274d4c0-1d30-44ad-b71d-9f542d291186 tempest-VolumesAssistedSnapshotsTest-1416743605 tempest-VolumesAssistedSnapshotsTest-1416743605-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 550.100677] env[62066]: DEBUG nova.virt.hardware [None req-d274d4c0-1d30-44ad-b71d-9f542d291186 tempest-VolumesAssistedSnapshotsTest-1416743605 tempest-VolumesAssistedSnapshotsTest-1416743605-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 550.101110] env[62066]: DEBUG nova.virt.hardware [None req-d274d4c0-1d30-44ad-b71d-9f542d291186 tempest-VolumesAssistedSnapshotsTest-1416743605 tempest-VolumesAssistedSnapshotsTest-1416743605-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 550.103417] env[62066]: DEBUG nova.virt.hardware [None req-d274d4c0-1d30-44ad-b71d-9f542d291186 tempest-VolumesAssistedSnapshotsTest-1416743605 tempest-VolumesAssistedSnapshotsTest-1416743605-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 550.103417] env[62066]: DEBUG nova.virt.hardware [None req-d274d4c0-1d30-44ad-b71d-9f542d291186 tempest-VolumesAssistedSnapshotsTest-1416743605 tempest-VolumesAssistedSnapshotsTest-1416743605-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 550.103417] env[62066]: DEBUG nova.virt.hardware [None req-d274d4c0-1d30-44ad-b71d-9f542d291186 tempest-VolumesAssistedSnapshotsTest-1416743605 tempest-VolumesAssistedSnapshotsTest-1416743605-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 550.103417] env[62066]: DEBUG nova.virt.hardware [None req-d274d4c0-1d30-44ad-b71d-9f542d291186 tempest-VolumesAssistedSnapshotsTest-1416743605 tempest-VolumesAssistedSnapshotsTest-1416743605-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 550.103417] env[62066]: DEBUG nova.virt.hardware [None req-d274d4c0-1d30-44ad-b71d-9f542d291186 tempest-VolumesAssistedSnapshotsTest-1416743605 tempest-VolumesAssistedSnapshotsTest-1416743605-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 550.103698] env[62066]: DEBUG nova.virt.hardware [None req-d274d4c0-1d30-44ad-b71d-9f542d291186 tempest-VolumesAssistedSnapshotsTest-1416743605 tempest-VolumesAssistedSnapshotsTest-1416743605-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 550.103698] env[62066]: DEBUG nova.virt.hardware [None req-d274d4c0-1d30-44ad-b71d-9f542d291186 tempest-VolumesAssistedSnapshotsTest-1416743605 tempest-VolumesAssistedSnapshotsTest-1416743605-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 550.103698] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f79bd66-dccf-4fbf-81b0-6ee7565c7ee6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.139145] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-222b6dda-e351-421f-81de-f5061993cd25 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.144447] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b065685-9538-4ab5-954b-c00bd178c4e9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.161412] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01082c51-4b6f-4d8a-a179-f978d90a7285 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.175767] env[62066]: DEBUG nova.compute.provider_tree [None req-68481586-e4e9-4c46-8b51-aa7845aca824 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 550.681101] env[62066]: DEBUG nova.scheduler.client.report [None req-68481586-e4e9-4c46-8b51-aa7845aca824 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 551.183258] env[62066]: ERROR nova.compute.manager [None req-d274d4c0-1d30-44ad-b71d-9f542d291186 tempest-VolumesAssistedSnapshotsTest-1416743605 tempest-VolumesAssistedSnapshotsTest-1416743605-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 4bb5d522-c456-4fd5-babb-0d4606b2df97, please check neutron logs for more information. [ 551.183258] env[62066]: ERROR nova.compute.manager Traceback (most recent call last): [ 551.183258] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 551.183258] env[62066]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 551.183258] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 551.183258] env[62066]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 551.183258] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 551.183258] env[62066]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 551.183258] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 551.183258] env[62066]: ERROR nova.compute.manager self.force_reraise() [ 551.183258] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 551.183258] env[62066]: ERROR nova.compute.manager raise self.value [ 551.183258] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 551.183258] env[62066]: ERROR nova.compute.manager updated_port = self._update_port( [ 551.183258] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 551.183258] env[62066]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 551.183757] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 551.183757] env[62066]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 551.183757] env[62066]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 4bb5d522-c456-4fd5-babb-0d4606b2df97, please check neutron logs for more information. [ 551.183757] env[62066]: ERROR nova.compute.manager [ 551.183757] env[62066]: Traceback (most recent call last): [ 551.183757] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 551.183757] env[62066]: listener.cb(fileno) [ 551.183757] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 551.183757] env[62066]: result = function(*args, **kwargs) [ 551.183757] env[62066]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 551.183757] env[62066]: return func(*args, **kwargs) [ 551.183757] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 551.183757] env[62066]: raise e [ 551.183757] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 551.183757] env[62066]: nwinfo = self.network_api.allocate_for_instance( [ 551.183757] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 551.183757] env[62066]: created_port_ids = self._update_ports_for_instance( [ 551.183757] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 551.183757] env[62066]: with excutils.save_and_reraise_exception(): [ 551.183757] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 551.183757] env[62066]: self.force_reraise() [ 551.183757] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 551.183757] env[62066]: raise self.value [ 551.183757] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 551.183757] env[62066]: updated_port = self._update_port( [ 551.183757] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 551.183757] env[62066]: _ensure_no_port_binding_failure(port) [ 551.183757] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 551.183757] env[62066]: raise exception.PortBindingFailed(port_id=port['id']) [ 551.187419] env[62066]: nova.exception.PortBindingFailed: Binding failed for port 4bb5d522-c456-4fd5-babb-0d4606b2df97, please check neutron logs for more information. [ 551.187419] env[62066]: Removing descriptor: 20 [ 551.187419] env[62066]: DEBUG oslo_concurrency.lockutils [None req-68481586-e4e9-4c46-8b51-aa7845aca824 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.136s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 551.187419] env[62066]: ERROR nova.compute.manager [None req-68481586-e4e9-4c46-8b51-aa7845aca824 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 5d597d92-a2ed-414a-babb-5eacc2e2e5c2, please check neutron logs for more information. [ 551.187419] env[62066]: ERROR nova.compute.manager [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] Traceback (most recent call last): [ 551.187419] env[62066]: ERROR nova.compute.manager [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 551.187419] env[62066]: ERROR nova.compute.manager [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] self.driver.spawn(context, instance, image_meta, [ 551.187419] env[62066]: ERROR nova.compute.manager [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 551.187419] env[62066]: ERROR nova.compute.manager [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] self._vmops.spawn(context, instance, image_meta, injected_files, [ 551.187419] env[62066]: ERROR nova.compute.manager [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 551.187886] env[62066]: ERROR nova.compute.manager [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] vm_ref = self.build_virtual_machine(instance, [ 551.187886] env[62066]: ERROR nova.compute.manager [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 551.187886] env[62066]: ERROR nova.compute.manager [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] vif_infos = vmwarevif.get_vif_info(self._session, [ 551.187886] env[62066]: ERROR nova.compute.manager [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 551.187886] env[62066]: ERROR nova.compute.manager [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] for vif in network_info: [ 551.187886] env[62066]: ERROR nova.compute.manager [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 551.187886] env[62066]: ERROR nova.compute.manager [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] return self._sync_wrapper(fn, *args, **kwargs) [ 551.187886] env[62066]: ERROR nova.compute.manager [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 551.187886] env[62066]: ERROR nova.compute.manager [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] self.wait() [ 551.187886] env[62066]: ERROR nova.compute.manager [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 551.187886] env[62066]: ERROR nova.compute.manager [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] self[:] = self._gt.wait() [ 551.187886] env[62066]: ERROR nova.compute.manager [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 551.187886] env[62066]: ERROR nova.compute.manager [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] return self._exit_event.wait() [ 551.188298] env[62066]: ERROR nova.compute.manager [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 551.188298] env[62066]: ERROR nova.compute.manager [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] result = hub.switch() [ 551.188298] env[62066]: ERROR nova.compute.manager [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 551.188298] env[62066]: ERROR nova.compute.manager [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] return self.greenlet.switch() [ 551.188298] env[62066]: ERROR nova.compute.manager [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 551.188298] env[62066]: ERROR nova.compute.manager [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] result = function(*args, **kwargs) [ 551.188298] env[62066]: ERROR nova.compute.manager [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 551.188298] env[62066]: ERROR nova.compute.manager [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] return func(*args, **kwargs) [ 551.188298] env[62066]: ERROR nova.compute.manager [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 551.188298] env[62066]: ERROR nova.compute.manager [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] raise e [ 551.188298] env[62066]: ERROR nova.compute.manager [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 551.188298] env[62066]: ERROR nova.compute.manager [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] nwinfo = self.network_api.allocate_for_instance( [ 551.188298] env[62066]: ERROR nova.compute.manager [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 551.188766] env[62066]: ERROR nova.compute.manager [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] created_port_ids = self._update_ports_for_instance( [ 551.188766] env[62066]: ERROR nova.compute.manager [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 551.188766] env[62066]: ERROR nova.compute.manager [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] with excutils.save_and_reraise_exception(): [ 551.188766] env[62066]: ERROR nova.compute.manager [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 551.188766] env[62066]: ERROR nova.compute.manager [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] self.force_reraise() [ 551.188766] env[62066]: ERROR nova.compute.manager [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 551.188766] env[62066]: ERROR nova.compute.manager [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] raise self.value [ 551.188766] env[62066]: ERROR nova.compute.manager [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 551.188766] env[62066]: ERROR nova.compute.manager [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] updated_port = self._update_port( [ 551.188766] env[62066]: ERROR nova.compute.manager [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 551.188766] env[62066]: ERROR nova.compute.manager [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] _ensure_no_port_binding_failure(port) [ 551.188766] env[62066]: ERROR nova.compute.manager [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 551.188766] env[62066]: ERROR nova.compute.manager [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] raise exception.PortBindingFailed(port_id=port['id']) [ 551.189252] env[62066]: ERROR nova.compute.manager [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] nova.exception.PortBindingFailed: Binding failed for port 5d597d92-a2ed-414a-babb-5eacc2e2e5c2, please check neutron logs for more information. [ 551.189252] env[62066]: ERROR nova.compute.manager [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] [ 551.189252] env[62066]: DEBUG nova.compute.utils [None req-68481586-e4e9-4c46-8b51-aa7845aca824 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] Binding failed for port 5d597d92-a2ed-414a-babb-5eacc2e2e5c2, please check neutron logs for more information. {{(pid=62066) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 551.189252] env[62066]: ERROR nova.compute.manager [None req-d274d4c0-1d30-44ad-b71d-9f542d291186 tempest-VolumesAssistedSnapshotsTest-1416743605 tempest-VolumesAssistedSnapshotsTest-1416743605-project-member] [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 4bb5d522-c456-4fd5-babb-0d4606b2df97, please check neutron logs for more information. [ 551.189252] env[62066]: ERROR nova.compute.manager [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] Traceback (most recent call last): [ 551.189252] env[62066]: ERROR nova.compute.manager [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 551.189252] env[62066]: ERROR nova.compute.manager [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] yield resources [ 551.189252] env[62066]: ERROR nova.compute.manager [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 551.189252] env[62066]: ERROR nova.compute.manager [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] self.driver.spawn(context, instance, image_meta, [ 551.189533] env[62066]: ERROR nova.compute.manager [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 551.189533] env[62066]: ERROR nova.compute.manager [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] self._vmops.spawn(context, instance, image_meta, injected_files, [ 551.189533] env[62066]: ERROR nova.compute.manager [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 551.189533] env[62066]: ERROR nova.compute.manager [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] vm_ref = self.build_virtual_machine(instance, [ 551.189533] env[62066]: ERROR nova.compute.manager [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 551.189533] env[62066]: ERROR nova.compute.manager [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] vif_infos = vmwarevif.get_vif_info(self._session, [ 551.189533] env[62066]: ERROR nova.compute.manager [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 551.189533] env[62066]: ERROR nova.compute.manager [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] for vif in network_info: [ 551.189533] env[62066]: ERROR nova.compute.manager [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 551.189533] env[62066]: ERROR nova.compute.manager [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] return self._sync_wrapper(fn, *args, **kwargs) [ 551.189533] env[62066]: ERROR nova.compute.manager [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 551.189533] env[62066]: ERROR nova.compute.manager [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] self.wait() [ 551.189533] env[62066]: ERROR nova.compute.manager [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 551.190283] env[62066]: ERROR nova.compute.manager [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] self[:] = self._gt.wait() [ 551.190283] env[62066]: ERROR nova.compute.manager [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 551.190283] env[62066]: ERROR nova.compute.manager [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] return self._exit_event.wait() [ 551.190283] env[62066]: ERROR nova.compute.manager [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 551.190283] env[62066]: ERROR nova.compute.manager [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] result = hub.switch() [ 551.190283] env[62066]: ERROR nova.compute.manager [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 551.190283] env[62066]: ERROR nova.compute.manager [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] return self.greenlet.switch() [ 551.190283] env[62066]: ERROR nova.compute.manager [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 551.190283] env[62066]: ERROR nova.compute.manager [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] result = function(*args, **kwargs) [ 551.190283] env[62066]: ERROR nova.compute.manager [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 551.190283] env[62066]: ERROR nova.compute.manager [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] return func(*args, **kwargs) [ 551.190283] env[62066]: ERROR nova.compute.manager [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 551.190283] env[62066]: ERROR nova.compute.manager [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] raise e [ 551.190772] env[62066]: ERROR nova.compute.manager [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 551.190772] env[62066]: ERROR nova.compute.manager [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] nwinfo = self.network_api.allocate_for_instance( [ 551.190772] env[62066]: ERROR nova.compute.manager [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 551.190772] env[62066]: ERROR nova.compute.manager [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] created_port_ids = self._update_ports_for_instance( [ 551.190772] env[62066]: ERROR nova.compute.manager [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 551.190772] env[62066]: ERROR nova.compute.manager [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] with excutils.save_and_reraise_exception(): [ 551.190772] env[62066]: ERROR nova.compute.manager [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 551.190772] env[62066]: ERROR nova.compute.manager [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] self.force_reraise() [ 551.190772] env[62066]: ERROR nova.compute.manager [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 551.190772] env[62066]: ERROR nova.compute.manager [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] raise self.value [ 551.190772] env[62066]: ERROR nova.compute.manager [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 551.190772] env[62066]: ERROR nova.compute.manager [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] updated_port = self._update_port( [ 551.190772] env[62066]: ERROR nova.compute.manager [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 551.191231] env[62066]: ERROR nova.compute.manager [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] _ensure_no_port_binding_failure(port) [ 551.191231] env[62066]: ERROR nova.compute.manager [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 551.191231] env[62066]: ERROR nova.compute.manager [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] raise exception.PortBindingFailed(port_id=port['id']) [ 551.191231] env[62066]: ERROR nova.compute.manager [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] nova.exception.PortBindingFailed: Binding failed for port 4bb5d522-c456-4fd5-babb-0d4606b2df97, please check neutron logs for more information. [ 551.191231] env[62066]: ERROR nova.compute.manager [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] [ 551.191231] env[62066]: INFO nova.compute.manager [None req-d274d4c0-1d30-44ad-b71d-9f542d291186 tempest-VolumesAssistedSnapshotsTest-1416743605 tempest-VolumesAssistedSnapshotsTest-1416743605-project-member] [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] Terminating instance [ 551.191496] env[62066]: DEBUG nova.compute.manager [None req-68481586-e4e9-4c46-8b51-aa7845aca824 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] Build of instance d522ee4a-9bd9-402b-b3bd-c36f93366249 was re-scheduled: Binding failed for port 5d597d92-a2ed-414a-babb-5eacc2e2e5c2, please check neutron logs for more information. {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 551.193043] env[62066]: DEBUG nova.compute.manager [None req-68481586-e4e9-4c46-8b51-aa7845aca824 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] Unplugging VIFs for instance {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 551.193043] env[62066]: DEBUG oslo_concurrency.lockutils [None req-68481586-e4e9-4c46-8b51-aa7845aca824 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] Acquiring lock "refresh_cache-d522ee4a-9bd9-402b-b3bd-c36f93366249" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 551.193043] env[62066]: DEBUG oslo_concurrency.lockutils [None req-68481586-e4e9-4c46-8b51-aa7845aca824 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] Acquired lock "refresh_cache-d522ee4a-9bd9-402b-b3bd-c36f93366249" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 551.193043] env[62066]: DEBUG nova.network.neutron [None req-68481586-e4e9-4c46-8b51-aa7845aca824 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 551.194169] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c4a0feae-1b12-4166-bd78-82a96fe8ac4c tempest-ServersWithSpecificFlavorTestJSON-184391817 tempest-ServersWithSpecificFlavorTestJSON-184391817-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.767s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 551.199938] env[62066]: INFO nova.compute.claims [None req-c4a0feae-1b12-4166-bd78-82a96fe8ac4c tempest-ServersWithSpecificFlavorTestJSON-184391817 tempest-ServersWithSpecificFlavorTestJSON-184391817-project-member] [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 551.217785] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d274d4c0-1d30-44ad-b71d-9f542d291186 tempest-VolumesAssistedSnapshotsTest-1416743605 tempest-VolumesAssistedSnapshotsTest-1416743605-project-member] Acquiring lock "refresh_cache-1fde7ae0-c88c-4d7f-b657-355c20e2cb10" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 551.218670] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d274d4c0-1d30-44ad-b71d-9f542d291186 tempest-VolumesAssistedSnapshotsTest-1416743605 tempest-VolumesAssistedSnapshotsTest-1416743605-project-member] Acquired lock "refresh_cache-1fde7ae0-c88c-4d7f-b657-355c20e2cb10" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 551.218670] env[62066]: DEBUG nova.network.neutron [None req-d274d4c0-1d30-44ad-b71d-9f542d291186 tempest-VolumesAssistedSnapshotsTest-1416743605 tempest-VolumesAssistedSnapshotsTest-1416743605-project-member] [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 551.452741] env[62066]: DEBUG nova.compute.manager [req-96ba6997-b312-4f77-92dc-503d5d8b5097 req-f31ad177-c26b-43b4-9999-94f9e8b25dba service nova] [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] Received event network-changed-4bb5d522-c456-4fd5-babb-0d4606b2df97 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 551.452741] env[62066]: DEBUG nova.compute.manager [req-96ba6997-b312-4f77-92dc-503d5d8b5097 req-f31ad177-c26b-43b4-9999-94f9e8b25dba service nova] [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] Refreshing instance network info cache due to event network-changed-4bb5d522-c456-4fd5-babb-0d4606b2df97. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 551.452741] env[62066]: DEBUG oslo_concurrency.lockutils [req-96ba6997-b312-4f77-92dc-503d5d8b5097 req-f31ad177-c26b-43b4-9999-94f9e8b25dba service nova] Acquiring lock "refresh_cache-1fde7ae0-c88c-4d7f-b657-355c20e2cb10" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 551.739224] env[62066]: DEBUG nova.network.neutron [None req-68481586-e4e9-4c46-8b51-aa7845aca824 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 551.741922] env[62066]: DEBUG nova.network.neutron [None req-d274d4c0-1d30-44ad-b71d-9f542d291186 tempest-VolumesAssistedSnapshotsTest-1416743605 tempest-VolumesAssistedSnapshotsTest-1416743605-project-member] [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 551.857181] env[62066]: DEBUG nova.network.neutron [None req-d274d4c0-1d30-44ad-b71d-9f542d291186 tempest-VolumesAssistedSnapshotsTest-1416743605 tempest-VolumesAssistedSnapshotsTest-1416743605-project-member] [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 551.906282] env[62066]: DEBUG nova.network.neutron [None req-68481586-e4e9-4c46-8b51-aa7845aca824 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 552.364795] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d274d4c0-1d30-44ad-b71d-9f542d291186 tempest-VolumesAssistedSnapshotsTest-1416743605 tempest-VolumesAssistedSnapshotsTest-1416743605-project-member] Releasing lock "refresh_cache-1fde7ae0-c88c-4d7f-b657-355c20e2cb10" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 552.364795] env[62066]: DEBUG nova.compute.manager [None req-d274d4c0-1d30-44ad-b71d-9f542d291186 tempest-VolumesAssistedSnapshotsTest-1416743605 tempest-VolumesAssistedSnapshotsTest-1416743605-project-member] [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 552.364948] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-d274d4c0-1d30-44ad-b71d-9f542d291186 tempest-VolumesAssistedSnapshotsTest-1416743605 tempest-VolumesAssistedSnapshotsTest-1416743605-project-member] [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 552.365615] env[62066]: DEBUG oslo_concurrency.lockutils [req-96ba6997-b312-4f77-92dc-503d5d8b5097 req-f31ad177-c26b-43b4-9999-94f9e8b25dba service nova] Acquired lock "refresh_cache-1fde7ae0-c88c-4d7f-b657-355c20e2cb10" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 552.365847] env[62066]: DEBUG nova.network.neutron [req-96ba6997-b312-4f77-92dc-503d5d8b5097 req-f31ad177-c26b-43b4-9999-94f9e8b25dba service nova] [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] Refreshing network info cache for port 4bb5d522-c456-4fd5-babb-0d4606b2df97 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 552.366934] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d3ec22b9-66e0-47ae-9adb-f305a55b8e41 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.381077] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24f3de5d-4214-4ab0-9925-514f30e330fd {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.409208] env[62066]: DEBUG oslo_concurrency.lockutils [None req-68481586-e4e9-4c46-8b51-aa7845aca824 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] Releasing lock "refresh_cache-d522ee4a-9bd9-402b-b3bd-c36f93366249" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 552.409536] env[62066]: DEBUG nova.compute.manager [None req-68481586-e4e9-4c46-8b51-aa7845aca824 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 552.409767] env[62066]: DEBUG nova.compute.manager [None req-68481586-e4e9-4c46-8b51-aa7845aca824 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 552.409978] env[62066]: DEBUG nova.network.neutron [None req-68481586-e4e9-4c46-8b51-aa7845aca824 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 552.411912] env[62066]: WARNING nova.virt.vmwareapi.vmops [None req-d274d4c0-1d30-44ad-b71d-9f542d291186 tempest-VolumesAssistedSnapshotsTest-1416743605 tempest-VolumesAssistedSnapshotsTest-1416743605-project-member] [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 1fde7ae0-c88c-4d7f-b657-355c20e2cb10 could not be found. [ 552.412177] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-d274d4c0-1d30-44ad-b71d-9f542d291186 tempest-VolumesAssistedSnapshotsTest-1416743605 tempest-VolumesAssistedSnapshotsTest-1416743605-project-member] [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 552.412738] env[62066]: INFO nova.compute.manager [None req-d274d4c0-1d30-44ad-b71d-9f542d291186 tempest-VolumesAssistedSnapshotsTest-1416743605 tempest-VolumesAssistedSnapshotsTest-1416743605-project-member] [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] Took 0.05 seconds to destroy the instance on the hypervisor. [ 552.412738] env[62066]: DEBUG oslo.service.loopingcall [None req-d274d4c0-1d30-44ad-b71d-9f542d291186 tempest-VolumesAssistedSnapshotsTest-1416743605 tempest-VolumesAssistedSnapshotsTest-1416743605-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 552.415506] env[62066]: DEBUG nova.compute.manager [-] [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 552.415643] env[62066]: DEBUG nova.network.neutron [-] [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 552.438862] env[62066]: DEBUG nova.network.neutron [-] [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 552.441287] env[62066]: DEBUG nova.network.neutron [None req-68481586-e4e9-4c46-8b51-aa7845aca824 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 552.775398] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5adc03fa-3be7-4a38-8639-bf65725e9920 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.784959] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d423ea3d-609e-4a6e-a296-0d7e8a6cb8fe {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.817953] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e98dc00-8bea-4f10-a538-d1bb9e74e3c5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.825926] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1910fffb-3dc9-4e59-b4b7-95659857152e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.843192] env[62066]: DEBUG nova.compute.provider_tree [None req-c4a0feae-1b12-4166-bd78-82a96fe8ac4c tempest-ServersWithSpecificFlavorTestJSON-184391817 tempest-ServersWithSpecificFlavorTestJSON-184391817-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 552.897338] env[62066]: DEBUG nova.network.neutron [req-96ba6997-b312-4f77-92dc-503d5d8b5097 req-f31ad177-c26b-43b4-9999-94f9e8b25dba service nova] [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 552.946531] env[62066]: DEBUG nova.network.neutron [None req-68481586-e4e9-4c46-8b51-aa7845aca824 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 552.947489] env[62066]: DEBUG nova.network.neutron [-] [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 552.948479] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3515d98a-2869-41bb-bf8e-e38d995a4d22 tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Acquiring lock "e795e53e-af33-4d0c-8dbd-9c59f236acd7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 552.948710] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3515d98a-2869-41bb-bf8e-e38d995a4d22 tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Lock "e795e53e-af33-4d0c-8dbd-9c59f236acd7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 553.044391] env[62066]: DEBUG nova.network.neutron [req-96ba6997-b312-4f77-92dc-503d5d8b5097 req-f31ad177-c26b-43b4-9999-94f9e8b25dba service nova] [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 553.190339] env[62066]: DEBUG oslo_concurrency.lockutils [None req-60a20a80-792e-4a03-84da-c17f8b28425d tempest-InstanceActionsNegativeTestJSON-1895685779 tempest-InstanceActionsNegativeTestJSON-1895685779-project-member] Acquiring lock "a690ad15-2878-4fa7-a810-d155ce6b1a8c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 553.190339] env[62066]: DEBUG oslo_concurrency.lockutils [None req-60a20a80-792e-4a03-84da-c17f8b28425d tempest-InstanceActionsNegativeTestJSON-1895685779 tempest-InstanceActionsNegativeTestJSON-1895685779-project-member] Lock "a690ad15-2878-4fa7-a810-d155ce6b1a8c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 553.349235] env[62066]: DEBUG nova.scheduler.client.report [None req-c4a0feae-1b12-4166-bd78-82a96fe8ac4c tempest-ServersWithSpecificFlavorTestJSON-184391817 tempest-ServersWithSpecificFlavorTestJSON-184391817-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 553.450104] env[62066]: INFO nova.compute.manager [-] [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] Took 1.03 seconds to deallocate network for instance. [ 553.450696] env[62066]: INFO nova.compute.manager [None req-68481586-e4e9-4c46-8b51-aa7845aca824 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] [instance: d522ee4a-9bd9-402b-b3bd-c36f93366249] Took 1.04 seconds to deallocate network for instance. [ 553.462716] env[62066]: DEBUG nova.compute.claims [None req-d274d4c0-1d30-44ad-b71d-9f542d291186 tempest-VolumesAssistedSnapshotsTest-1416743605 tempest-VolumesAssistedSnapshotsTest-1416743605-project-member] [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] Aborting claim: {{(pid=62066) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 553.462716] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d274d4c0-1d30-44ad-b71d-9f542d291186 tempest-VolumesAssistedSnapshotsTest-1416743605 tempest-VolumesAssistedSnapshotsTest-1416743605-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 553.483106] env[62066]: DEBUG nova.compute.manager [req-97d72caa-8368-461e-9ddb-1ec7c43e9c55 req-73077fbf-d600-4b7a-9482-1f2a3e2db334 service nova] [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] Received event network-vif-deleted-4bb5d522-c456-4fd5-babb-0d4606b2df97 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 553.546986] env[62066]: DEBUG oslo_concurrency.lockutils [req-96ba6997-b312-4f77-92dc-503d5d8b5097 req-f31ad177-c26b-43b4-9999-94f9e8b25dba service nova] Releasing lock "refresh_cache-1fde7ae0-c88c-4d7f-b657-355c20e2cb10" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 553.852906] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c4a0feae-1b12-4166-bd78-82a96fe8ac4c tempest-ServersWithSpecificFlavorTestJSON-184391817 tempest-ServersWithSpecificFlavorTestJSON-184391817-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.659s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 553.854977] env[62066]: DEBUG nova.compute.manager [None req-c4a0feae-1b12-4166-bd78-82a96fe8ac4c tempest-ServersWithSpecificFlavorTestJSON-184391817 tempest-ServersWithSpecificFlavorTestJSON-184391817-project-member] [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 553.857337] env[62066]: DEBUG oslo_concurrency.lockutils [None req-99d923d5-ef6f-43f2-959c-4b252ac1d398 tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 26.711s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 554.351272] env[62066]: DEBUG oslo_concurrency.lockutils [None req-19be2ac0-adbe-403f-ab5a-6d1895be1574 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] Acquiring lock "5fad2258-f694-4993-bf64-bfa16abc09ba" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 554.351272] env[62066]: DEBUG oslo_concurrency.lockutils [None req-19be2ac0-adbe-403f-ab5a-6d1895be1574 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] Lock "5fad2258-f694-4993-bf64-bfa16abc09ba" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.003s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 554.366843] env[62066]: DEBUG nova.compute.utils [None req-c4a0feae-1b12-4166-bd78-82a96fe8ac4c tempest-ServersWithSpecificFlavorTestJSON-184391817 tempest-ServersWithSpecificFlavorTestJSON-184391817-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 554.369058] env[62066]: DEBUG nova.compute.manager [None req-c4a0feae-1b12-4166-bd78-82a96fe8ac4c tempest-ServersWithSpecificFlavorTestJSON-184391817 tempest-ServersWithSpecificFlavorTestJSON-184391817-project-member] [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 554.369412] env[62066]: DEBUG nova.network.neutron [None req-c4a0feae-1b12-4166-bd78-82a96fe8ac4c tempest-ServersWithSpecificFlavorTestJSON-184391817 tempest-ServersWithSpecificFlavorTestJSON-184391817-project-member] [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 554.453656] env[62066]: DEBUG nova.policy [None req-c4a0feae-1b12-4166-bd78-82a96fe8ac4c tempest-ServersWithSpecificFlavorTestJSON-184391817 tempest-ServersWithSpecificFlavorTestJSON-184391817-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1a2e2d88d2074ec79aa1387a246e53bb', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f8e6ed0e9a9348df93e655302f8bd178', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 554.498617] env[62066]: INFO nova.scheduler.client.report [None req-68481586-e4e9-4c46-8b51-aa7845aca824 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] Deleted allocations for instance d522ee4a-9bd9-402b-b3bd-c36f93366249 [ 554.875287] env[62066]: DEBUG nova.compute.manager [None req-c4a0feae-1b12-4166-bd78-82a96fe8ac4c tempest-ServersWithSpecificFlavorTestJSON-184391817 tempest-ServersWithSpecificFlavorTestJSON-184391817-project-member] [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 554.903493] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f99f3c45-441a-411f-8b1f-2a207c3e9204 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.913202] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-419ba12a-fe1f-433b-a69b-608019b52846 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.949293] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63a1df75-7d68-4e3e-8bf5-115c096f203b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.958204] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82e73382-a858-4f9b-8c65-f1976463c490 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.972240] env[62066]: DEBUG nova.compute.provider_tree [None req-99d923d5-ef6f-43f2-959c-4b252ac1d398 tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 555.014617] env[62066]: DEBUG oslo_concurrency.lockutils [None req-68481586-e4e9-4c46-8b51-aa7845aca824 tempest-ServersAdminTestJSON-830117193 tempest-ServersAdminTestJSON-830117193-project-member] Lock "d522ee4a-9bd9-402b-b3bd-c36f93366249" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 56.391s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 555.166071] env[62066]: DEBUG nova.network.neutron [None req-c4a0feae-1b12-4166-bd78-82a96fe8ac4c tempest-ServersWithSpecificFlavorTestJSON-184391817 tempest-ServersWithSpecificFlavorTestJSON-184391817-project-member] [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] Successfully created port: 8f4adee9-38df-4ef0-88ca-4c8c443fe300 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 555.476392] env[62066]: DEBUG nova.scheduler.client.report [None req-99d923d5-ef6f-43f2-959c-4b252ac1d398 tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 555.518151] env[62066]: DEBUG nova.compute.manager [None req-4fa8ac7b-ee00-4ef3-97aa-07a2c2ae6378 tempest-ServerExternalEventsTest-41922728 tempest-ServerExternalEventsTest-41922728-project-member] [instance: 39cd204a-49c2-4309-9030-555056ab125e] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 555.888939] env[62066]: DEBUG nova.compute.manager [None req-c4a0feae-1b12-4166-bd78-82a96fe8ac4c tempest-ServersWithSpecificFlavorTestJSON-184391817 tempest-ServersWithSpecificFlavorTestJSON-184391817-project-member] [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 555.922171] env[62066]: DEBUG nova.virt.hardware [None req-c4a0feae-1b12-4166-bd78-82a96fe8ac4c tempest-ServersWithSpecificFlavorTestJSON-184391817 tempest-ServersWithSpecificFlavorTestJSON-184391817-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-23T13:42:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='1621836986',id=23,is_public=True,memory_mb=192,name='tempest-flavor_with_ephemeral_0-2115743874',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-23T13:40:23Z,direct_url=,disk_format='vmdk',id=50ff584c-3b50-4395-af07-3e66769bc9f7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='eb52f7069a374c61ae946f052007c6d9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-23T13:40:24Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 555.922487] env[62066]: DEBUG nova.virt.hardware [None req-c4a0feae-1b12-4166-bd78-82a96fe8ac4c tempest-ServersWithSpecificFlavorTestJSON-184391817 tempest-ServersWithSpecificFlavorTestJSON-184391817-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 555.922691] env[62066]: DEBUG nova.virt.hardware [None req-c4a0feae-1b12-4166-bd78-82a96fe8ac4c tempest-ServersWithSpecificFlavorTestJSON-184391817 tempest-ServersWithSpecificFlavorTestJSON-184391817-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 555.923260] env[62066]: DEBUG nova.virt.hardware [None req-c4a0feae-1b12-4166-bd78-82a96fe8ac4c tempest-ServersWithSpecificFlavorTestJSON-184391817 tempest-ServersWithSpecificFlavorTestJSON-184391817-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 555.924044] env[62066]: DEBUG nova.virt.hardware [None req-c4a0feae-1b12-4166-bd78-82a96fe8ac4c tempest-ServersWithSpecificFlavorTestJSON-184391817 tempest-ServersWithSpecificFlavorTestJSON-184391817-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 555.924255] env[62066]: DEBUG nova.virt.hardware [None req-c4a0feae-1b12-4166-bd78-82a96fe8ac4c tempest-ServersWithSpecificFlavorTestJSON-184391817 tempest-ServersWithSpecificFlavorTestJSON-184391817-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 555.924551] env[62066]: DEBUG nova.virt.hardware [None req-c4a0feae-1b12-4166-bd78-82a96fe8ac4c tempest-ServersWithSpecificFlavorTestJSON-184391817 tempest-ServersWithSpecificFlavorTestJSON-184391817-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 555.924762] env[62066]: DEBUG nova.virt.hardware [None req-c4a0feae-1b12-4166-bd78-82a96fe8ac4c tempest-ServersWithSpecificFlavorTestJSON-184391817 tempest-ServersWithSpecificFlavorTestJSON-184391817-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 555.924985] env[62066]: DEBUG nova.virt.hardware [None req-c4a0feae-1b12-4166-bd78-82a96fe8ac4c tempest-ServersWithSpecificFlavorTestJSON-184391817 tempest-ServersWithSpecificFlavorTestJSON-184391817-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 555.925437] env[62066]: DEBUG nova.virt.hardware [None req-c4a0feae-1b12-4166-bd78-82a96fe8ac4c tempest-ServersWithSpecificFlavorTestJSON-184391817 tempest-ServersWithSpecificFlavorTestJSON-184391817-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 555.925668] env[62066]: DEBUG nova.virt.hardware [None req-c4a0feae-1b12-4166-bd78-82a96fe8ac4c tempest-ServersWithSpecificFlavorTestJSON-184391817 tempest-ServersWithSpecificFlavorTestJSON-184391817-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 555.926757] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e801e9fd-cc0f-4ec7-ad73-c3de13ff136e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.935688] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88ae7b89-0b23-4dbe-8369-951ce9b7d85e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.987227] env[62066]: DEBUG oslo_concurrency.lockutils [None req-99d923d5-ef6f-43f2-959c-4b252ac1d398 tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.127s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 555.987227] env[62066]: ERROR nova.compute.manager [None req-99d923d5-ef6f-43f2-959c-4b252ac1d398 tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 04cc7ab7-caaa-4afc-afd7-b4230a6e2326, please check neutron logs for more information. [ 555.987227] env[62066]: ERROR nova.compute.manager [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] Traceback (most recent call last): [ 555.987227] env[62066]: ERROR nova.compute.manager [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 555.987227] env[62066]: ERROR nova.compute.manager [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] self.driver.spawn(context, instance, image_meta, [ 555.987227] env[62066]: ERROR nova.compute.manager [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 555.987227] env[62066]: ERROR nova.compute.manager [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 555.987227] env[62066]: ERROR nova.compute.manager [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 555.987227] env[62066]: ERROR nova.compute.manager [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] vm_ref = self.build_virtual_machine(instance, [ 555.987787] env[62066]: ERROR nova.compute.manager [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 555.987787] env[62066]: ERROR nova.compute.manager [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] vif_infos = vmwarevif.get_vif_info(self._session, [ 555.987787] env[62066]: ERROR nova.compute.manager [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 555.987787] env[62066]: ERROR nova.compute.manager [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] for vif in network_info: [ 555.987787] env[62066]: ERROR nova.compute.manager [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 555.987787] env[62066]: ERROR nova.compute.manager [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] return self._sync_wrapper(fn, *args, **kwargs) [ 555.987787] env[62066]: ERROR nova.compute.manager [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 555.987787] env[62066]: ERROR nova.compute.manager [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] self.wait() [ 555.987787] env[62066]: ERROR nova.compute.manager [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 555.987787] env[62066]: ERROR nova.compute.manager [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] self[:] = self._gt.wait() [ 555.987787] env[62066]: ERROR nova.compute.manager [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 555.987787] env[62066]: ERROR nova.compute.manager [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] return self._exit_event.wait() [ 555.987787] env[62066]: ERROR nova.compute.manager [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 555.988207] env[62066]: ERROR nova.compute.manager [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] result = hub.switch() [ 555.988207] env[62066]: ERROR nova.compute.manager [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 555.988207] env[62066]: ERROR nova.compute.manager [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] return self.greenlet.switch() [ 555.988207] env[62066]: ERROR nova.compute.manager [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 555.988207] env[62066]: ERROR nova.compute.manager [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] result = function(*args, **kwargs) [ 555.988207] env[62066]: ERROR nova.compute.manager [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 555.988207] env[62066]: ERROR nova.compute.manager [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] return func(*args, **kwargs) [ 555.988207] env[62066]: ERROR nova.compute.manager [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 555.988207] env[62066]: ERROR nova.compute.manager [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] raise e [ 555.988207] env[62066]: ERROR nova.compute.manager [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 555.988207] env[62066]: ERROR nova.compute.manager [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] nwinfo = self.network_api.allocate_for_instance( [ 555.988207] env[62066]: ERROR nova.compute.manager [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 555.988207] env[62066]: ERROR nova.compute.manager [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] created_port_ids = self._update_ports_for_instance( [ 555.988722] env[62066]: ERROR nova.compute.manager [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 555.988722] env[62066]: ERROR nova.compute.manager [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] with excutils.save_and_reraise_exception(): [ 555.988722] env[62066]: ERROR nova.compute.manager [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 555.988722] env[62066]: ERROR nova.compute.manager [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] self.force_reraise() [ 555.988722] env[62066]: ERROR nova.compute.manager [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 555.988722] env[62066]: ERROR nova.compute.manager [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] raise self.value [ 555.988722] env[62066]: ERROR nova.compute.manager [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 555.988722] env[62066]: ERROR nova.compute.manager [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] updated_port = self._update_port( [ 555.988722] env[62066]: ERROR nova.compute.manager [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 555.988722] env[62066]: ERROR nova.compute.manager [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] _ensure_no_port_binding_failure(port) [ 555.988722] env[62066]: ERROR nova.compute.manager [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 555.988722] env[62066]: ERROR nova.compute.manager [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] raise exception.PortBindingFailed(port_id=port['id']) [ 555.989149] env[62066]: ERROR nova.compute.manager [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] nova.exception.PortBindingFailed: Binding failed for port 04cc7ab7-caaa-4afc-afd7-b4230a6e2326, please check neutron logs for more information. [ 555.989149] env[62066]: ERROR nova.compute.manager [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] [ 555.989149] env[62066]: DEBUG nova.compute.utils [None req-99d923d5-ef6f-43f2-959c-4b252ac1d398 tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] Binding failed for port 04cc7ab7-caaa-4afc-afd7-b4230a6e2326, please check neutron logs for more information. {{(pid=62066) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 555.991328] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cb29a201-afba-4eb8-8336-fe27a681c8f8 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.759s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 555.992951] env[62066]: INFO nova.compute.claims [None req-cb29a201-afba-4eb8-8336-fe27a681c8f8 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 556.006755] env[62066]: DEBUG nova.compute.manager [None req-99d923d5-ef6f-43f2-959c-4b252ac1d398 tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] Build of instance 68b65162-23a6-464f-a2f9-1635bea8786f was re-scheduled: Binding failed for port 04cc7ab7-caaa-4afc-afd7-b4230a6e2326, please check neutron logs for more information. {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 556.006755] env[62066]: DEBUG nova.compute.manager [None req-99d923d5-ef6f-43f2-959c-4b252ac1d398 tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] Unplugging VIFs for instance {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 556.006755] env[62066]: DEBUG oslo_concurrency.lockutils [None req-99d923d5-ef6f-43f2-959c-4b252ac1d398 tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] Acquiring lock "refresh_cache-68b65162-23a6-464f-a2f9-1635bea8786f" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 556.006755] env[62066]: DEBUG oslo_concurrency.lockutils [None req-99d923d5-ef6f-43f2-959c-4b252ac1d398 tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] Acquired lock "refresh_cache-68b65162-23a6-464f-a2f9-1635bea8786f" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 556.006946] env[62066]: DEBUG nova.network.neutron [None req-99d923d5-ef6f-43f2-959c-4b252ac1d398 tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 556.060333] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4fa8ac7b-ee00-4ef3-97aa-07a2c2ae6378 tempest-ServerExternalEventsTest-41922728 tempest-ServerExternalEventsTest-41922728-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 556.539350] env[62066]: DEBUG nova.network.neutron [None req-99d923d5-ef6f-43f2-959c-4b252ac1d398 tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 556.727642] env[62066]: DEBUG nova.compute.manager [req-7b98913a-a99c-4535-8e96-a164ff112cda req-916cebbd-31b4-4f31-8f9d-353c197a1a66 service nova] [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] Received event network-changed-8f4adee9-38df-4ef0-88ca-4c8c443fe300 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 556.728015] env[62066]: DEBUG nova.compute.manager [req-7b98913a-a99c-4535-8e96-a164ff112cda req-916cebbd-31b4-4f31-8f9d-353c197a1a66 service nova] [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] Refreshing instance network info cache due to event network-changed-8f4adee9-38df-4ef0-88ca-4c8c443fe300. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 556.728209] env[62066]: DEBUG oslo_concurrency.lockutils [req-7b98913a-a99c-4535-8e96-a164ff112cda req-916cebbd-31b4-4f31-8f9d-353c197a1a66 service nova] Acquiring lock "refresh_cache-4c67e82d-b51b-4820-af81-e6eefc7bcb36" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 556.730025] env[62066]: DEBUG oslo_concurrency.lockutils [req-7b98913a-a99c-4535-8e96-a164ff112cda req-916cebbd-31b4-4f31-8f9d-353c197a1a66 service nova] Acquired lock "refresh_cache-4c67e82d-b51b-4820-af81-e6eefc7bcb36" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 556.730025] env[62066]: DEBUG nova.network.neutron [req-7b98913a-a99c-4535-8e96-a164ff112cda req-916cebbd-31b4-4f31-8f9d-353c197a1a66 service nova] [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] Refreshing network info cache for port 8f4adee9-38df-4ef0-88ca-4c8c443fe300 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 556.874659] env[62066]: DEBUG nova.network.neutron [None req-99d923d5-ef6f-43f2-959c-4b252ac1d398 tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 557.273352] env[62066]: DEBUG nova.network.neutron [req-7b98913a-a99c-4535-8e96-a164ff112cda req-916cebbd-31b4-4f31-8f9d-353c197a1a66 service nova] [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 557.278145] env[62066]: ERROR nova.compute.manager [None req-c4a0feae-1b12-4166-bd78-82a96fe8ac4c tempest-ServersWithSpecificFlavorTestJSON-184391817 tempest-ServersWithSpecificFlavorTestJSON-184391817-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 8f4adee9-38df-4ef0-88ca-4c8c443fe300, please check neutron logs for more information. [ 557.278145] env[62066]: ERROR nova.compute.manager Traceback (most recent call last): [ 557.278145] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 557.278145] env[62066]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 557.278145] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 557.278145] env[62066]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 557.278145] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 557.278145] env[62066]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 557.278145] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 557.278145] env[62066]: ERROR nova.compute.manager self.force_reraise() [ 557.278145] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 557.278145] env[62066]: ERROR nova.compute.manager raise self.value [ 557.278145] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 557.278145] env[62066]: ERROR nova.compute.manager updated_port = self._update_port( [ 557.278145] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 557.278145] env[62066]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 557.279022] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 557.279022] env[62066]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 557.279022] env[62066]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 8f4adee9-38df-4ef0-88ca-4c8c443fe300, please check neutron logs for more information. [ 557.279022] env[62066]: ERROR nova.compute.manager [ 557.279022] env[62066]: Traceback (most recent call last): [ 557.279022] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 557.279022] env[62066]: listener.cb(fileno) [ 557.279022] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 557.279022] env[62066]: result = function(*args, **kwargs) [ 557.279022] env[62066]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 557.279022] env[62066]: return func(*args, **kwargs) [ 557.279022] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 557.279022] env[62066]: raise e [ 557.279022] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 557.279022] env[62066]: nwinfo = self.network_api.allocate_for_instance( [ 557.279022] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 557.279022] env[62066]: created_port_ids = self._update_ports_for_instance( [ 557.279022] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 557.279022] env[62066]: with excutils.save_and_reraise_exception(): [ 557.279022] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 557.279022] env[62066]: self.force_reraise() [ 557.279022] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 557.279022] env[62066]: raise self.value [ 557.279022] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 557.279022] env[62066]: updated_port = self._update_port( [ 557.279022] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 557.279022] env[62066]: _ensure_no_port_binding_failure(port) [ 557.279022] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 557.279022] env[62066]: raise exception.PortBindingFailed(port_id=port['id']) [ 557.280393] env[62066]: nova.exception.PortBindingFailed: Binding failed for port 8f4adee9-38df-4ef0-88ca-4c8c443fe300, please check neutron logs for more information. [ 557.280393] env[62066]: Removing descriptor: 20 [ 557.280393] env[62066]: ERROR nova.compute.manager [None req-c4a0feae-1b12-4166-bd78-82a96fe8ac4c tempest-ServersWithSpecificFlavorTestJSON-184391817 tempest-ServersWithSpecificFlavorTestJSON-184391817-project-member] [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 8f4adee9-38df-4ef0-88ca-4c8c443fe300, please check neutron logs for more information. [ 557.280393] env[62066]: ERROR nova.compute.manager [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] Traceback (most recent call last): [ 557.280393] env[62066]: ERROR nova.compute.manager [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 557.280393] env[62066]: ERROR nova.compute.manager [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] yield resources [ 557.280393] env[62066]: ERROR nova.compute.manager [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 557.280393] env[62066]: ERROR nova.compute.manager [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] self.driver.spawn(context, instance, image_meta, [ 557.280393] env[62066]: ERROR nova.compute.manager [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 557.280393] env[62066]: ERROR nova.compute.manager [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] self._vmops.spawn(context, instance, image_meta, injected_files, [ 557.280393] env[62066]: ERROR nova.compute.manager [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 557.280393] env[62066]: ERROR nova.compute.manager [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] vm_ref = self.build_virtual_machine(instance, [ 557.280918] env[62066]: ERROR nova.compute.manager [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 557.280918] env[62066]: ERROR nova.compute.manager [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] vif_infos = vmwarevif.get_vif_info(self._session, [ 557.280918] env[62066]: ERROR nova.compute.manager [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 557.280918] env[62066]: ERROR nova.compute.manager [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] for vif in network_info: [ 557.280918] env[62066]: ERROR nova.compute.manager [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 557.280918] env[62066]: ERROR nova.compute.manager [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] return self._sync_wrapper(fn, *args, **kwargs) [ 557.280918] env[62066]: ERROR nova.compute.manager [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 557.280918] env[62066]: ERROR nova.compute.manager [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] self.wait() [ 557.280918] env[62066]: ERROR nova.compute.manager [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 557.280918] env[62066]: ERROR nova.compute.manager [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] self[:] = self._gt.wait() [ 557.280918] env[62066]: ERROR nova.compute.manager [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 557.280918] env[62066]: ERROR nova.compute.manager [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] return self._exit_event.wait() [ 557.280918] env[62066]: ERROR nova.compute.manager [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 557.281452] env[62066]: ERROR nova.compute.manager [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] result = hub.switch() [ 557.281452] env[62066]: ERROR nova.compute.manager [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 557.281452] env[62066]: ERROR nova.compute.manager [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] return self.greenlet.switch() [ 557.281452] env[62066]: ERROR nova.compute.manager [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 557.281452] env[62066]: ERROR nova.compute.manager [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] result = function(*args, **kwargs) [ 557.281452] env[62066]: ERROR nova.compute.manager [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 557.281452] env[62066]: ERROR nova.compute.manager [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] return func(*args, **kwargs) [ 557.281452] env[62066]: ERROR nova.compute.manager [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 557.281452] env[62066]: ERROR nova.compute.manager [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] raise e [ 557.281452] env[62066]: ERROR nova.compute.manager [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 557.281452] env[62066]: ERROR nova.compute.manager [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] nwinfo = self.network_api.allocate_for_instance( [ 557.281452] env[62066]: ERROR nova.compute.manager [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 557.281452] env[62066]: ERROR nova.compute.manager [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] created_port_ids = self._update_ports_for_instance( [ 557.282216] env[62066]: ERROR nova.compute.manager [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 557.282216] env[62066]: ERROR nova.compute.manager [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] with excutils.save_and_reraise_exception(): [ 557.282216] env[62066]: ERROR nova.compute.manager [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 557.282216] env[62066]: ERROR nova.compute.manager [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] self.force_reraise() [ 557.282216] env[62066]: ERROR nova.compute.manager [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 557.282216] env[62066]: ERROR nova.compute.manager [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] raise self.value [ 557.282216] env[62066]: ERROR nova.compute.manager [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 557.282216] env[62066]: ERROR nova.compute.manager [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] updated_port = self._update_port( [ 557.282216] env[62066]: ERROR nova.compute.manager [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 557.282216] env[62066]: ERROR nova.compute.manager [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] _ensure_no_port_binding_failure(port) [ 557.282216] env[62066]: ERROR nova.compute.manager [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 557.282216] env[62066]: ERROR nova.compute.manager [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] raise exception.PortBindingFailed(port_id=port['id']) [ 557.282669] env[62066]: ERROR nova.compute.manager [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] nova.exception.PortBindingFailed: Binding failed for port 8f4adee9-38df-4ef0-88ca-4c8c443fe300, please check neutron logs for more information. [ 557.282669] env[62066]: ERROR nova.compute.manager [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] [ 557.282669] env[62066]: INFO nova.compute.manager [None req-c4a0feae-1b12-4166-bd78-82a96fe8ac4c tempest-ServersWithSpecificFlavorTestJSON-184391817 tempest-ServersWithSpecificFlavorTestJSON-184391817-project-member] [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] Terminating instance [ 557.282669] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c4a0feae-1b12-4166-bd78-82a96fe8ac4c tempest-ServersWithSpecificFlavorTestJSON-184391817 tempest-ServersWithSpecificFlavorTestJSON-184391817-project-member] Acquiring lock "refresh_cache-4c67e82d-b51b-4820-af81-e6eefc7bcb36" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 557.378035] env[62066]: DEBUG oslo_concurrency.lockutils [None req-99d923d5-ef6f-43f2-959c-4b252ac1d398 tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] Releasing lock "refresh_cache-68b65162-23a6-464f-a2f9-1635bea8786f" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 557.378657] env[62066]: DEBUG nova.compute.manager [None req-99d923d5-ef6f-43f2-959c-4b252ac1d398 tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 557.378843] env[62066]: DEBUG nova.compute.manager [None req-99d923d5-ef6f-43f2-959c-4b252ac1d398 tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 557.379057] env[62066]: DEBUG nova.network.neutron [None req-99d923d5-ef6f-43f2-959c-4b252ac1d398 tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 557.424358] env[62066]: DEBUG nova.network.neutron [None req-99d923d5-ef6f-43f2-959c-4b252ac1d398 tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 557.456343] env[62066]: DEBUG nova.network.neutron [req-7b98913a-a99c-4535-8e96-a164ff112cda req-916cebbd-31b4-4f31-8f9d-353c197a1a66 service nova] [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 557.603092] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46161530-d9af-4947-bf77-ef59dd11f3ef {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.612913] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-629712ab-e559-4780-90f6-373edf90e51c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.656634] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f7a5950-5ea6-4926-8881-2e62e943b9af {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.666202] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0dab837-8ff0-4a52-af5d-1d029aae6a1c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.681282] env[62066]: DEBUG nova.compute.provider_tree [None req-cb29a201-afba-4eb8-8336-fe27a681c8f8 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 557.930645] env[62066]: DEBUG nova.network.neutron [None req-99d923d5-ef6f-43f2-959c-4b252ac1d398 tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 557.963407] env[62066]: DEBUG oslo_concurrency.lockutils [req-7b98913a-a99c-4535-8e96-a164ff112cda req-916cebbd-31b4-4f31-8f9d-353c197a1a66 service nova] Releasing lock "refresh_cache-4c67e82d-b51b-4820-af81-e6eefc7bcb36" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 557.963734] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c4a0feae-1b12-4166-bd78-82a96fe8ac4c tempest-ServersWithSpecificFlavorTestJSON-184391817 tempest-ServersWithSpecificFlavorTestJSON-184391817-project-member] Acquired lock "refresh_cache-4c67e82d-b51b-4820-af81-e6eefc7bcb36" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 557.963920] env[62066]: DEBUG nova.network.neutron [None req-c4a0feae-1b12-4166-bd78-82a96fe8ac4c tempest-ServersWithSpecificFlavorTestJSON-184391817 tempest-ServersWithSpecificFlavorTestJSON-184391817-project-member] [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 558.185717] env[62066]: DEBUG nova.scheduler.client.report [None req-cb29a201-afba-4eb8-8336-fe27a681c8f8 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 558.376246] env[62066]: DEBUG oslo_concurrency.lockutils [None req-96f72479-6fed-4ff5-9d17-40fa20c1f416 tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] Acquiring lock "45c32181-01a8-4e1e-9e0e-37035a298d55" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 558.376246] env[62066]: DEBUG oslo_concurrency.lockutils [None req-96f72479-6fed-4ff5-9d17-40fa20c1f416 tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] Lock "45c32181-01a8-4e1e-9e0e-37035a298d55" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 558.435881] env[62066]: INFO nova.compute.manager [None req-99d923d5-ef6f-43f2-959c-4b252ac1d398 tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] [instance: 68b65162-23a6-464f-a2f9-1635bea8786f] Took 1.05 seconds to deallocate network for instance. [ 558.483650] env[62066]: DEBUG nova.network.neutron [None req-c4a0feae-1b12-4166-bd78-82a96fe8ac4c tempest-ServersWithSpecificFlavorTestJSON-184391817 tempest-ServersWithSpecificFlavorTestJSON-184391817-project-member] [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 558.611309] env[62066]: DEBUG nova.network.neutron [None req-c4a0feae-1b12-4166-bd78-82a96fe8ac4c tempest-ServersWithSpecificFlavorTestJSON-184391817 tempest-ServersWithSpecificFlavorTestJSON-184391817-project-member] [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 558.690809] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cb29a201-afba-4eb8-8336-fe27a681c8f8 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.703s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 558.691116] env[62066]: DEBUG nova.compute.manager [None req-cb29a201-afba-4eb8-8336-fe27a681c8f8 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 558.693696] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9838e0c6-e0b2-45d5-9c5b-a0a0b9c60219 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 26.541s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 558.946632] env[62066]: DEBUG nova.compute.manager [req-44cf28ae-95e0-405a-822f-839016ad29d7 req-baad9fde-b1a6-46de-9787-5f02df96e059 service nova] [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] Received event network-vif-deleted-8f4adee9-38df-4ef0-88ca-4c8c443fe300 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 559.113963] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c4a0feae-1b12-4166-bd78-82a96fe8ac4c tempest-ServersWithSpecificFlavorTestJSON-184391817 tempest-ServersWithSpecificFlavorTestJSON-184391817-project-member] Releasing lock "refresh_cache-4c67e82d-b51b-4820-af81-e6eefc7bcb36" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 559.114437] env[62066]: DEBUG nova.compute.manager [None req-c4a0feae-1b12-4166-bd78-82a96fe8ac4c tempest-ServersWithSpecificFlavorTestJSON-184391817 tempest-ServersWithSpecificFlavorTestJSON-184391817-project-member] [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 559.114662] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-c4a0feae-1b12-4166-bd78-82a96fe8ac4c tempest-ServersWithSpecificFlavorTestJSON-184391817 tempest-ServersWithSpecificFlavorTestJSON-184391817-project-member] [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 559.116087] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9186edda-046b-4595-8624-e7a24b362f28 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.129715] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7d8c662-2b22-4122-9eec-124adae7d429 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.156356] env[62066]: WARNING nova.virt.vmwareapi.vmops [None req-c4a0feae-1b12-4166-bd78-82a96fe8ac4c tempest-ServersWithSpecificFlavorTestJSON-184391817 tempest-ServersWithSpecificFlavorTestJSON-184391817-project-member] [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 4c67e82d-b51b-4820-af81-e6eefc7bcb36 could not be found. [ 559.156409] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-c4a0feae-1b12-4166-bd78-82a96fe8ac4c tempest-ServersWithSpecificFlavorTestJSON-184391817 tempest-ServersWithSpecificFlavorTestJSON-184391817-project-member] [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 559.156889] env[62066]: INFO nova.compute.manager [None req-c4a0feae-1b12-4166-bd78-82a96fe8ac4c tempest-ServersWithSpecificFlavorTestJSON-184391817 tempest-ServersWithSpecificFlavorTestJSON-184391817-project-member] [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] Took 0.04 seconds to destroy the instance on the hypervisor. [ 559.157230] env[62066]: DEBUG oslo.service.loopingcall [None req-c4a0feae-1b12-4166-bd78-82a96fe8ac4c tempest-ServersWithSpecificFlavorTestJSON-184391817 tempest-ServersWithSpecificFlavorTestJSON-184391817-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 559.157481] env[62066]: DEBUG nova.compute.manager [-] [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 559.157578] env[62066]: DEBUG nova.network.neutron [-] [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 559.181886] env[62066]: DEBUG nova.network.neutron [-] [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 559.202113] env[62066]: DEBUG nova.compute.utils [None req-cb29a201-afba-4eb8-8336-fe27a681c8f8 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 559.206621] env[62066]: DEBUG nova.compute.manager [None req-cb29a201-afba-4eb8-8336-fe27a681c8f8 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 559.206724] env[62066]: DEBUG nova.network.neutron [None req-cb29a201-afba-4eb8-8336-fe27a681c8f8 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 559.251539] env[62066]: DEBUG nova.policy [None req-cb29a201-afba-4eb8-8336-fe27a681c8f8 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '617259d206d647fbb20653e3d91083b3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '504e1c99f04c4d4fab8634c89cb017f4', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 559.476820] env[62066]: INFO nova.scheduler.client.report [None req-99d923d5-ef6f-43f2-959c-4b252ac1d398 tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] Deleted allocations for instance 68b65162-23a6-464f-a2f9-1635bea8786f [ 559.686859] env[62066]: DEBUG nova.network.neutron [-] [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 559.708446] env[62066]: DEBUG nova.compute.manager [None req-cb29a201-afba-4eb8-8336-fe27a681c8f8 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 559.715957] env[62066]: DEBUG nova.network.neutron [None req-cb29a201-afba-4eb8-8336-fe27a681c8f8 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] Successfully created port: c18dc899-1b90-41a9-8d63-1bf7f9d1a206 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 559.816015] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94fe405e-81c0-41a4-aca3-2293cd66d989 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.828943] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fae7fee-3d9d-4c29-9dc0-0828c29c93ac {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.865361] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d538226-0e69-4db8-b0e5-127e118f49f4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.873341] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b58a6d0c-1e3c-4c06-9676-0c1a8100df46 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.888831] env[62066]: DEBUG nova.compute.provider_tree [None req-9838e0c6-e0b2-45d5-9c5b-a0a0b9c60219 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 559.998472] env[62066]: DEBUG oslo_concurrency.lockutils [None req-99d923d5-ef6f-43f2-959c-4b252ac1d398 tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] Lock "68b65162-23a6-464f-a2f9-1635bea8786f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 61.240s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 560.190634] env[62066]: INFO nova.compute.manager [-] [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] Took 1.03 seconds to deallocate network for instance. [ 560.193275] env[62066]: DEBUG nova.compute.claims [None req-c4a0feae-1b12-4166-bd78-82a96fe8ac4c tempest-ServersWithSpecificFlavorTestJSON-184391817 tempest-ServersWithSpecificFlavorTestJSON-184391817-project-member] [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] Aborting claim: {{(pid=62066) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 560.193458] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c4a0feae-1b12-4166-bd78-82a96fe8ac4c tempest-ServersWithSpecificFlavorTestJSON-184391817 tempest-ServersWithSpecificFlavorTestJSON-184391817-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 560.394328] env[62066]: DEBUG nova.scheduler.client.report [None req-9838e0c6-e0b2-45d5-9c5b-a0a0b9c60219 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 560.503909] env[62066]: DEBUG nova.compute.manager [None req-1e3e30d6-8bb5-4347-b6c2-c2d648c37b67 tempest-ImagesOneServerNegativeTestJSON-310236551 tempest-ImagesOneServerNegativeTestJSON-310236551-project-member] [instance: 194588b4-fe40-4286-8036-874a7c410327] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 560.721072] env[62066]: DEBUG nova.compute.manager [None req-cb29a201-afba-4eb8-8336-fe27a681c8f8 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 560.759150] env[62066]: DEBUG nova.virt.hardware [None req-cb29a201-afba-4eb8-8336-fe27a681c8f8 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-23T13:40:41Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-23T13:40:23Z,direct_url=,disk_format='vmdk',id=50ff584c-3b50-4395-af07-3e66769bc9f7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='eb52f7069a374c61ae946f052007c6d9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-23T13:40:24Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 560.759150] env[62066]: DEBUG nova.virt.hardware [None req-cb29a201-afba-4eb8-8336-fe27a681c8f8 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 560.759150] env[62066]: DEBUG nova.virt.hardware [None req-cb29a201-afba-4eb8-8336-fe27a681c8f8 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 560.759339] env[62066]: DEBUG nova.virt.hardware [None req-cb29a201-afba-4eb8-8336-fe27a681c8f8 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 560.759339] env[62066]: DEBUG nova.virt.hardware [None req-cb29a201-afba-4eb8-8336-fe27a681c8f8 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 560.759339] env[62066]: DEBUG nova.virt.hardware [None req-cb29a201-afba-4eb8-8336-fe27a681c8f8 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 560.759449] env[62066]: DEBUG nova.virt.hardware [None req-cb29a201-afba-4eb8-8336-fe27a681c8f8 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 560.759561] env[62066]: DEBUG nova.virt.hardware [None req-cb29a201-afba-4eb8-8336-fe27a681c8f8 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 560.759671] env[62066]: DEBUG nova.virt.hardware [None req-cb29a201-afba-4eb8-8336-fe27a681c8f8 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 560.759829] env[62066]: DEBUG nova.virt.hardware [None req-cb29a201-afba-4eb8-8336-fe27a681c8f8 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 560.759993] env[62066]: DEBUG nova.virt.hardware [None req-cb29a201-afba-4eb8-8336-fe27a681c8f8 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 560.760881] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d9161bd-90ca-4395-a747-19e513d09759 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.773241] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-516578ca-f1f6-4daf-8bea-dfb16414be86 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.832697] env[62066]: DEBUG nova.compute.manager [req-fdc8577f-17b0-4b0c-b763-e9f1f7430ae9 req-1776b1d9-d247-41e6-983c-46b16dd0fe44 service nova] [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] Received event network-changed-c18dc899-1b90-41a9-8d63-1bf7f9d1a206 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 560.832880] env[62066]: DEBUG nova.compute.manager [req-fdc8577f-17b0-4b0c-b763-e9f1f7430ae9 req-1776b1d9-d247-41e6-983c-46b16dd0fe44 service nova] [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] Refreshing instance network info cache due to event network-changed-c18dc899-1b90-41a9-8d63-1bf7f9d1a206. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 560.834630] env[62066]: DEBUG oslo_concurrency.lockutils [req-fdc8577f-17b0-4b0c-b763-e9f1f7430ae9 req-1776b1d9-d247-41e6-983c-46b16dd0fe44 service nova] Acquiring lock "refresh_cache-69d99063-a6a7-4832-a864-dc9fd6024ea8" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 560.834630] env[62066]: DEBUG oslo_concurrency.lockutils [req-fdc8577f-17b0-4b0c-b763-e9f1f7430ae9 req-1776b1d9-d247-41e6-983c-46b16dd0fe44 service nova] Acquired lock "refresh_cache-69d99063-a6a7-4832-a864-dc9fd6024ea8" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 560.834630] env[62066]: DEBUG nova.network.neutron [req-fdc8577f-17b0-4b0c-b763-e9f1f7430ae9 req-1776b1d9-d247-41e6-983c-46b16dd0fe44 service nova] [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] Refreshing network info cache for port c18dc899-1b90-41a9-8d63-1bf7f9d1a206 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 560.871738] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b2b06eaa-4237-476b-9167-ed8be81ba3b9 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Acquiring lock "f9a8e863-6e7f-4f00-b54a-78802659cd59" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 560.872971] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b2b06eaa-4237-476b-9167-ed8be81ba3b9 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Lock "f9a8e863-6e7f-4f00-b54a-78802659cd59" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 560.900442] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9838e0c6-e0b2-45d5-9c5b-a0a0b9c60219 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.207s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 560.900996] env[62066]: ERROR nova.compute.manager [None req-9838e0c6-e0b2-45d5-9c5b-a0a0b9c60219 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] [instance: 64f9698e-16f5-447a-914a-9e936e05d101] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 8a431920-fe14-44e1-85f9-31b9af5b10c2, please check neutron logs for more information. [ 560.900996] env[62066]: ERROR nova.compute.manager [instance: 64f9698e-16f5-447a-914a-9e936e05d101] Traceback (most recent call last): [ 560.900996] env[62066]: ERROR nova.compute.manager [instance: 64f9698e-16f5-447a-914a-9e936e05d101] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 560.900996] env[62066]: ERROR nova.compute.manager [instance: 64f9698e-16f5-447a-914a-9e936e05d101] self.driver.spawn(context, instance, image_meta, [ 560.900996] env[62066]: ERROR nova.compute.manager [instance: 64f9698e-16f5-447a-914a-9e936e05d101] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 560.900996] env[62066]: ERROR nova.compute.manager [instance: 64f9698e-16f5-447a-914a-9e936e05d101] self._vmops.spawn(context, instance, image_meta, injected_files, [ 560.900996] env[62066]: ERROR nova.compute.manager [instance: 64f9698e-16f5-447a-914a-9e936e05d101] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 560.900996] env[62066]: ERROR nova.compute.manager [instance: 64f9698e-16f5-447a-914a-9e936e05d101] vm_ref = self.build_virtual_machine(instance, [ 560.900996] env[62066]: ERROR nova.compute.manager [instance: 64f9698e-16f5-447a-914a-9e936e05d101] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 560.900996] env[62066]: ERROR nova.compute.manager [instance: 64f9698e-16f5-447a-914a-9e936e05d101] vif_infos = vmwarevif.get_vif_info(self._session, [ 560.900996] env[62066]: ERROR nova.compute.manager [instance: 64f9698e-16f5-447a-914a-9e936e05d101] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 560.901408] env[62066]: ERROR nova.compute.manager [instance: 64f9698e-16f5-447a-914a-9e936e05d101] for vif in network_info: [ 560.901408] env[62066]: ERROR nova.compute.manager [instance: 64f9698e-16f5-447a-914a-9e936e05d101] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 560.901408] env[62066]: ERROR nova.compute.manager [instance: 64f9698e-16f5-447a-914a-9e936e05d101] return self._sync_wrapper(fn, *args, **kwargs) [ 560.901408] env[62066]: ERROR nova.compute.manager [instance: 64f9698e-16f5-447a-914a-9e936e05d101] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 560.901408] env[62066]: ERROR nova.compute.manager [instance: 64f9698e-16f5-447a-914a-9e936e05d101] self.wait() [ 560.901408] env[62066]: ERROR nova.compute.manager [instance: 64f9698e-16f5-447a-914a-9e936e05d101] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 560.901408] env[62066]: ERROR nova.compute.manager [instance: 64f9698e-16f5-447a-914a-9e936e05d101] self[:] = self._gt.wait() [ 560.901408] env[62066]: ERROR nova.compute.manager [instance: 64f9698e-16f5-447a-914a-9e936e05d101] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 560.901408] env[62066]: ERROR nova.compute.manager [instance: 64f9698e-16f5-447a-914a-9e936e05d101] return self._exit_event.wait() [ 560.901408] env[62066]: ERROR nova.compute.manager [instance: 64f9698e-16f5-447a-914a-9e936e05d101] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 560.901408] env[62066]: ERROR nova.compute.manager [instance: 64f9698e-16f5-447a-914a-9e936e05d101] result = hub.switch() [ 560.901408] env[62066]: ERROR nova.compute.manager [instance: 64f9698e-16f5-447a-914a-9e936e05d101] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 560.901408] env[62066]: ERROR nova.compute.manager [instance: 64f9698e-16f5-447a-914a-9e936e05d101] return self.greenlet.switch() [ 560.901806] env[62066]: ERROR nova.compute.manager [instance: 64f9698e-16f5-447a-914a-9e936e05d101] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 560.901806] env[62066]: ERROR nova.compute.manager [instance: 64f9698e-16f5-447a-914a-9e936e05d101] result = function(*args, **kwargs) [ 560.901806] env[62066]: ERROR nova.compute.manager [instance: 64f9698e-16f5-447a-914a-9e936e05d101] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 560.901806] env[62066]: ERROR nova.compute.manager [instance: 64f9698e-16f5-447a-914a-9e936e05d101] return func(*args, **kwargs) [ 560.901806] env[62066]: ERROR nova.compute.manager [instance: 64f9698e-16f5-447a-914a-9e936e05d101] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 560.901806] env[62066]: ERROR nova.compute.manager [instance: 64f9698e-16f5-447a-914a-9e936e05d101] raise e [ 560.901806] env[62066]: ERROR nova.compute.manager [instance: 64f9698e-16f5-447a-914a-9e936e05d101] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 560.901806] env[62066]: ERROR nova.compute.manager [instance: 64f9698e-16f5-447a-914a-9e936e05d101] nwinfo = self.network_api.allocate_for_instance( [ 560.901806] env[62066]: ERROR nova.compute.manager [instance: 64f9698e-16f5-447a-914a-9e936e05d101] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 560.901806] env[62066]: ERROR nova.compute.manager [instance: 64f9698e-16f5-447a-914a-9e936e05d101] created_port_ids = self._update_ports_for_instance( [ 560.901806] env[62066]: ERROR nova.compute.manager [instance: 64f9698e-16f5-447a-914a-9e936e05d101] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 560.901806] env[62066]: ERROR nova.compute.manager [instance: 64f9698e-16f5-447a-914a-9e936e05d101] with excutils.save_and_reraise_exception(): [ 560.901806] env[62066]: ERROR nova.compute.manager [instance: 64f9698e-16f5-447a-914a-9e936e05d101] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 560.902336] env[62066]: ERROR nova.compute.manager [instance: 64f9698e-16f5-447a-914a-9e936e05d101] self.force_reraise() [ 560.902336] env[62066]: ERROR nova.compute.manager [instance: 64f9698e-16f5-447a-914a-9e936e05d101] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 560.902336] env[62066]: ERROR nova.compute.manager [instance: 64f9698e-16f5-447a-914a-9e936e05d101] raise self.value [ 560.902336] env[62066]: ERROR nova.compute.manager [instance: 64f9698e-16f5-447a-914a-9e936e05d101] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 560.902336] env[62066]: ERROR nova.compute.manager [instance: 64f9698e-16f5-447a-914a-9e936e05d101] updated_port = self._update_port( [ 560.902336] env[62066]: ERROR nova.compute.manager [instance: 64f9698e-16f5-447a-914a-9e936e05d101] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 560.902336] env[62066]: ERROR nova.compute.manager [instance: 64f9698e-16f5-447a-914a-9e936e05d101] _ensure_no_port_binding_failure(port) [ 560.902336] env[62066]: ERROR nova.compute.manager [instance: 64f9698e-16f5-447a-914a-9e936e05d101] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 560.902336] env[62066]: ERROR nova.compute.manager [instance: 64f9698e-16f5-447a-914a-9e936e05d101] raise exception.PortBindingFailed(port_id=port['id']) [ 560.902336] env[62066]: ERROR nova.compute.manager [instance: 64f9698e-16f5-447a-914a-9e936e05d101] nova.exception.PortBindingFailed: Binding failed for port 8a431920-fe14-44e1-85f9-31b9af5b10c2, please check neutron logs for more information. [ 560.902336] env[62066]: ERROR nova.compute.manager [instance: 64f9698e-16f5-447a-914a-9e936e05d101] [ 560.902679] env[62066]: DEBUG nova.compute.utils [None req-9838e0c6-e0b2-45d5-9c5b-a0a0b9c60219 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] [instance: 64f9698e-16f5-447a-914a-9e936e05d101] Binding failed for port 8a431920-fe14-44e1-85f9-31b9af5b10c2, please check neutron logs for more information. {{(pid=62066) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 560.903784] env[62066]: ERROR nova.compute.manager [None req-cb29a201-afba-4eb8-8336-fe27a681c8f8 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port c18dc899-1b90-41a9-8d63-1bf7f9d1a206, please check neutron logs for more information. [ 560.903784] env[62066]: ERROR nova.compute.manager Traceback (most recent call last): [ 560.903784] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 560.903784] env[62066]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 560.903784] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 560.903784] env[62066]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 560.903784] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 560.903784] env[62066]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 560.903784] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 560.903784] env[62066]: ERROR nova.compute.manager self.force_reraise() [ 560.903784] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 560.903784] env[62066]: ERROR nova.compute.manager raise self.value [ 560.903784] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 560.903784] env[62066]: ERROR nova.compute.manager updated_port = self._update_port( [ 560.903784] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 560.903784] env[62066]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 560.904402] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 560.904402] env[62066]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 560.904402] env[62066]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port c18dc899-1b90-41a9-8d63-1bf7f9d1a206, please check neutron logs for more information. [ 560.904402] env[62066]: ERROR nova.compute.manager [ 560.904402] env[62066]: Traceback (most recent call last): [ 560.904402] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 560.904402] env[62066]: listener.cb(fileno) [ 560.904402] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 560.904402] env[62066]: result = function(*args, **kwargs) [ 560.904402] env[62066]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 560.904402] env[62066]: return func(*args, **kwargs) [ 560.904402] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 560.904402] env[62066]: raise e [ 560.904402] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 560.904402] env[62066]: nwinfo = self.network_api.allocate_for_instance( [ 560.904402] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 560.904402] env[62066]: created_port_ids = self._update_ports_for_instance( [ 560.904402] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 560.904402] env[62066]: with excutils.save_and_reraise_exception(): [ 560.904402] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 560.904402] env[62066]: self.force_reraise() [ 560.904402] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 560.904402] env[62066]: raise self.value [ 560.904402] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 560.904402] env[62066]: updated_port = self._update_port( [ 560.904402] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 560.904402] env[62066]: _ensure_no_port_binding_failure(port) [ 560.904402] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 560.904402] env[62066]: raise exception.PortBindingFailed(port_id=port['id']) [ 560.910791] env[62066]: nova.exception.PortBindingFailed: Binding failed for port c18dc899-1b90-41a9-8d63-1bf7f9d1a206, please check neutron logs for more information. [ 560.910791] env[62066]: Removing descriptor: 15 [ 560.910791] env[62066]: DEBUG oslo_concurrency.lockutils [None req-65bf05da-7d56-4074-8d13-63fde71b9619 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.041s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 560.910791] env[62066]: INFO nova.compute.claims [None req-65bf05da-7d56-4074-8d13-63fde71b9619 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 560.910791] env[62066]: ERROR nova.compute.manager [None req-cb29a201-afba-4eb8-8336-fe27a681c8f8 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port c18dc899-1b90-41a9-8d63-1bf7f9d1a206, please check neutron logs for more information. [ 560.910791] env[62066]: ERROR nova.compute.manager [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] Traceback (most recent call last): [ 560.910791] env[62066]: ERROR nova.compute.manager [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 560.910791] env[62066]: ERROR nova.compute.manager [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] yield resources [ 560.910791] env[62066]: ERROR nova.compute.manager [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 560.911289] env[62066]: ERROR nova.compute.manager [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] self.driver.spawn(context, instance, image_meta, [ 560.911289] env[62066]: ERROR nova.compute.manager [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 560.911289] env[62066]: ERROR nova.compute.manager [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] self._vmops.spawn(context, instance, image_meta, injected_files, [ 560.911289] env[62066]: ERROR nova.compute.manager [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 560.911289] env[62066]: ERROR nova.compute.manager [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] vm_ref = self.build_virtual_machine(instance, [ 560.911289] env[62066]: ERROR nova.compute.manager [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 560.911289] env[62066]: ERROR nova.compute.manager [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] vif_infos = vmwarevif.get_vif_info(self._session, [ 560.911289] env[62066]: ERROR nova.compute.manager [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 560.911289] env[62066]: ERROR nova.compute.manager [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] for vif in network_info: [ 560.911289] env[62066]: ERROR nova.compute.manager [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 560.911289] env[62066]: ERROR nova.compute.manager [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] return self._sync_wrapper(fn, *args, **kwargs) [ 560.911289] env[62066]: ERROR nova.compute.manager [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 560.911289] env[62066]: ERROR nova.compute.manager [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] self.wait() [ 560.911958] env[62066]: ERROR nova.compute.manager [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 560.911958] env[62066]: ERROR nova.compute.manager [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] self[:] = self._gt.wait() [ 560.911958] env[62066]: ERROR nova.compute.manager [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 560.911958] env[62066]: ERROR nova.compute.manager [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] return self._exit_event.wait() [ 560.911958] env[62066]: ERROR nova.compute.manager [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 560.911958] env[62066]: ERROR nova.compute.manager [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] result = hub.switch() [ 560.911958] env[62066]: ERROR nova.compute.manager [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 560.911958] env[62066]: ERROR nova.compute.manager [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] return self.greenlet.switch() [ 560.911958] env[62066]: ERROR nova.compute.manager [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 560.911958] env[62066]: ERROR nova.compute.manager [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] result = function(*args, **kwargs) [ 560.911958] env[62066]: ERROR nova.compute.manager [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 560.911958] env[62066]: ERROR nova.compute.manager [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] return func(*args, **kwargs) [ 560.911958] env[62066]: ERROR nova.compute.manager [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 560.912362] env[62066]: ERROR nova.compute.manager [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] raise e [ 560.912362] env[62066]: ERROR nova.compute.manager [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 560.912362] env[62066]: ERROR nova.compute.manager [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] nwinfo = self.network_api.allocate_for_instance( [ 560.912362] env[62066]: ERROR nova.compute.manager [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 560.912362] env[62066]: ERROR nova.compute.manager [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] created_port_ids = self._update_ports_for_instance( [ 560.912362] env[62066]: ERROR nova.compute.manager [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 560.912362] env[62066]: ERROR nova.compute.manager [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] with excutils.save_and_reraise_exception(): [ 560.912362] env[62066]: ERROR nova.compute.manager [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 560.912362] env[62066]: ERROR nova.compute.manager [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] self.force_reraise() [ 560.912362] env[62066]: ERROR nova.compute.manager [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 560.912362] env[62066]: ERROR nova.compute.manager [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] raise self.value [ 560.912362] env[62066]: ERROR nova.compute.manager [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 560.912362] env[62066]: ERROR nova.compute.manager [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] updated_port = self._update_port( [ 560.912735] env[62066]: ERROR nova.compute.manager [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 560.912735] env[62066]: ERROR nova.compute.manager [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] _ensure_no_port_binding_failure(port) [ 560.912735] env[62066]: ERROR nova.compute.manager [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 560.912735] env[62066]: ERROR nova.compute.manager [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] raise exception.PortBindingFailed(port_id=port['id']) [ 560.912735] env[62066]: ERROR nova.compute.manager [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] nova.exception.PortBindingFailed: Binding failed for port c18dc899-1b90-41a9-8d63-1bf7f9d1a206, please check neutron logs for more information. [ 560.912735] env[62066]: ERROR nova.compute.manager [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] [ 560.912735] env[62066]: INFO nova.compute.manager [None req-cb29a201-afba-4eb8-8336-fe27a681c8f8 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] Terminating instance [ 560.912735] env[62066]: DEBUG nova.compute.manager [None req-9838e0c6-e0b2-45d5-9c5b-a0a0b9c60219 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] [instance: 64f9698e-16f5-447a-914a-9e936e05d101] Build of instance 64f9698e-16f5-447a-914a-9e936e05d101 was re-scheduled: Binding failed for port 8a431920-fe14-44e1-85f9-31b9af5b10c2, please check neutron logs for more information. {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 560.913015] env[62066]: DEBUG nova.compute.manager [None req-9838e0c6-e0b2-45d5-9c5b-a0a0b9c60219 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] [instance: 64f9698e-16f5-447a-914a-9e936e05d101] Unplugging VIFs for instance {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 560.913015] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9838e0c6-e0b2-45d5-9c5b-a0a0b9c60219 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] Acquiring lock "refresh_cache-64f9698e-16f5-447a-914a-9e936e05d101" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 560.913015] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9838e0c6-e0b2-45d5-9c5b-a0a0b9c60219 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] Acquired lock "refresh_cache-64f9698e-16f5-447a-914a-9e936e05d101" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 560.913015] env[62066]: DEBUG nova.network.neutron [None req-9838e0c6-e0b2-45d5-9c5b-a0a0b9c60219 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] [instance: 64f9698e-16f5-447a-914a-9e936e05d101] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 560.913015] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cb29a201-afba-4eb8-8336-fe27a681c8f8 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] Acquiring lock "refresh_cache-69d99063-a6a7-4832-a864-dc9fd6024ea8" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 561.028323] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1e3e30d6-8bb5-4347-b6c2-c2d648c37b67 tempest-ImagesOneServerNegativeTestJSON-310236551 tempest-ImagesOneServerNegativeTestJSON-310236551-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 561.358669] env[62066]: DEBUG nova.network.neutron [req-fdc8577f-17b0-4b0c-b763-e9f1f7430ae9 req-1776b1d9-d247-41e6-983c-46b16dd0fe44 service nova] [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 561.439204] env[62066]: DEBUG nova.network.neutron [None req-9838e0c6-e0b2-45d5-9c5b-a0a0b9c60219 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] [instance: 64f9698e-16f5-447a-914a-9e936e05d101] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 561.481971] env[62066]: DEBUG nova.network.neutron [req-fdc8577f-17b0-4b0c-b763-e9f1f7430ae9 req-1776b1d9-d247-41e6-983c-46b16dd0fe44 service nova] [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 561.567912] env[62066]: DEBUG nova.network.neutron [None req-9838e0c6-e0b2-45d5-9c5b-a0a0b9c60219 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] [instance: 64f9698e-16f5-447a-914a-9e936e05d101] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 561.974274] env[62066]: DEBUG oslo_concurrency.lockutils [None req-84147b21-d606-4daf-b0b9-f581072be1cb tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] Acquiring lock "2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 561.974492] env[62066]: DEBUG oslo_concurrency.lockutils [None req-84147b21-d606-4daf-b0b9-f581072be1cb tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] Lock "2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 561.987169] env[62066]: DEBUG oslo_concurrency.lockutils [req-fdc8577f-17b0-4b0c-b763-e9f1f7430ae9 req-1776b1d9-d247-41e6-983c-46b16dd0fe44 service nova] Releasing lock "refresh_cache-69d99063-a6a7-4832-a864-dc9fd6024ea8" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 561.987580] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cb29a201-afba-4eb8-8336-fe27a681c8f8 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] Acquired lock "refresh_cache-69d99063-a6a7-4832-a864-dc9fd6024ea8" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 561.987681] env[62066]: DEBUG nova.network.neutron [None req-cb29a201-afba-4eb8-8336-fe27a681c8f8 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 562.072514] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9838e0c6-e0b2-45d5-9c5b-a0a0b9c60219 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] Releasing lock "refresh_cache-64f9698e-16f5-447a-914a-9e936e05d101" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 562.072730] env[62066]: DEBUG nova.compute.manager [None req-9838e0c6-e0b2-45d5-9c5b-a0a0b9c60219 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 562.072915] env[62066]: DEBUG nova.compute.manager [None req-9838e0c6-e0b2-45d5-9c5b-a0a0b9c60219 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] [instance: 64f9698e-16f5-447a-914a-9e936e05d101] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 562.073124] env[62066]: DEBUG nova.network.neutron [None req-9838e0c6-e0b2-45d5-9c5b-a0a0b9c60219 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] [instance: 64f9698e-16f5-447a-914a-9e936e05d101] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 562.102180] env[62066]: DEBUG nova.network.neutron [None req-9838e0c6-e0b2-45d5-9c5b-a0a0b9c60219 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] [instance: 64f9698e-16f5-447a-914a-9e936e05d101] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 562.380015] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10c93586-5121-48a6-922e-60aaa1af62c3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.385465] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0340708-b3ba-4821-b97a-fd69058e8174 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.415666] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80582fad-9f23-417d-9bd2-b555aecc91c3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.422994] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76731abf-fc94-4c55-976e-42559d53d7f2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.436086] env[62066]: DEBUG nova.compute.provider_tree [None req-65bf05da-7d56-4074-8d13-63fde71b9619 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 562.506175] env[62066]: DEBUG nova.network.neutron [None req-cb29a201-afba-4eb8-8336-fe27a681c8f8 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 562.613862] env[62066]: DEBUG nova.network.neutron [None req-9838e0c6-e0b2-45d5-9c5b-a0a0b9c60219 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] [instance: 64f9698e-16f5-447a-914a-9e936e05d101] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 562.625751] env[62066]: DEBUG nova.network.neutron [None req-cb29a201-afba-4eb8-8336-fe27a681c8f8 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 562.896278] env[62066]: DEBUG nova.compute.manager [req-7d4ae4fe-0327-4f31-b6a3-c3742a042ba9 req-5d077615-e854-4804-b8b9-3fd04be8f444 service nova] [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] Received event network-vif-deleted-c18dc899-1b90-41a9-8d63-1bf7f9d1a206 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 562.939521] env[62066]: DEBUG nova.scheduler.client.report [None req-65bf05da-7d56-4074-8d13-63fde71b9619 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 563.119873] env[62066]: INFO nova.compute.manager [None req-9838e0c6-e0b2-45d5-9c5b-a0a0b9c60219 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] [instance: 64f9698e-16f5-447a-914a-9e936e05d101] Took 1.05 seconds to deallocate network for instance. [ 563.130786] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cb29a201-afba-4eb8-8336-fe27a681c8f8 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] Releasing lock "refresh_cache-69d99063-a6a7-4832-a864-dc9fd6024ea8" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 563.131186] env[62066]: DEBUG nova.compute.manager [None req-cb29a201-afba-4eb8-8336-fe27a681c8f8 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 563.131374] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-cb29a201-afba-4eb8-8336-fe27a681c8f8 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 563.131666] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e9f2aad4-114a-4d38-87dd-e8ff17fba80d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.141245] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61242cbb-89fd-4744-9a4e-2239e29dd6a3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.166432] env[62066]: WARNING nova.virt.vmwareapi.vmops [None req-cb29a201-afba-4eb8-8336-fe27a681c8f8 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 69d99063-a6a7-4832-a864-dc9fd6024ea8 could not be found. [ 563.166765] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-cb29a201-afba-4eb8-8336-fe27a681c8f8 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 563.166969] env[62066]: INFO nova.compute.manager [None req-cb29a201-afba-4eb8-8336-fe27a681c8f8 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] Took 0.04 seconds to destroy the instance on the hypervisor. [ 563.167246] env[62066]: DEBUG oslo.service.loopingcall [None req-cb29a201-afba-4eb8-8336-fe27a681c8f8 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 563.168176] env[62066]: DEBUG nova.compute.manager [-] [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 563.168276] env[62066]: DEBUG nova.network.neutron [-] [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 563.357924] env[62066]: DEBUG nova.network.neutron [-] [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 563.445139] env[62066]: DEBUG oslo_concurrency.lockutils [None req-65bf05da-7d56-4074-8d13-63fde71b9619 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.541s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 563.445788] env[62066]: DEBUG nova.compute.manager [None req-65bf05da-7d56-4074-8d13-63fde71b9619 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 563.448747] env[62066]: DEBUG oslo_concurrency.lockutils [None req-659c7f6d-78fb-4b73-9024-3866f9052a37 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 28.970s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 563.449118] env[62066]: DEBUG nova.objects.instance [None req-659c7f6d-78fb-4b73-9024-3866f9052a37 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] Lazy-loading 'resources' on Instance uuid 4b34ff57-e137-4abc-8aed-76dd2ec8d313 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 563.865897] env[62066]: DEBUG nova.network.neutron [-] [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 563.952896] env[62066]: DEBUG nova.compute.utils [None req-65bf05da-7d56-4074-8d13-63fde71b9619 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 563.957985] env[62066]: DEBUG nova.compute.manager [None req-65bf05da-7d56-4074-8d13-63fde71b9619 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 563.957985] env[62066]: DEBUG nova.network.neutron [None req-65bf05da-7d56-4074-8d13-63fde71b9619 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 564.021722] env[62066]: DEBUG nova.policy [None req-65bf05da-7d56-4074-8d13-63fde71b9619 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '617259d206d647fbb20653e3d91083b3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '504e1c99f04c4d4fab8634c89cb017f4', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 564.157262] env[62066]: INFO nova.scheduler.client.report [None req-9838e0c6-e0b2-45d5-9c5b-a0a0b9c60219 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] Deleted allocations for instance 64f9698e-16f5-447a-914a-9e936e05d101 [ 564.371336] env[62066]: INFO nova.compute.manager [-] [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] Took 1.20 seconds to deallocate network for instance. [ 564.373866] env[62066]: DEBUG nova.compute.claims [None req-cb29a201-afba-4eb8-8336-fe27a681c8f8 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] Aborting claim: {{(pid=62066) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 564.374146] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cb29a201-afba-4eb8-8336-fe27a681c8f8 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 564.457606] env[62066]: DEBUG nova.compute.manager [None req-65bf05da-7d56-4074-8d13-63fde71b9619 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 564.476258] env[62066]: DEBUG nova.network.neutron [None req-65bf05da-7d56-4074-8d13-63fde71b9619 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] Successfully created port: 39ab0f84-58c5-4569-bf28-6dbda905c21d {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 564.487297] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f571454-5315-4cfd-9ce5-a960c63495d8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.494664] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b073cc5d-8b4b-40b2-807a-c30b6451802a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.531622] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-732be29d-e309-4ab0-9853-48c290a9ec2e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.538671] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2214131-15e9-4f7e-a2e2-bdf84655617b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.551944] env[62066]: DEBUG nova.compute.provider_tree [None req-659c7f6d-78fb-4b73-9024-3866f9052a37 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 564.668108] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9838e0c6-e0b2-45d5-9c5b-a0a0b9c60219 tempest-ServerRescueNegativeTestJSON-981773185 tempest-ServerRescueNegativeTestJSON-981773185-project-member] Lock "64f9698e-16f5-447a-914a-9e936e05d101" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 64.957s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 565.055672] env[62066]: DEBUG nova.scheduler.client.report [None req-659c7f6d-78fb-4b73-9024-3866f9052a37 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 565.171321] env[62066]: DEBUG nova.compute.manager [None req-fc3cf2dd-2491-4541-a245-5eefc9a5c042 tempest-TenantUsagesTestJSON-970395509 tempest-TenantUsagesTestJSON-970395509-project-member] [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 565.337230] env[62066]: DEBUG oslo_concurrency.lockutils [None req-8d368cc7-6ac0-44cd-b71e-9136d8337cb4 tempest-AttachInterfacesUnderV243Test-1832696447 tempest-AttachInterfacesUnderV243Test-1832696447-project-member] Acquiring lock "f91a90f7-be73-424c-966c-1be6f37a0864" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 565.337230] env[62066]: DEBUG oslo_concurrency.lockutils [None req-8d368cc7-6ac0-44cd-b71e-9136d8337cb4 tempest-AttachInterfacesUnderV243Test-1832696447 tempest-AttachInterfacesUnderV243Test-1832696447-project-member] Lock "f91a90f7-be73-424c-966c-1be6f37a0864" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 565.466242] env[62066]: DEBUG nova.compute.manager [None req-65bf05da-7d56-4074-8d13-63fde71b9619 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 565.491580] env[62066]: DEBUG nova.virt.hardware [None req-65bf05da-7d56-4074-8d13-63fde71b9619 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-23T13:40:41Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-23T13:40:23Z,direct_url=,disk_format='vmdk',id=50ff584c-3b50-4395-af07-3e66769bc9f7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='eb52f7069a374c61ae946f052007c6d9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-23T13:40:24Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 565.491804] env[62066]: DEBUG nova.virt.hardware [None req-65bf05da-7d56-4074-8d13-63fde71b9619 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 565.492033] env[62066]: DEBUG nova.virt.hardware [None req-65bf05da-7d56-4074-8d13-63fde71b9619 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 565.492200] env[62066]: DEBUG nova.virt.hardware [None req-65bf05da-7d56-4074-8d13-63fde71b9619 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 565.492346] env[62066]: DEBUG nova.virt.hardware [None req-65bf05da-7d56-4074-8d13-63fde71b9619 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 565.492489] env[62066]: DEBUG nova.virt.hardware [None req-65bf05da-7d56-4074-8d13-63fde71b9619 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 565.493765] env[62066]: DEBUG nova.virt.hardware [None req-65bf05da-7d56-4074-8d13-63fde71b9619 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 565.493955] env[62066]: DEBUG nova.virt.hardware [None req-65bf05da-7d56-4074-8d13-63fde71b9619 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 565.494137] env[62066]: DEBUG nova.virt.hardware [None req-65bf05da-7d56-4074-8d13-63fde71b9619 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 565.494306] env[62066]: DEBUG nova.virt.hardware [None req-65bf05da-7d56-4074-8d13-63fde71b9619 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 565.494573] env[62066]: DEBUG nova.virt.hardware [None req-65bf05da-7d56-4074-8d13-63fde71b9619 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 565.498984] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8200b7e8-cbab-4a53-8d43-1cd939e64ade {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.508610] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c04b1113-5a7b-4d45-853a-b38ee87ad198 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.561687] env[62066]: DEBUG oslo_concurrency.lockutils [None req-659c7f6d-78fb-4b73-9024-3866f9052a37 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.113s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 565.563982] env[62066]: DEBUG oslo_concurrency.lockutils [None req-67ee84a7-af8a-43fc-bff6-48212ddfc8ae tempest-ServersNegativeTestMultiTenantJSON-1325431503 tempest-ServersNegativeTestMultiTenantJSON-1325431503-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 29.777s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 565.581181] env[62066]: INFO nova.scheduler.client.report [None req-659c7f6d-78fb-4b73-9024-3866f9052a37 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] Deleted allocations for instance 4b34ff57-e137-4abc-8aed-76dd2ec8d313 [ 565.617401] env[62066]: DEBUG nova.compute.manager [req-307958cb-761e-4e5b-a4b5-b1058ba4d7bd req-e1974890-86c7-4a44-8d34-704661c37c91 service nova] [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] Received event network-changed-39ab0f84-58c5-4569-bf28-6dbda905c21d {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 565.617401] env[62066]: DEBUG nova.compute.manager [req-307958cb-761e-4e5b-a4b5-b1058ba4d7bd req-e1974890-86c7-4a44-8d34-704661c37c91 service nova] [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] Refreshing instance network info cache due to event network-changed-39ab0f84-58c5-4569-bf28-6dbda905c21d. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 565.617401] env[62066]: DEBUG oslo_concurrency.lockutils [req-307958cb-761e-4e5b-a4b5-b1058ba4d7bd req-e1974890-86c7-4a44-8d34-704661c37c91 service nova] Acquiring lock "refresh_cache-eacdd472-5bb0-48bc-8603-9e1d08b41f55" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 565.617401] env[62066]: DEBUG oslo_concurrency.lockutils [req-307958cb-761e-4e5b-a4b5-b1058ba4d7bd req-e1974890-86c7-4a44-8d34-704661c37c91 service nova] Acquired lock "refresh_cache-eacdd472-5bb0-48bc-8603-9e1d08b41f55" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 565.617401] env[62066]: DEBUG nova.network.neutron [req-307958cb-761e-4e5b-a4b5-b1058ba4d7bd req-e1974890-86c7-4a44-8d34-704661c37c91 service nova] [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] Refreshing network info cache for port 39ab0f84-58c5-4569-bf28-6dbda905c21d {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 565.691237] env[62066]: DEBUG oslo_concurrency.lockutils [None req-fc3cf2dd-2491-4541-a245-5eefc9a5c042 tempest-TenantUsagesTestJSON-970395509 tempest-TenantUsagesTestJSON-970395509-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 565.905714] env[62066]: ERROR nova.compute.manager [None req-65bf05da-7d56-4074-8d13-63fde71b9619 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 39ab0f84-58c5-4569-bf28-6dbda905c21d, please check neutron logs for more information. [ 565.905714] env[62066]: ERROR nova.compute.manager Traceback (most recent call last): [ 565.905714] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 565.905714] env[62066]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 565.905714] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 565.905714] env[62066]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 565.905714] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 565.905714] env[62066]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 565.905714] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 565.905714] env[62066]: ERROR nova.compute.manager self.force_reraise() [ 565.905714] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 565.905714] env[62066]: ERROR nova.compute.manager raise self.value [ 565.905714] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 565.905714] env[62066]: ERROR nova.compute.manager updated_port = self._update_port( [ 565.905714] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 565.905714] env[62066]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 565.906236] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 565.906236] env[62066]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 565.906236] env[62066]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 39ab0f84-58c5-4569-bf28-6dbda905c21d, please check neutron logs for more information. [ 565.906236] env[62066]: ERROR nova.compute.manager [ 565.906236] env[62066]: Traceback (most recent call last): [ 565.906236] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 565.906236] env[62066]: listener.cb(fileno) [ 565.906236] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 565.906236] env[62066]: result = function(*args, **kwargs) [ 565.906236] env[62066]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 565.906236] env[62066]: return func(*args, **kwargs) [ 565.906236] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 565.906236] env[62066]: raise e [ 565.906236] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 565.906236] env[62066]: nwinfo = self.network_api.allocate_for_instance( [ 565.906236] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 565.906236] env[62066]: created_port_ids = self._update_ports_for_instance( [ 565.906236] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 565.906236] env[62066]: with excutils.save_and_reraise_exception(): [ 565.906236] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 565.906236] env[62066]: self.force_reraise() [ 565.906236] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 565.906236] env[62066]: raise self.value [ 565.906236] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 565.906236] env[62066]: updated_port = self._update_port( [ 565.906236] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 565.906236] env[62066]: _ensure_no_port_binding_failure(port) [ 565.906236] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 565.906236] env[62066]: raise exception.PortBindingFailed(port_id=port['id']) [ 565.907107] env[62066]: nova.exception.PortBindingFailed: Binding failed for port 39ab0f84-58c5-4569-bf28-6dbda905c21d, please check neutron logs for more information. [ 565.907107] env[62066]: Removing descriptor: 20 [ 565.907107] env[62066]: ERROR nova.compute.manager [None req-65bf05da-7d56-4074-8d13-63fde71b9619 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 39ab0f84-58c5-4569-bf28-6dbda905c21d, please check neutron logs for more information. [ 565.907107] env[62066]: ERROR nova.compute.manager [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] Traceback (most recent call last): [ 565.907107] env[62066]: ERROR nova.compute.manager [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 565.907107] env[62066]: ERROR nova.compute.manager [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] yield resources [ 565.907107] env[62066]: ERROR nova.compute.manager [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 565.907107] env[62066]: ERROR nova.compute.manager [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] self.driver.spawn(context, instance, image_meta, [ 565.907107] env[62066]: ERROR nova.compute.manager [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 565.907107] env[62066]: ERROR nova.compute.manager [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] self._vmops.spawn(context, instance, image_meta, injected_files, [ 565.907107] env[62066]: ERROR nova.compute.manager [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 565.907107] env[62066]: ERROR nova.compute.manager [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] vm_ref = self.build_virtual_machine(instance, [ 565.907534] env[62066]: ERROR nova.compute.manager [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 565.907534] env[62066]: ERROR nova.compute.manager [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] vif_infos = vmwarevif.get_vif_info(self._session, [ 565.907534] env[62066]: ERROR nova.compute.manager [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 565.907534] env[62066]: ERROR nova.compute.manager [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] for vif in network_info: [ 565.907534] env[62066]: ERROR nova.compute.manager [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 565.907534] env[62066]: ERROR nova.compute.manager [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] return self._sync_wrapper(fn, *args, **kwargs) [ 565.907534] env[62066]: ERROR nova.compute.manager [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 565.907534] env[62066]: ERROR nova.compute.manager [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] self.wait() [ 565.907534] env[62066]: ERROR nova.compute.manager [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 565.907534] env[62066]: ERROR nova.compute.manager [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] self[:] = self._gt.wait() [ 565.907534] env[62066]: ERROR nova.compute.manager [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 565.907534] env[62066]: ERROR nova.compute.manager [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] return self._exit_event.wait() [ 565.907534] env[62066]: ERROR nova.compute.manager [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 565.907942] env[62066]: ERROR nova.compute.manager [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] result = hub.switch() [ 565.907942] env[62066]: ERROR nova.compute.manager [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 565.907942] env[62066]: ERROR nova.compute.manager [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] return self.greenlet.switch() [ 565.907942] env[62066]: ERROR nova.compute.manager [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 565.907942] env[62066]: ERROR nova.compute.manager [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] result = function(*args, **kwargs) [ 565.907942] env[62066]: ERROR nova.compute.manager [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 565.907942] env[62066]: ERROR nova.compute.manager [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] return func(*args, **kwargs) [ 565.907942] env[62066]: ERROR nova.compute.manager [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 565.907942] env[62066]: ERROR nova.compute.manager [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] raise e [ 565.907942] env[62066]: ERROR nova.compute.manager [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 565.907942] env[62066]: ERROR nova.compute.manager [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] nwinfo = self.network_api.allocate_for_instance( [ 565.907942] env[62066]: ERROR nova.compute.manager [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 565.907942] env[62066]: ERROR nova.compute.manager [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] created_port_ids = self._update_ports_for_instance( [ 565.908340] env[62066]: ERROR nova.compute.manager [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 565.908340] env[62066]: ERROR nova.compute.manager [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] with excutils.save_and_reraise_exception(): [ 565.908340] env[62066]: ERROR nova.compute.manager [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 565.908340] env[62066]: ERROR nova.compute.manager [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] self.force_reraise() [ 565.908340] env[62066]: ERROR nova.compute.manager [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 565.908340] env[62066]: ERROR nova.compute.manager [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] raise self.value [ 565.908340] env[62066]: ERROR nova.compute.manager [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 565.908340] env[62066]: ERROR nova.compute.manager [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] updated_port = self._update_port( [ 565.908340] env[62066]: ERROR nova.compute.manager [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 565.908340] env[62066]: ERROR nova.compute.manager [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] _ensure_no_port_binding_failure(port) [ 565.908340] env[62066]: ERROR nova.compute.manager [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 565.908340] env[62066]: ERROR nova.compute.manager [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] raise exception.PortBindingFailed(port_id=port['id']) [ 565.908699] env[62066]: ERROR nova.compute.manager [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] nova.exception.PortBindingFailed: Binding failed for port 39ab0f84-58c5-4569-bf28-6dbda905c21d, please check neutron logs for more information. [ 565.908699] env[62066]: ERROR nova.compute.manager [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] [ 565.908699] env[62066]: INFO nova.compute.manager [None req-65bf05da-7d56-4074-8d13-63fde71b9619 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] Terminating instance [ 565.909666] env[62066]: DEBUG oslo_concurrency.lockutils [None req-65bf05da-7d56-4074-8d13-63fde71b9619 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] Acquiring lock "refresh_cache-eacdd472-5bb0-48bc-8603-9e1d08b41f55" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 566.089152] env[62066]: DEBUG oslo_concurrency.lockutils [None req-659c7f6d-78fb-4b73-9024-3866f9052a37 tempest-ServerDiagnosticsV248Test-1870196446 tempest-ServerDiagnosticsV248Test-1870196446-project-member] Lock "4b34ff57-e137-4abc-8aed-76dd2ec8d313" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 35.548s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 566.142875] env[62066]: DEBUG nova.network.neutron [req-307958cb-761e-4e5b-a4b5-b1058ba4d7bd req-e1974890-86c7-4a44-8d34-704661c37c91 service nova] [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 566.316903] env[62066]: DEBUG nova.network.neutron [req-307958cb-761e-4e5b-a4b5-b1058ba4d7bd req-e1974890-86c7-4a44-8d34-704661c37c91 service nova] [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 566.494867] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1996f70-2b7c-44b3-a485-d98ecb18b6b6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.506046] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da407b4e-c6d6-47cc-b450-135e58c6a672 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.532881] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-153da8bf-d631-41a8-b757-18b4e074fb23 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.540398] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-947c4538-3bc0-4341-8e97-4fa1a51c4040 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.553581] env[62066]: DEBUG nova.compute.provider_tree [None req-67ee84a7-af8a-43fc-bff6-48212ddfc8ae tempest-ServersNegativeTestMultiTenantJSON-1325431503 tempest-ServersNegativeTestMultiTenantJSON-1325431503-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 566.821085] env[62066]: DEBUG oslo_concurrency.lockutils [req-307958cb-761e-4e5b-a4b5-b1058ba4d7bd req-e1974890-86c7-4a44-8d34-704661c37c91 service nova] Releasing lock "refresh_cache-eacdd472-5bb0-48bc-8603-9e1d08b41f55" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 566.821486] env[62066]: DEBUG oslo_concurrency.lockutils [None req-65bf05da-7d56-4074-8d13-63fde71b9619 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] Acquired lock "refresh_cache-eacdd472-5bb0-48bc-8603-9e1d08b41f55" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 566.821669] env[62066]: DEBUG nova.network.neutron [None req-65bf05da-7d56-4074-8d13-63fde71b9619 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 567.057261] env[62066]: DEBUG nova.scheduler.client.report [None req-67ee84a7-af8a-43fc-bff6-48212ddfc8ae tempest-ServersNegativeTestMultiTenantJSON-1325431503 tempest-ServersNegativeTestMultiTenantJSON-1325431503-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 567.350810] env[62066]: DEBUG nova.network.neutron [None req-65bf05da-7d56-4074-8d13-63fde71b9619 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 567.471486] env[62066]: DEBUG nova.network.neutron [None req-65bf05da-7d56-4074-8d13-63fde71b9619 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 567.563229] env[62066]: DEBUG oslo_concurrency.lockutils [None req-67ee84a7-af8a-43fc-bff6-48212ddfc8ae tempest-ServersNegativeTestMultiTenantJSON-1325431503 tempest-ServersNegativeTestMultiTenantJSON-1325431503-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.999s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 567.563888] env[62066]: ERROR nova.compute.manager [None req-67ee84a7-af8a-43fc-bff6-48212ddfc8ae tempest-ServersNegativeTestMultiTenantJSON-1325431503 tempest-ServersNegativeTestMultiTenantJSON-1325431503-project-member] [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 88f48a5f-c406-4d4f-8c70-39da5145d0f6, please check neutron logs for more information. [ 567.563888] env[62066]: ERROR nova.compute.manager [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] Traceback (most recent call last): [ 567.563888] env[62066]: ERROR nova.compute.manager [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 567.563888] env[62066]: ERROR nova.compute.manager [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] self.driver.spawn(context, instance, image_meta, [ 567.563888] env[62066]: ERROR nova.compute.manager [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 567.563888] env[62066]: ERROR nova.compute.manager [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] self._vmops.spawn(context, instance, image_meta, injected_files, [ 567.563888] env[62066]: ERROR nova.compute.manager [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 567.563888] env[62066]: ERROR nova.compute.manager [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] vm_ref = self.build_virtual_machine(instance, [ 567.563888] env[62066]: ERROR nova.compute.manager [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 567.563888] env[62066]: ERROR nova.compute.manager [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] vif_infos = vmwarevif.get_vif_info(self._session, [ 567.563888] env[62066]: ERROR nova.compute.manager [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 567.564272] env[62066]: ERROR nova.compute.manager [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] for vif in network_info: [ 567.564272] env[62066]: ERROR nova.compute.manager [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 567.564272] env[62066]: ERROR nova.compute.manager [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] return self._sync_wrapper(fn, *args, **kwargs) [ 567.564272] env[62066]: ERROR nova.compute.manager [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 567.564272] env[62066]: ERROR nova.compute.manager [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] self.wait() [ 567.564272] env[62066]: ERROR nova.compute.manager [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 567.564272] env[62066]: ERROR nova.compute.manager [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] self[:] = self._gt.wait() [ 567.564272] env[62066]: ERROR nova.compute.manager [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 567.564272] env[62066]: ERROR nova.compute.manager [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] return self._exit_event.wait() [ 567.564272] env[62066]: ERROR nova.compute.manager [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 567.564272] env[62066]: ERROR nova.compute.manager [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] result = hub.switch() [ 567.564272] env[62066]: ERROR nova.compute.manager [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 567.564272] env[62066]: ERROR nova.compute.manager [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] return self.greenlet.switch() [ 567.564661] env[62066]: ERROR nova.compute.manager [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 567.564661] env[62066]: ERROR nova.compute.manager [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] result = function(*args, **kwargs) [ 567.564661] env[62066]: ERROR nova.compute.manager [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 567.564661] env[62066]: ERROR nova.compute.manager [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] return func(*args, **kwargs) [ 567.564661] env[62066]: ERROR nova.compute.manager [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 567.564661] env[62066]: ERROR nova.compute.manager [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] raise e [ 567.564661] env[62066]: ERROR nova.compute.manager [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 567.564661] env[62066]: ERROR nova.compute.manager [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] nwinfo = self.network_api.allocate_for_instance( [ 567.564661] env[62066]: ERROR nova.compute.manager [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 567.564661] env[62066]: ERROR nova.compute.manager [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] created_port_ids = self._update_ports_for_instance( [ 567.564661] env[62066]: ERROR nova.compute.manager [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 567.564661] env[62066]: ERROR nova.compute.manager [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] with excutils.save_and_reraise_exception(): [ 567.564661] env[62066]: ERROR nova.compute.manager [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 567.565057] env[62066]: ERROR nova.compute.manager [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] self.force_reraise() [ 567.565057] env[62066]: ERROR nova.compute.manager [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 567.565057] env[62066]: ERROR nova.compute.manager [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] raise self.value [ 567.565057] env[62066]: ERROR nova.compute.manager [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 567.565057] env[62066]: ERROR nova.compute.manager [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] updated_port = self._update_port( [ 567.565057] env[62066]: ERROR nova.compute.manager [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 567.565057] env[62066]: ERROR nova.compute.manager [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] _ensure_no_port_binding_failure(port) [ 567.565057] env[62066]: ERROR nova.compute.manager [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 567.565057] env[62066]: ERROR nova.compute.manager [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] raise exception.PortBindingFailed(port_id=port['id']) [ 567.565057] env[62066]: ERROR nova.compute.manager [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] nova.exception.PortBindingFailed: Binding failed for port 88f48a5f-c406-4d4f-8c70-39da5145d0f6, please check neutron logs for more information. [ 567.565057] env[62066]: ERROR nova.compute.manager [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] [ 567.565456] env[62066]: DEBUG nova.compute.utils [None req-67ee84a7-af8a-43fc-bff6-48212ddfc8ae tempest-ServersNegativeTestMultiTenantJSON-1325431503 tempest-ServersNegativeTestMultiTenantJSON-1325431503-project-member] [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] Binding failed for port 88f48a5f-c406-4d4f-8c70-39da5145d0f6, please check neutron logs for more information. {{(pid=62066) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 567.569166] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a16a38bd-97b1-4d18-bd95-9847ae1f8463 tempest-ServersAdmin275Test-1729403672 tempest-ServersAdmin275Test-1729403672-project-admin] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 29.904s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 567.569166] env[62066]: DEBUG nova.objects.instance [None req-a16a38bd-97b1-4d18-bd95-9847ae1f8463 tempest-ServersAdmin275Test-1729403672 tempest-ServersAdmin275Test-1729403672-project-admin] [instance: 56eee633-2c01-4df2-8d2e-58ad712942f6] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62066) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 567.574391] env[62066]: DEBUG nova.compute.manager [None req-67ee84a7-af8a-43fc-bff6-48212ddfc8ae tempest-ServersNegativeTestMultiTenantJSON-1325431503 tempest-ServersNegativeTestMultiTenantJSON-1325431503-project-member] [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] Build of instance 95bb27b9-e5a3-418f-a078-ae22ebf40013 was re-scheduled: Binding failed for port 88f48a5f-c406-4d4f-8c70-39da5145d0f6, please check neutron logs for more information. {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 567.574391] env[62066]: DEBUG nova.compute.manager [None req-67ee84a7-af8a-43fc-bff6-48212ddfc8ae tempest-ServersNegativeTestMultiTenantJSON-1325431503 tempest-ServersNegativeTestMultiTenantJSON-1325431503-project-member] [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] Unplugging VIFs for instance {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 567.574391] env[62066]: DEBUG oslo_concurrency.lockutils [None req-67ee84a7-af8a-43fc-bff6-48212ddfc8ae tempest-ServersNegativeTestMultiTenantJSON-1325431503 tempest-ServersNegativeTestMultiTenantJSON-1325431503-project-member] Acquiring lock "refresh_cache-95bb27b9-e5a3-418f-a078-ae22ebf40013" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 567.574391] env[62066]: DEBUG oslo_concurrency.lockutils [None req-67ee84a7-af8a-43fc-bff6-48212ddfc8ae tempest-ServersNegativeTestMultiTenantJSON-1325431503 tempest-ServersNegativeTestMultiTenantJSON-1325431503-project-member] Acquired lock "refresh_cache-95bb27b9-e5a3-418f-a078-ae22ebf40013" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 567.574776] env[62066]: DEBUG nova.network.neutron [None req-67ee84a7-af8a-43fc-bff6-48212ddfc8ae tempest-ServersNegativeTestMultiTenantJSON-1325431503 tempest-ServersNegativeTestMultiTenantJSON-1325431503-project-member] [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 567.788791] env[62066]: DEBUG nova.compute.manager [req-385f3389-8f57-4340-a937-7a8714d6af6f req-1ed91cd9-fed5-4b0a-9293-0547fe6bb600 service nova] [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] Received event network-vif-deleted-39ab0f84-58c5-4569-bf28-6dbda905c21d {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 567.977016] env[62066]: DEBUG oslo_concurrency.lockutils [None req-65bf05da-7d56-4074-8d13-63fde71b9619 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] Releasing lock "refresh_cache-eacdd472-5bb0-48bc-8603-9e1d08b41f55" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 567.977016] env[62066]: DEBUG nova.compute.manager [None req-65bf05da-7d56-4074-8d13-63fde71b9619 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 567.977016] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-65bf05da-7d56-4074-8d13-63fde71b9619 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 567.977016] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-30041a32-62b5-40f1-a1e4-ddbb42a22496 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.989661] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74693bb3-e99f-45ee-9757-b49ea74de8ae {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.013620] env[62066]: WARNING nova.virt.vmwareapi.vmops [None req-65bf05da-7d56-4074-8d13-63fde71b9619 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance eacdd472-5bb0-48bc-8603-9e1d08b41f55 could not be found. [ 568.013833] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-65bf05da-7d56-4074-8d13-63fde71b9619 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 568.014032] env[62066]: INFO nova.compute.manager [None req-65bf05da-7d56-4074-8d13-63fde71b9619 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] Took 0.04 seconds to destroy the instance on the hypervisor. [ 568.014269] env[62066]: DEBUG oslo.service.loopingcall [None req-65bf05da-7d56-4074-8d13-63fde71b9619 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 568.014496] env[62066]: DEBUG nova.compute.manager [-] [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 568.014585] env[62066]: DEBUG nova.network.neutron [-] [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 568.036825] env[62066]: DEBUG nova.network.neutron [-] [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 568.108126] env[62066]: DEBUG nova.network.neutron [None req-67ee84a7-af8a-43fc-bff6-48212ddfc8ae tempest-ServersNegativeTestMultiTenantJSON-1325431503 tempest-ServersNegativeTestMultiTenantJSON-1325431503-project-member] [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 568.191607] env[62066]: DEBUG nova.network.neutron [None req-67ee84a7-af8a-43fc-bff6-48212ddfc8ae tempest-ServersNegativeTestMultiTenantJSON-1325431503 tempest-ServersNegativeTestMultiTenantJSON-1325431503-project-member] [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 568.539674] env[62066]: DEBUG nova.network.neutron [-] [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 568.581717] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a16a38bd-97b1-4d18-bd95-9847ae1f8463 tempest-ServersAdmin275Test-1729403672 tempest-ServersAdmin275Test-1729403672-project-admin] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.014s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 568.581717] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b105d8cd-4f01-4e61-a779-8b209763af2b tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 25.277s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 568.581717] env[62066]: DEBUG nova.objects.instance [None req-b105d8cd-4f01-4e61-a779-8b209763af2b tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Lazy-loading 'resources' on Instance uuid 56eee633-2c01-4df2-8d2e-58ad712942f6 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 568.695252] env[62066]: DEBUG oslo_concurrency.lockutils [None req-67ee84a7-af8a-43fc-bff6-48212ddfc8ae tempest-ServersNegativeTestMultiTenantJSON-1325431503 tempest-ServersNegativeTestMultiTenantJSON-1325431503-project-member] Releasing lock "refresh_cache-95bb27b9-e5a3-418f-a078-ae22ebf40013" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 568.695352] env[62066]: DEBUG nova.compute.manager [None req-67ee84a7-af8a-43fc-bff6-48212ddfc8ae tempest-ServersNegativeTestMultiTenantJSON-1325431503 tempest-ServersNegativeTestMultiTenantJSON-1325431503-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 568.695701] env[62066]: DEBUG nova.compute.manager [None req-67ee84a7-af8a-43fc-bff6-48212ddfc8ae tempest-ServersNegativeTestMultiTenantJSON-1325431503 tempest-ServersNegativeTestMultiTenantJSON-1325431503-project-member] [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 568.695701] env[62066]: DEBUG nova.network.neutron [None req-67ee84a7-af8a-43fc-bff6-48212ddfc8ae tempest-ServersNegativeTestMultiTenantJSON-1325431503 tempest-ServersNegativeTestMultiTenantJSON-1325431503-project-member] [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 568.727651] env[62066]: DEBUG nova.network.neutron [None req-67ee84a7-af8a-43fc-bff6-48212ddfc8ae tempest-ServersNegativeTestMultiTenantJSON-1325431503 tempest-ServersNegativeTestMultiTenantJSON-1325431503-project-member] [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 569.042570] env[62066]: INFO nova.compute.manager [-] [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] Took 1.03 seconds to deallocate network for instance. [ 569.045107] env[62066]: DEBUG nova.compute.claims [None req-65bf05da-7d56-4074-8d13-63fde71b9619 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] Aborting claim: {{(pid=62066) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 569.045328] env[62066]: DEBUG oslo_concurrency.lockutils [None req-65bf05da-7d56-4074-8d13-63fde71b9619 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 569.176306] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3d3acd2e-756a-442f-acf6-9a75ba7d933c tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Acquiring lock "0ff2f9f4-6d85-4465-8e0a-fb39f2a25c9b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 569.176593] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3d3acd2e-756a-442f-acf6-9a75ba7d933c tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Lock "0ff2f9f4-6d85-4465-8e0a-fb39f2a25c9b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 569.230256] env[62066]: DEBUG nova.network.neutron [None req-67ee84a7-af8a-43fc-bff6-48212ddfc8ae tempest-ServersNegativeTestMultiTenantJSON-1325431503 tempest-ServersNegativeTestMultiTenantJSON-1325431503-project-member] [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 569.504621] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42b741c1-df31-43c3-b75e-64075c9b2873 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.512155] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3a490f6-ee62-41a3-b311-e97c688bbb42 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.548838] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db2d25cf-a3bb-439c-89ab-d5fba31ca477 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.556539] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1c56c26-0539-4b0b-9dd5-6f35c36e0c13 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.572800] env[62066]: DEBUG nova.compute.provider_tree [None req-b105d8cd-4f01-4e61-a779-8b209763af2b tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 569.733506] env[62066]: INFO nova.compute.manager [None req-67ee84a7-af8a-43fc-bff6-48212ddfc8ae tempest-ServersNegativeTestMultiTenantJSON-1325431503 tempest-ServersNegativeTestMultiTenantJSON-1325431503-project-member] [instance: 95bb27b9-e5a3-418f-a078-ae22ebf40013] Took 1.04 seconds to deallocate network for instance. [ 570.076069] env[62066]: DEBUG nova.scheduler.client.report [None req-b105d8cd-4f01-4e61-a779-8b209763af2b tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 570.581190] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b105d8cd-4f01-4e61-a779-8b209763af2b tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 570.583476] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4ae0c9b4-adca-4b03-8ac7-409a4ee035f6 tempest-InstanceActionsTestJSON-1008103477 tempest-InstanceActionsTestJSON-1008103477-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.472s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 570.585982] env[62066]: INFO nova.compute.claims [None req-4ae0c9b4-adca-4b03-8ac7-409a4ee035f6 tempest-InstanceActionsTestJSON-1008103477 tempest-InstanceActionsTestJSON-1008103477-project-member] [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 570.608820] env[62066]: INFO nova.scheduler.client.report [None req-b105d8cd-4f01-4e61-a779-8b209763af2b tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Deleted allocations for instance 56eee633-2c01-4df2-8d2e-58ad712942f6 [ 570.767470] env[62066]: INFO nova.scheduler.client.report [None req-67ee84a7-af8a-43fc-bff6-48212ddfc8ae tempest-ServersNegativeTestMultiTenantJSON-1325431503 tempest-ServersNegativeTestMultiTenantJSON-1325431503-project-member] Deleted allocations for instance 95bb27b9-e5a3-418f-a078-ae22ebf40013 [ 571.118499] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b105d8cd-4f01-4e61-a779-8b209763af2b tempest-ServersAdmin275Test-1100995386 tempest-ServersAdmin275Test-1100995386-project-member] Lock "56eee633-2c01-4df2-8d2e-58ad712942f6" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 31.709s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 571.278176] env[62066]: DEBUG oslo_concurrency.lockutils [None req-67ee84a7-af8a-43fc-bff6-48212ddfc8ae tempest-ServersNegativeTestMultiTenantJSON-1325431503 tempest-ServersNegativeTestMultiTenantJSON-1325431503-project-member] Lock "95bb27b9-e5a3-418f-a078-ae22ebf40013" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 71.566s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 571.780466] env[62066]: DEBUG nova.compute.manager [None req-e7221104-9243-4d76-b03e-5782ae8f55d4 tempest-FloatingIPsAssociationNegativeTestJSON-961065153 tempest-FloatingIPsAssociationNegativeTestJSON-961065153-project-member] [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 572.093767] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4c20a79-e3bb-4742-a2f6-02d5773df55b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.101788] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-012e7f4d-8a21-46dd-b5a9-a6df5b90df9b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.131910] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3babe169-dc16-4e49-b857-246b7e9bb48e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.139842] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dedcdde4-bffe-4639-9fe4-37bd1edc9450 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.153194] env[62066]: DEBUG nova.compute.provider_tree [None req-4ae0c9b4-adca-4b03-8ac7-409a4ee035f6 tempest-InstanceActionsTestJSON-1008103477 tempest-InstanceActionsTestJSON-1008103477-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 572.314917] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e7221104-9243-4d76-b03e-5782ae8f55d4 tempest-FloatingIPsAssociationNegativeTestJSON-961065153 tempest-FloatingIPsAssociationNegativeTestJSON-961065153-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 572.629281] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5b5d7867-70eb-472d-8d1a-c2bb035ec6db tempest-InstanceActionsV221TestJSON-1098427857 tempest-InstanceActionsV221TestJSON-1098427857-project-member] Acquiring lock "d7de90c0-95e9-4899-92a0-81658831713a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 572.629480] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5b5d7867-70eb-472d-8d1a-c2bb035ec6db tempest-InstanceActionsV221TestJSON-1098427857 tempest-InstanceActionsV221TestJSON-1098427857-project-member] Lock "d7de90c0-95e9-4899-92a0-81658831713a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 572.661851] env[62066]: DEBUG nova.scheduler.client.report [None req-4ae0c9b4-adca-4b03-8ac7-409a4ee035f6 tempest-InstanceActionsTestJSON-1008103477 tempest-InstanceActionsTestJSON-1008103477-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 573.168481] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4ae0c9b4-adca-4b03-8ac7-409a4ee035f6 tempest-InstanceActionsTestJSON-1008103477 tempest-InstanceActionsTestJSON-1008103477-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.583s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 573.168481] env[62066]: DEBUG nova.compute.manager [None req-4ae0c9b4-adca-4b03-8ac7-409a4ee035f6 tempest-InstanceActionsTestJSON-1008103477 tempest-InstanceActionsTestJSON-1008103477-project-member] [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 573.172033] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cba923da-ec24-4fd6-9695-23f810adaa32 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.081s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 573.175791] env[62066]: INFO nova.compute.claims [None req-cba923da-ec24-4fd6-9695-23f810adaa32 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 573.678114] env[62066]: DEBUG nova.compute.utils [None req-4ae0c9b4-adca-4b03-8ac7-409a4ee035f6 tempest-InstanceActionsTestJSON-1008103477 tempest-InstanceActionsTestJSON-1008103477-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 573.679502] env[62066]: DEBUG nova.compute.manager [None req-4ae0c9b4-adca-4b03-8ac7-409a4ee035f6 tempest-InstanceActionsTestJSON-1008103477 tempest-InstanceActionsTestJSON-1008103477-project-member] [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 573.683018] env[62066]: DEBUG nova.network.neutron [None req-4ae0c9b4-adca-4b03-8ac7-409a4ee035f6 tempest-InstanceActionsTestJSON-1008103477 tempest-InstanceActionsTestJSON-1008103477-project-member] [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 573.768945] env[62066]: DEBUG nova.policy [None req-4ae0c9b4-adca-4b03-8ac7-409a4ee035f6 tempest-InstanceActionsTestJSON-1008103477 tempest-InstanceActionsTestJSON-1008103477-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '26505fa868434735a6746b9028569b76', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd5c3f4493d0a4d2e9a1ef9198282067b', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 574.186257] env[62066]: DEBUG nova.compute.manager [None req-4ae0c9b4-adca-4b03-8ac7-409a4ee035f6 tempest-InstanceActionsTestJSON-1008103477 tempest-InstanceActionsTestJSON-1008103477-project-member] [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 574.211201] env[62066]: DEBUG nova.network.neutron [None req-4ae0c9b4-adca-4b03-8ac7-409a4ee035f6 tempest-InstanceActionsTestJSON-1008103477 tempest-InstanceActionsTestJSON-1008103477-project-member] [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] Successfully created port: 895cb3c4-534b-41ae-8230-af8fc3c155f1 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 574.700284] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe99695c-e4e9-4e89-8bca-32f464f09942 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.709475] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-370b698f-3a32-4e41-b0d4-d3a73c86ca05 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.742123] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-addf6e12-8265-427c-ab31-38113dd4d42d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.750038] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c10b5a7c-7dae-4415-a7de-a274ba21148b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.766940] env[62066]: DEBUG nova.compute.provider_tree [None req-cba923da-ec24-4fd6-9695-23f810adaa32 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 575.196346] env[62066]: DEBUG nova.compute.manager [None req-4ae0c9b4-adca-4b03-8ac7-409a4ee035f6 tempest-InstanceActionsTestJSON-1008103477 tempest-InstanceActionsTestJSON-1008103477-project-member] [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 575.229783] env[62066]: DEBUG nova.virt.hardware [None req-4ae0c9b4-adca-4b03-8ac7-409a4ee035f6 tempest-InstanceActionsTestJSON-1008103477 tempest-InstanceActionsTestJSON-1008103477-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-23T13:40:41Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-23T13:40:23Z,direct_url=,disk_format='vmdk',id=50ff584c-3b50-4395-af07-3e66769bc9f7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='eb52f7069a374c61ae946f052007c6d9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-23T13:40:24Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 575.230252] env[62066]: DEBUG nova.virt.hardware [None req-4ae0c9b4-adca-4b03-8ac7-409a4ee035f6 tempest-InstanceActionsTestJSON-1008103477 tempest-InstanceActionsTestJSON-1008103477-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 575.230683] env[62066]: DEBUG nova.virt.hardware [None req-4ae0c9b4-adca-4b03-8ac7-409a4ee035f6 tempest-InstanceActionsTestJSON-1008103477 tempest-InstanceActionsTestJSON-1008103477-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 575.231169] env[62066]: DEBUG nova.virt.hardware [None req-4ae0c9b4-adca-4b03-8ac7-409a4ee035f6 tempest-InstanceActionsTestJSON-1008103477 tempest-InstanceActionsTestJSON-1008103477-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 575.231423] env[62066]: DEBUG nova.virt.hardware [None req-4ae0c9b4-adca-4b03-8ac7-409a4ee035f6 tempest-InstanceActionsTestJSON-1008103477 tempest-InstanceActionsTestJSON-1008103477-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 575.231661] env[62066]: DEBUG nova.virt.hardware [None req-4ae0c9b4-adca-4b03-8ac7-409a4ee035f6 tempest-InstanceActionsTestJSON-1008103477 tempest-InstanceActionsTestJSON-1008103477-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 575.231968] env[62066]: DEBUG nova.virt.hardware [None req-4ae0c9b4-adca-4b03-8ac7-409a4ee035f6 tempest-InstanceActionsTestJSON-1008103477 tempest-InstanceActionsTestJSON-1008103477-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 575.232312] env[62066]: DEBUG nova.virt.hardware [None req-4ae0c9b4-adca-4b03-8ac7-409a4ee035f6 tempest-InstanceActionsTestJSON-1008103477 tempest-InstanceActionsTestJSON-1008103477-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 575.232603] env[62066]: DEBUG nova.virt.hardware [None req-4ae0c9b4-adca-4b03-8ac7-409a4ee035f6 tempest-InstanceActionsTestJSON-1008103477 tempest-InstanceActionsTestJSON-1008103477-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 575.233086] env[62066]: DEBUG nova.virt.hardware [None req-4ae0c9b4-adca-4b03-8ac7-409a4ee035f6 tempest-InstanceActionsTestJSON-1008103477 tempest-InstanceActionsTestJSON-1008103477-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 575.233366] env[62066]: DEBUG nova.virt.hardware [None req-4ae0c9b4-adca-4b03-8ac7-409a4ee035f6 tempest-InstanceActionsTestJSON-1008103477 tempest-InstanceActionsTestJSON-1008103477-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 575.234352] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-448e0c07-360d-4fca-a14a-a2171cc58c56 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.243582] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-923c94f0-8298-4dd3-b087-2a6c5bf45072 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.269406] env[62066]: DEBUG nova.scheduler.client.report [None req-cba923da-ec24-4fd6-9695-23f810adaa32 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 575.409214] env[62066]: DEBUG nova.compute.manager [req-d80a34fa-9964-49e3-bad1-616eac398673 req-49379aae-0a16-424e-a8b0-e02be3c5bc87 service nova] [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] Received event network-changed-895cb3c4-534b-41ae-8230-af8fc3c155f1 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 575.410165] env[62066]: DEBUG nova.compute.manager [req-d80a34fa-9964-49e3-bad1-616eac398673 req-49379aae-0a16-424e-a8b0-e02be3c5bc87 service nova] [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] Refreshing instance network info cache due to event network-changed-895cb3c4-534b-41ae-8230-af8fc3c155f1. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 575.410443] env[62066]: DEBUG oslo_concurrency.lockutils [req-d80a34fa-9964-49e3-bad1-616eac398673 req-49379aae-0a16-424e-a8b0-e02be3c5bc87 service nova] Acquiring lock "refresh_cache-454a790d-3cb2-4ef2-995f-97e0eb1e2469" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 575.410636] env[62066]: DEBUG oslo_concurrency.lockutils [req-d80a34fa-9964-49e3-bad1-616eac398673 req-49379aae-0a16-424e-a8b0-e02be3c5bc87 service nova] Acquired lock "refresh_cache-454a790d-3cb2-4ef2-995f-97e0eb1e2469" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 575.410824] env[62066]: DEBUG nova.network.neutron [req-d80a34fa-9964-49e3-bad1-616eac398673 req-49379aae-0a16-424e-a8b0-e02be3c5bc87 service nova] [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] Refreshing network info cache for port 895cb3c4-534b-41ae-8230-af8fc3c155f1 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 575.557547] env[62066]: ERROR nova.compute.manager [None req-4ae0c9b4-adca-4b03-8ac7-409a4ee035f6 tempest-InstanceActionsTestJSON-1008103477 tempest-InstanceActionsTestJSON-1008103477-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 895cb3c4-534b-41ae-8230-af8fc3c155f1, please check neutron logs for more information. [ 575.557547] env[62066]: ERROR nova.compute.manager Traceback (most recent call last): [ 575.557547] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 575.557547] env[62066]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 575.557547] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 575.557547] env[62066]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 575.557547] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 575.557547] env[62066]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 575.557547] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 575.557547] env[62066]: ERROR nova.compute.manager self.force_reraise() [ 575.557547] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 575.557547] env[62066]: ERROR nova.compute.manager raise self.value [ 575.557547] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 575.557547] env[62066]: ERROR nova.compute.manager updated_port = self._update_port( [ 575.557547] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 575.557547] env[62066]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 575.557905] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 575.557905] env[62066]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 575.557905] env[62066]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 895cb3c4-534b-41ae-8230-af8fc3c155f1, please check neutron logs for more information. [ 575.557905] env[62066]: ERROR nova.compute.manager [ 575.557905] env[62066]: Traceback (most recent call last): [ 575.557905] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 575.557905] env[62066]: listener.cb(fileno) [ 575.557905] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 575.557905] env[62066]: result = function(*args, **kwargs) [ 575.557905] env[62066]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 575.557905] env[62066]: return func(*args, **kwargs) [ 575.557905] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 575.557905] env[62066]: raise e [ 575.557905] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 575.557905] env[62066]: nwinfo = self.network_api.allocate_for_instance( [ 575.557905] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 575.557905] env[62066]: created_port_ids = self._update_ports_for_instance( [ 575.557905] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 575.557905] env[62066]: with excutils.save_and_reraise_exception(): [ 575.557905] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 575.557905] env[62066]: self.force_reraise() [ 575.557905] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 575.557905] env[62066]: raise self.value [ 575.557905] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 575.557905] env[62066]: updated_port = self._update_port( [ 575.557905] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 575.557905] env[62066]: _ensure_no_port_binding_failure(port) [ 575.557905] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 575.557905] env[62066]: raise exception.PortBindingFailed(port_id=port['id']) [ 575.558646] env[62066]: nova.exception.PortBindingFailed: Binding failed for port 895cb3c4-534b-41ae-8230-af8fc3c155f1, please check neutron logs for more information. [ 575.558646] env[62066]: Removing descriptor: 20 [ 575.558646] env[62066]: ERROR nova.compute.manager [None req-4ae0c9b4-adca-4b03-8ac7-409a4ee035f6 tempest-InstanceActionsTestJSON-1008103477 tempest-InstanceActionsTestJSON-1008103477-project-member] [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 895cb3c4-534b-41ae-8230-af8fc3c155f1, please check neutron logs for more information. [ 575.558646] env[62066]: ERROR nova.compute.manager [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] Traceback (most recent call last): [ 575.558646] env[62066]: ERROR nova.compute.manager [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 575.558646] env[62066]: ERROR nova.compute.manager [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] yield resources [ 575.558646] env[62066]: ERROR nova.compute.manager [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 575.558646] env[62066]: ERROR nova.compute.manager [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] self.driver.spawn(context, instance, image_meta, [ 575.558646] env[62066]: ERROR nova.compute.manager [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 575.558646] env[62066]: ERROR nova.compute.manager [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] self._vmops.spawn(context, instance, image_meta, injected_files, [ 575.558646] env[62066]: ERROR nova.compute.manager [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 575.558646] env[62066]: ERROR nova.compute.manager [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] vm_ref = self.build_virtual_machine(instance, [ 575.558909] env[62066]: ERROR nova.compute.manager [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 575.558909] env[62066]: ERROR nova.compute.manager [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] vif_infos = vmwarevif.get_vif_info(self._session, [ 575.558909] env[62066]: ERROR nova.compute.manager [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 575.558909] env[62066]: ERROR nova.compute.manager [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] for vif in network_info: [ 575.558909] env[62066]: ERROR nova.compute.manager [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 575.558909] env[62066]: ERROR nova.compute.manager [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] return self._sync_wrapper(fn, *args, **kwargs) [ 575.558909] env[62066]: ERROR nova.compute.manager [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 575.558909] env[62066]: ERROR nova.compute.manager [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] self.wait() [ 575.558909] env[62066]: ERROR nova.compute.manager [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 575.558909] env[62066]: ERROR nova.compute.manager [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] self[:] = self._gt.wait() [ 575.558909] env[62066]: ERROR nova.compute.manager [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 575.558909] env[62066]: ERROR nova.compute.manager [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] return self._exit_event.wait() [ 575.558909] env[62066]: ERROR nova.compute.manager [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 575.559194] env[62066]: ERROR nova.compute.manager [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] result = hub.switch() [ 575.559194] env[62066]: ERROR nova.compute.manager [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 575.559194] env[62066]: ERROR nova.compute.manager [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] return self.greenlet.switch() [ 575.559194] env[62066]: ERROR nova.compute.manager [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 575.559194] env[62066]: ERROR nova.compute.manager [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] result = function(*args, **kwargs) [ 575.559194] env[62066]: ERROR nova.compute.manager [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 575.559194] env[62066]: ERROR nova.compute.manager [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] return func(*args, **kwargs) [ 575.559194] env[62066]: ERROR nova.compute.manager [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 575.559194] env[62066]: ERROR nova.compute.manager [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] raise e [ 575.559194] env[62066]: ERROR nova.compute.manager [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 575.559194] env[62066]: ERROR nova.compute.manager [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] nwinfo = self.network_api.allocate_for_instance( [ 575.559194] env[62066]: ERROR nova.compute.manager [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 575.559194] env[62066]: ERROR nova.compute.manager [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] created_port_ids = self._update_ports_for_instance( [ 575.559495] env[62066]: ERROR nova.compute.manager [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 575.559495] env[62066]: ERROR nova.compute.manager [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] with excutils.save_and_reraise_exception(): [ 575.559495] env[62066]: ERROR nova.compute.manager [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 575.559495] env[62066]: ERROR nova.compute.manager [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] self.force_reraise() [ 575.559495] env[62066]: ERROR nova.compute.manager [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 575.559495] env[62066]: ERROR nova.compute.manager [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] raise self.value [ 575.559495] env[62066]: ERROR nova.compute.manager [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 575.559495] env[62066]: ERROR nova.compute.manager [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] updated_port = self._update_port( [ 575.559495] env[62066]: ERROR nova.compute.manager [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 575.559495] env[62066]: ERROR nova.compute.manager [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] _ensure_no_port_binding_failure(port) [ 575.559495] env[62066]: ERROR nova.compute.manager [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 575.559495] env[62066]: ERROR nova.compute.manager [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] raise exception.PortBindingFailed(port_id=port['id']) [ 575.559726] env[62066]: ERROR nova.compute.manager [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] nova.exception.PortBindingFailed: Binding failed for port 895cb3c4-534b-41ae-8230-af8fc3c155f1, please check neutron logs for more information. [ 575.559726] env[62066]: ERROR nova.compute.manager [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] [ 575.559726] env[62066]: INFO nova.compute.manager [None req-4ae0c9b4-adca-4b03-8ac7-409a4ee035f6 tempest-InstanceActionsTestJSON-1008103477 tempest-InstanceActionsTestJSON-1008103477-project-member] [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] Terminating instance [ 575.561178] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4ae0c9b4-adca-4b03-8ac7-409a4ee035f6 tempest-InstanceActionsTestJSON-1008103477 tempest-InstanceActionsTestJSON-1008103477-project-member] Acquiring lock "refresh_cache-454a790d-3cb2-4ef2-995f-97e0eb1e2469" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 575.775900] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cba923da-ec24-4fd6-9695-23f810adaa32 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.605s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 575.776682] env[62066]: DEBUG nova.compute.manager [None req-cba923da-ec24-4fd6-9695-23f810adaa32 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 575.782362] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d274d4c0-1d30-44ad-b71d-9f542d291186 tempest-VolumesAssistedSnapshotsTest-1416743605 tempest-VolumesAssistedSnapshotsTest-1416743605-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 22.321s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 575.930136] env[62066]: DEBUG nova.network.neutron [req-d80a34fa-9964-49e3-bad1-616eac398673 req-49379aae-0a16-424e-a8b0-e02be3c5bc87 service nova] [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 576.030565] env[62066]: DEBUG nova.network.neutron [req-d80a34fa-9964-49e3-bad1-616eac398673 req-49379aae-0a16-424e-a8b0-e02be3c5bc87 service nova] [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 576.287353] env[62066]: DEBUG nova.compute.utils [None req-cba923da-ec24-4fd6-9695-23f810adaa32 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 576.294461] env[62066]: DEBUG nova.compute.manager [None req-cba923da-ec24-4fd6-9695-23f810adaa32 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 576.294461] env[62066]: DEBUG nova.network.neutron [None req-cba923da-ec24-4fd6-9695-23f810adaa32 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 576.350888] env[62066]: DEBUG nova.policy [None req-cba923da-ec24-4fd6-9695-23f810adaa32 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '617259d206d647fbb20653e3d91083b3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '504e1c99f04c4d4fab8634c89cb017f4', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 576.534782] env[62066]: DEBUG oslo_concurrency.lockutils [req-d80a34fa-9964-49e3-bad1-616eac398673 req-49379aae-0a16-424e-a8b0-e02be3c5bc87 service nova] Releasing lock "refresh_cache-454a790d-3cb2-4ef2-995f-97e0eb1e2469" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 576.534782] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4ae0c9b4-adca-4b03-8ac7-409a4ee035f6 tempest-InstanceActionsTestJSON-1008103477 tempest-InstanceActionsTestJSON-1008103477-project-member] Acquired lock "refresh_cache-454a790d-3cb2-4ef2-995f-97e0eb1e2469" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 576.534782] env[62066]: DEBUG nova.network.neutron [None req-4ae0c9b4-adca-4b03-8ac7-409a4ee035f6 tempest-InstanceActionsTestJSON-1008103477 tempest-InstanceActionsTestJSON-1008103477-project-member] [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 576.797367] env[62066]: DEBUG nova.compute.manager [None req-cba923da-ec24-4fd6-9695-23f810adaa32 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 576.837953] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc17b3ce-181b-45e1-98a9-8617108d8270 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.846862] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8c0a8ac-49dd-46ba-9a65-52f9736b74af {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.878725] env[62066]: DEBUG nova.network.neutron [None req-cba923da-ec24-4fd6-9695-23f810adaa32 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] Successfully created port: 19cd4055-4906-4fce-a9a8-b6d1dffd27f4 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 576.881101] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-047db2b6-4c95-4f61-bea8-9fb76c010419 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.888970] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e1b759f-972f-49ae-866c-ca176129d38d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.901967] env[62066]: DEBUG nova.compute.provider_tree [None req-d274d4c0-1d30-44ad-b71d-9f542d291186 tempest-VolumesAssistedSnapshotsTest-1416743605 tempest-VolumesAssistedSnapshotsTest-1416743605-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 577.055484] env[62066]: DEBUG nova.network.neutron [None req-4ae0c9b4-adca-4b03-8ac7-409a4ee035f6 tempest-InstanceActionsTestJSON-1008103477 tempest-InstanceActionsTestJSON-1008103477-project-member] [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 577.174353] env[62066]: DEBUG nova.network.neutron [None req-4ae0c9b4-adca-4b03-8ac7-409a4ee035f6 tempest-InstanceActionsTestJSON-1008103477 tempest-InstanceActionsTestJSON-1008103477-project-member] [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 577.406726] env[62066]: DEBUG nova.scheduler.client.report [None req-d274d4c0-1d30-44ad-b71d-9f542d291186 tempest-VolumesAssistedSnapshotsTest-1416743605 tempest-VolumesAssistedSnapshotsTest-1416743605-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 577.678049] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4ae0c9b4-adca-4b03-8ac7-409a4ee035f6 tempest-InstanceActionsTestJSON-1008103477 tempest-InstanceActionsTestJSON-1008103477-project-member] Releasing lock "refresh_cache-454a790d-3cb2-4ef2-995f-97e0eb1e2469" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 577.678351] env[62066]: DEBUG nova.compute.manager [None req-4ae0c9b4-adca-4b03-8ac7-409a4ee035f6 tempest-InstanceActionsTestJSON-1008103477 tempest-InstanceActionsTestJSON-1008103477-project-member] [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 577.679303] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-4ae0c9b4-adca-4b03-8ac7-409a4ee035f6 tempest-InstanceActionsTestJSON-1008103477 tempest-InstanceActionsTestJSON-1008103477-project-member] [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 577.680172] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4bcefb88-3dd9-4656-9445-f6f767926dcb {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.694104] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2a2d7b4-1a3b-4350-971c-cb981f26d278 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.732842] env[62066]: WARNING nova.virt.vmwareapi.vmops [None req-4ae0c9b4-adca-4b03-8ac7-409a4ee035f6 tempest-InstanceActionsTestJSON-1008103477 tempest-InstanceActionsTestJSON-1008103477-project-member] [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 454a790d-3cb2-4ef2-995f-97e0eb1e2469 could not be found. [ 577.733233] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-4ae0c9b4-adca-4b03-8ac7-409a4ee035f6 tempest-InstanceActionsTestJSON-1008103477 tempest-InstanceActionsTestJSON-1008103477-project-member] [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 577.733711] env[62066]: INFO nova.compute.manager [None req-4ae0c9b4-adca-4b03-8ac7-409a4ee035f6 tempest-InstanceActionsTestJSON-1008103477 tempest-InstanceActionsTestJSON-1008103477-project-member] [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] Took 0.05 seconds to destroy the instance on the hypervisor. [ 577.733911] env[62066]: DEBUG oslo.service.loopingcall [None req-4ae0c9b4-adca-4b03-8ac7-409a4ee035f6 tempest-InstanceActionsTestJSON-1008103477 tempest-InstanceActionsTestJSON-1008103477-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 577.734267] env[62066]: DEBUG nova.compute.manager [-] [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 577.734415] env[62066]: DEBUG nova.network.neutron [-] [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 577.759672] env[62066]: DEBUG nova.network.neutron [-] [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 577.809499] env[62066]: DEBUG nova.compute.manager [None req-cba923da-ec24-4fd6-9695-23f810adaa32 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 577.846386] env[62066]: DEBUG nova.virt.hardware [None req-cba923da-ec24-4fd6-9695-23f810adaa32 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-23T13:40:43Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-23T13:40:23Z,direct_url=,disk_format='vmdk',id=50ff584c-3b50-4395-af07-3e66769bc9f7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='eb52f7069a374c61ae946f052007c6d9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-23T13:40:24Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 577.846820] env[62066]: DEBUG nova.virt.hardware [None req-cba923da-ec24-4fd6-9695-23f810adaa32 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 577.847091] env[62066]: DEBUG nova.virt.hardware [None req-cba923da-ec24-4fd6-9695-23f810adaa32 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 577.847329] env[62066]: DEBUG nova.virt.hardware [None req-cba923da-ec24-4fd6-9695-23f810adaa32 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 577.847511] env[62066]: DEBUG nova.virt.hardware [None req-cba923da-ec24-4fd6-9695-23f810adaa32 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 577.847690] env[62066]: DEBUG nova.virt.hardware [None req-cba923da-ec24-4fd6-9695-23f810adaa32 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 577.847956] env[62066]: DEBUG nova.virt.hardware [None req-cba923da-ec24-4fd6-9695-23f810adaa32 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 577.848212] env[62066]: DEBUG nova.virt.hardware [None req-cba923da-ec24-4fd6-9695-23f810adaa32 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 577.848405] env[62066]: DEBUG nova.virt.hardware [None req-cba923da-ec24-4fd6-9695-23f810adaa32 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 577.848604] env[62066]: DEBUG nova.virt.hardware [None req-cba923da-ec24-4fd6-9695-23f810adaa32 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 577.848802] env[62066]: DEBUG nova.virt.hardware [None req-cba923da-ec24-4fd6-9695-23f810adaa32 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 577.850186] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c529f426-a492-4168-9327-c8da28d5e32c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.866441] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9ed7ff6-2b2b-49f6-b819-5445c466d4f7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.913544] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d274d4c0-1d30-44ad-b71d-9f542d291186 tempest-VolumesAssistedSnapshotsTest-1416743605 tempest-VolumesAssistedSnapshotsTest-1416743605-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.133s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 577.914185] env[62066]: ERROR nova.compute.manager [None req-d274d4c0-1d30-44ad-b71d-9f542d291186 tempest-VolumesAssistedSnapshotsTest-1416743605 tempest-VolumesAssistedSnapshotsTest-1416743605-project-member] [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 4bb5d522-c456-4fd5-babb-0d4606b2df97, please check neutron logs for more information. [ 577.914185] env[62066]: ERROR nova.compute.manager [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] Traceback (most recent call last): [ 577.914185] env[62066]: ERROR nova.compute.manager [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 577.914185] env[62066]: ERROR nova.compute.manager [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] self.driver.spawn(context, instance, image_meta, [ 577.914185] env[62066]: ERROR nova.compute.manager [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 577.914185] env[62066]: ERROR nova.compute.manager [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] self._vmops.spawn(context, instance, image_meta, injected_files, [ 577.914185] env[62066]: ERROR nova.compute.manager [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 577.914185] env[62066]: ERROR nova.compute.manager [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] vm_ref = self.build_virtual_machine(instance, [ 577.914185] env[62066]: ERROR nova.compute.manager [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 577.914185] env[62066]: ERROR nova.compute.manager [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] vif_infos = vmwarevif.get_vif_info(self._session, [ 577.914185] env[62066]: ERROR nova.compute.manager [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 577.914955] env[62066]: ERROR nova.compute.manager [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] for vif in network_info: [ 577.914955] env[62066]: ERROR nova.compute.manager [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 577.914955] env[62066]: ERROR nova.compute.manager [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] return self._sync_wrapper(fn, *args, **kwargs) [ 577.914955] env[62066]: ERROR nova.compute.manager [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 577.914955] env[62066]: ERROR nova.compute.manager [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] self.wait() [ 577.914955] env[62066]: ERROR nova.compute.manager [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 577.914955] env[62066]: ERROR nova.compute.manager [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] self[:] = self._gt.wait() [ 577.914955] env[62066]: ERROR nova.compute.manager [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 577.914955] env[62066]: ERROR nova.compute.manager [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] return self._exit_event.wait() [ 577.914955] env[62066]: ERROR nova.compute.manager [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 577.914955] env[62066]: ERROR nova.compute.manager [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] result = hub.switch() [ 577.914955] env[62066]: ERROR nova.compute.manager [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 577.914955] env[62066]: ERROR nova.compute.manager [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] return self.greenlet.switch() [ 577.915267] env[62066]: ERROR nova.compute.manager [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 577.915267] env[62066]: ERROR nova.compute.manager [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] result = function(*args, **kwargs) [ 577.915267] env[62066]: ERROR nova.compute.manager [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 577.915267] env[62066]: ERROR nova.compute.manager [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] return func(*args, **kwargs) [ 577.915267] env[62066]: ERROR nova.compute.manager [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 577.915267] env[62066]: ERROR nova.compute.manager [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] raise e [ 577.915267] env[62066]: ERROR nova.compute.manager [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 577.915267] env[62066]: ERROR nova.compute.manager [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] nwinfo = self.network_api.allocate_for_instance( [ 577.915267] env[62066]: ERROR nova.compute.manager [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 577.915267] env[62066]: ERROR nova.compute.manager [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] created_port_ids = self._update_ports_for_instance( [ 577.915267] env[62066]: ERROR nova.compute.manager [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 577.915267] env[62066]: ERROR nova.compute.manager [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] with excutils.save_and_reraise_exception(): [ 577.915267] env[62066]: ERROR nova.compute.manager [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 577.915532] env[62066]: ERROR nova.compute.manager [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] self.force_reraise() [ 577.915532] env[62066]: ERROR nova.compute.manager [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 577.915532] env[62066]: ERROR nova.compute.manager [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] raise self.value [ 577.915532] env[62066]: ERROR nova.compute.manager [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 577.915532] env[62066]: ERROR nova.compute.manager [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] updated_port = self._update_port( [ 577.915532] env[62066]: ERROR nova.compute.manager [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 577.915532] env[62066]: ERROR nova.compute.manager [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] _ensure_no_port_binding_failure(port) [ 577.915532] env[62066]: ERROR nova.compute.manager [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 577.915532] env[62066]: ERROR nova.compute.manager [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] raise exception.PortBindingFailed(port_id=port['id']) [ 577.915532] env[62066]: ERROR nova.compute.manager [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] nova.exception.PortBindingFailed: Binding failed for port 4bb5d522-c456-4fd5-babb-0d4606b2df97, please check neutron logs for more information. [ 577.915532] env[62066]: ERROR nova.compute.manager [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] [ 577.915762] env[62066]: DEBUG nova.compute.utils [None req-d274d4c0-1d30-44ad-b71d-9f542d291186 tempest-VolumesAssistedSnapshotsTest-1416743605 tempest-VolumesAssistedSnapshotsTest-1416743605-project-member] [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] Binding failed for port 4bb5d522-c456-4fd5-babb-0d4606b2df97, please check neutron logs for more information. {{(pid=62066) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 577.916191] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4fa8ac7b-ee00-4ef3-97aa-07a2c2ae6378 tempest-ServerExternalEventsTest-41922728 tempest-ServerExternalEventsTest-41922728-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.857s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 577.918111] env[62066]: INFO nova.compute.claims [None req-4fa8ac7b-ee00-4ef3-97aa-07a2c2ae6378 tempest-ServerExternalEventsTest-41922728 tempest-ServerExternalEventsTest-41922728-project-member] [instance: 39cd204a-49c2-4309-9030-555056ab125e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 577.920641] env[62066]: DEBUG nova.compute.manager [None req-d274d4c0-1d30-44ad-b71d-9f542d291186 tempest-VolumesAssistedSnapshotsTest-1416743605 tempest-VolumesAssistedSnapshotsTest-1416743605-project-member] [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] Build of instance 1fde7ae0-c88c-4d7f-b657-355c20e2cb10 was re-scheduled: Binding failed for port 4bb5d522-c456-4fd5-babb-0d4606b2df97, please check neutron logs for more information. {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 577.921855] env[62066]: DEBUG nova.compute.manager [None req-d274d4c0-1d30-44ad-b71d-9f542d291186 tempest-VolumesAssistedSnapshotsTest-1416743605 tempest-VolumesAssistedSnapshotsTest-1416743605-project-member] [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] Unplugging VIFs for instance {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 577.921855] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d274d4c0-1d30-44ad-b71d-9f542d291186 tempest-VolumesAssistedSnapshotsTest-1416743605 tempest-VolumesAssistedSnapshotsTest-1416743605-project-member] Acquiring lock "refresh_cache-1fde7ae0-c88c-4d7f-b657-355c20e2cb10" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 577.921855] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d274d4c0-1d30-44ad-b71d-9f542d291186 tempest-VolumesAssistedSnapshotsTest-1416743605 tempest-VolumesAssistedSnapshotsTest-1416743605-project-member] Acquired lock "refresh_cache-1fde7ae0-c88c-4d7f-b657-355c20e2cb10" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 577.921855] env[62066]: DEBUG nova.network.neutron [None req-d274d4c0-1d30-44ad-b71d-9f542d291186 tempest-VolumesAssistedSnapshotsTest-1416743605 tempest-VolumesAssistedSnapshotsTest-1416743605-project-member] [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 578.022745] env[62066]: DEBUG nova.compute.manager [req-8c3a6928-febe-44ce-835a-481975c3683b req-156c09e2-69f6-4b56-b5cd-9a334fe868ae service nova] [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] Received event network-vif-deleted-895cb3c4-534b-41ae-8230-af8fc3c155f1 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 578.263822] env[62066]: DEBUG nova.network.neutron [-] [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 578.455994] env[62066]: DEBUG nova.network.neutron [None req-d274d4c0-1d30-44ad-b71d-9f542d291186 tempest-VolumesAssistedSnapshotsTest-1416743605 tempest-VolumesAssistedSnapshotsTest-1416743605-project-member] [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 578.606422] env[62066]: DEBUG nova.network.neutron [None req-d274d4c0-1d30-44ad-b71d-9f542d291186 tempest-VolumesAssistedSnapshotsTest-1416743605 tempest-VolumesAssistedSnapshotsTest-1416743605-project-member] [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 578.766773] env[62066]: INFO nova.compute.manager [-] [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] Took 1.03 seconds to deallocate network for instance. [ 578.772688] env[62066]: DEBUG nova.compute.claims [None req-4ae0c9b4-adca-4b03-8ac7-409a4ee035f6 tempest-InstanceActionsTestJSON-1008103477 tempest-InstanceActionsTestJSON-1008103477-project-member] [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] Aborting claim: {{(pid=62066) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 578.773275] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4ae0c9b4-adca-4b03-8ac7-409a4ee035f6 tempest-InstanceActionsTestJSON-1008103477 tempest-InstanceActionsTestJSON-1008103477-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 578.866912] env[62066]: ERROR nova.compute.manager [None req-cba923da-ec24-4fd6-9695-23f810adaa32 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 19cd4055-4906-4fce-a9a8-b6d1dffd27f4, please check neutron logs for more information. [ 578.866912] env[62066]: ERROR nova.compute.manager Traceback (most recent call last): [ 578.866912] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 578.866912] env[62066]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 578.866912] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 578.866912] env[62066]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 578.866912] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 578.866912] env[62066]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 578.866912] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 578.866912] env[62066]: ERROR nova.compute.manager self.force_reraise() [ 578.866912] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 578.866912] env[62066]: ERROR nova.compute.manager raise self.value [ 578.866912] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 578.866912] env[62066]: ERROR nova.compute.manager updated_port = self._update_port( [ 578.866912] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 578.866912] env[62066]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 578.867366] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 578.867366] env[62066]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 578.867366] env[62066]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 19cd4055-4906-4fce-a9a8-b6d1dffd27f4, please check neutron logs for more information. [ 578.867366] env[62066]: ERROR nova.compute.manager [ 578.867366] env[62066]: Traceback (most recent call last): [ 578.867366] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 578.867366] env[62066]: listener.cb(fileno) [ 578.867366] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 578.867366] env[62066]: result = function(*args, **kwargs) [ 578.867366] env[62066]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 578.867366] env[62066]: return func(*args, **kwargs) [ 578.867366] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 578.867366] env[62066]: raise e [ 578.867366] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 578.867366] env[62066]: nwinfo = self.network_api.allocate_for_instance( [ 578.867366] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 578.867366] env[62066]: created_port_ids = self._update_ports_for_instance( [ 578.867366] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 578.867366] env[62066]: with excutils.save_and_reraise_exception(): [ 578.867366] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 578.867366] env[62066]: self.force_reraise() [ 578.867366] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 578.867366] env[62066]: raise self.value [ 578.867366] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 578.867366] env[62066]: updated_port = self._update_port( [ 578.867366] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 578.867366] env[62066]: _ensure_no_port_binding_failure(port) [ 578.867366] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 578.867366] env[62066]: raise exception.PortBindingFailed(port_id=port['id']) [ 578.868311] env[62066]: nova.exception.PortBindingFailed: Binding failed for port 19cd4055-4906-4fce-a9a8-b6d1dffd27f4, please check neutron logs for more information. [ 578.868311] env[62066]: Removing descriptor: 20 [ 578.868311] env[62066]: ERROR nova.compute.manager [None req-cba923da-ec24-4fd6-9695-23f810adaa32 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 19cd4055-4906-4fce-a9a8-b6d1dffd27f4, please check neutron logs for more information. [ 578.868311] env[62066]: ERROR nova.compute.manager [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] Traceback (most recent call last): [ 578.868311] env[62066]: ERROR nova.compute.manager [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 578.868311] env[62066]: ERROR nova.compute.manager [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] yield resources [ 578.868311] env[62066]: ERROR nova.compute.manager [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 578.868311] env[62066]: ERROR nova.compute.manager [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] self.driver.spawn(context, instance, image_meta, [ 578.868311] env[62066]: ERROR nova.compute.manager [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 578.868311] env[62066]: ERROR nova.compute.manager [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] self._vmops.spawn(context, instance, image_meta, injected_files, [ 578.868311] env[62066]: ERROR nova.compute.manager [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 578.868311] env[62066]: ERROR nova.compute.manager [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] vm_ref = self.build_virtual_machine(instance, [ 578.868706] env[62066]: ERROR nova.compute.manager [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 578.868706] env[62066]: ERROR nova.compute.manager [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] vif_infos = vmwarevif.get_vif_info(self._session, [ 578.868706] env[62066]: ERROR nova.compute.manager [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 578.868706] env[62066]: ERROR nova.compute.manager [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] for vif in network_info: [ 578.868706] env[62066]: ERROR nova.compute.manager [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 578.868706] env[62066]: ERROR nova.compute.manager [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] return self._sync_wrapper(fn, *args, **kwargs) [ 578.868706] env[62066]: ERROR nova.compute.manager [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 578.868706] env[62066]: ERROR nova.compute.manager [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] self.wait() [ 578.868706] env[62066]: ERROR nova.compute.manager [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 578.868706] env[62066]: ERROR nova.compute.manager [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] self[:] = self._gt.wait() [ 578.868706] env[62066]: ERROR nova.compute.manager [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 578.868706] env[62066]: ERROR nova.compute.manager [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] return self._exit_event.wait() [ 578.868706] env[62066]: ERROR nova.compute.manager [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 578.869009] env[62066]: ERROR nova.compute.manager [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] result = hub.switch() [ 578.869009] env[62066]: ERROR nova.compute.manager [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 578.869009] env[62066]: ERROR nova.compute.manager [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] return self.greenlet.switch() [ 578.869009] env[62066]: ERROR nova.compute.manager [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 578.869009] env[62066]: ERROR nova.compute.manager [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] result = function(*args, **kwargs) [ 578.869009] env[62066]: ERROR nova.compute.manager [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 578.869009] env[62066]: ERROR nova.compute.manager [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] return func(*args, **kwargs) [ 578.869009] env[62066]: ERROR nova.compute.manager [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 578.869009] env[62066]: ERROR nova.compute.manager [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] raise e [ 578.869009] env[62066]: ERROR nova.compute.manager [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 578.869009] env[62066]: ERROR nova.compute.manager [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] nwinfo = self.network_api.allocate_for_instance( [ 578.869009] env[62066]: ERROR nova.compute.manager [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 578.869009] env[62066]: ERROR nova.compute.manager [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] created_port_ids = self._update_ports_for_instance( [ 578.869361] env[62066]: ERROR nova.compute.manager [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 578.869361] env[62066]: ERROR nova.compute.manager [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] with excutils.save_and_reraise_exception(): [ 578.869361] env[62066]: ERROR nova.compute.manager [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 578.869361] env[62066]: ERROR nova.compute.manager [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] self.force_reraise() [ 578.869361] env[62066]: ERROR nova.compute.manager [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 578.869361] env[62066]: ERROR nova.compute.manager [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] raise self.value [ 578.869361] env[62066]: ERROR nova.compute.manager [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 578.869361] env[62066]: ERROR nova.compute.manager [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] updated_port = self._update_port( [ 578.869361] env[62066]: ERROR nova.compute.manager [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 578.869361] env[62066]: ERROR nova.compute.manager [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] _ensure_no_port_binding_failure(port) [ 578.869361] env[62066]: ERROR nova.compute.manager [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 578.869361] env[62066]: ERROR nova.compute.manager [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] raise exception.PortBindingFailed(port_id=port['id']) [ 578.869640] env[62066]: ERROR nova.compute.manager [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] nova.exception.PortBindingFailed: Binding failed for port 19cd4055-4906-4fce-a9a8-b6d1dffd27f4, please check neutron logs for more information. [ 578.869640] env[62066]: ERROR nova.compute.manager [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] [ 578.869640] env[62066]: INFO nova.compute.manager [None req-cba923da-ec24-4fd6-9695-23f810adaa32 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] Terminating instance [ 578.871285] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cba923da-ec24-4fd6-9695-23f810adaa32 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] Acquiring lock "refresh_cache-0a08bc46-0eea-4802-b0be-eb24f3507f73" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 578.871439] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cba923da-ec24-4fd6-9695-23f810adaa32 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] Acquired lock "refresh_cache-0a08bc46-0eea-4802-b0be-eb24f3507f73" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 578.871612] env[62066]: DEBUG nova.network.neutron [None req-cba923da-ec24-4fd6-9695-23f810adaa32 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 579.111582] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d274d4c0-1d30-44ad-b71d-9f542d291186 tempest-VolumesAssistedSnapshotsTest-1416743605 tempest-VolumesAssistedSnapshotsTest-1416743605-project-member] Releasing lock "refresh_cache-1fde7ae0-c88c-4d7f-b657-355c20e2cb10" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 579.111816] env[62066]: DEBUG nova.compute.manager [None req-d274d4c0-1d30-44ad-b71d-9f542d291186 tempest-VolumesAssistedSnapshotsTest-1416743605 tempest-VolumesAssistedSnapshotsTest-1416743605-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 579.111992] env[62066]: DEBUG nova.compute.manager [None req-d274d4c0-1d30-44ad-b71d-9f542d291186 tempest-VolumesAssistedSnapshotsTest-1416743605 tempest-VolumesAssistedSnapshotsTest-1416743605-project-member] [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 579.112178] env[62066]: DEBUG nova.network.neutron [None req-d274d4c0-1d30-44ad-b71d-9f542d291186 tempest-VolumesAssistedSnapshotsTest-1416743605 tempest-VolumesAssistedSnapshotsTest-1416743605-project-member] [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 579.134662] env[62066]: DEBUG nova.network.neutron [None req-d274d4c0-1d30-44ad-b71d-9f542d291186 tempest-VolumesAssistedSnapshotsTest-1416743605 tempest-VolumesAssistedSnapshotsTest-1416743605-project-member] [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 579.394965] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86b7fb0e-bf7c-4a20-ba0e-a87396de7e05 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.403498] env[62066]: DEBUG nova.network.neutron [None req-cba923da-ec24-4fd6-9695-23f810adaa32 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 579.406345] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33c06929-920f-47a6-8ba6-8b55f6a465f0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.438960] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bdf2d263-4220-45e3-8730-5a609b4d4637 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.446644] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8830cc2-f18c-4ed0-a3b1-dd4fd2409140 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.460352] env[62066]: DEBUG nova.compute.provider_tree [None req-4fa8ac7b-ee00-4ef3-97aa-07a2c2ae6378 tempest-ServerExternalEventsTest-41922728 tempest-ServerExternalEventsTest-41922728-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 579.509039] env[62066]: DEBUG nova.network.neutron [None req-cba923da-ec24-4fd6-9695-23f810adaa32 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 579.637296] env[62066]: DEBUG nova.network.neutron [None req-d274d4c0-1d30-44ad-b71d-9f542d291186 tempest-VolumesAssistedSnapshotsTest-1416743605 tempest-VolumesAssistedSnapshotsTest-1416743605-project-member] [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 579.963593] env[62066]: DEBUG nova.scheduler.client.report [None req-4fa8ac7b-ee00-4ef3-97aa-07a2c2ae6378 tempest-ServerExternalEventsTest-41922728 tempest-ServerExternalEventsTest-41922728-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 580.011697] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cba923da-ec24-4fd6-9695-23f810adaa32 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] Releasing lock "refresh_cache-0a08bc46-0eea-4802-b0be-eb24f3507f73" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 580.012180] env[62066]: DEBUG nova.compute.manager [None req-cba923da-ec24-4fd6-9695-23f810adaa32 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 580.012411] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-cba923da-ec24-4fd6-9695-23f810adaa32 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 580.012728] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9b601412-7c66-47b6-91d2-e0eb32609d69 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.021809] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad6f284f-4a49-41dc-8f46-ad2abe913cd3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.043371] env[62066]: WARNING nova.virt.vmwareapi.vmops [None req-cba923da-ec24-4fd6-9695-23f810adaa32 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 0a08bc46-0eea-4802-b0be-eb24f3507f73 could not be found. [ 580.043613] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-cba923da-ec24-4fd6-9695-23f810adaa32 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 580.043794] env[62066]: INFO nova.compute.manager [None req-cba923da-ec24-4fd6-9695-23f810adaa32 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] Took 0.03 seconds to destroy the instance on the hypervisor. [ 580.044046] env[62066]: DEBUG oslo.service.loopingcall [None req-cba923da-ec24-4fd6-9695-23f810adaa32 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 580.044257] env[62066]: DEBUG nova.compute.manager [-] [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 580.044346] env[62066]: DEBUG nova.network.neutron [-] [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 580.093245] env[62066]: DEBUG nova.network.neutron [-] [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 580.139956] env[62066]: INFO nova.compute.manager [None req-d274d4c0-1d30-44ad-b71d-9f542d291186 tempest-VolumesAssistedSnapshotsTest-1416743605 tempest-VolumesAssistedSnapshotsTest-1416743605-project-member] [instance: 1fde7ae0-c88c-4d7f-b657-355c20e2cb10] Took 1.03 seconds to deallocate network for instance. [ 580.305374] env[62066]: DEBUG nova.compute.manager [req-58178c27-32f8-4692-b7ec-fffb4870a5dd req-356544b9-de21-41aa-bbfd-c7504c54f143 service nova] [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] Received event network-changed-19cd4055-4906-4fce-a9a8-b6d1dffd27f4 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 580.305374] env[62066]: DEBUG nova.compute.manager [req-58178c27-32f8-4692-b7ec-fffb4870a5dd req-356544b9-de21-41aa-bbfd-c7504c54f143 service nova] [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] Refreshing instance network info cache due to event network-changed-19cd4055-4906-4fce-a9a8-b6d1dffd27f4. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 580.305563] env[62066]: DEBUG oslo_concurrency.lockutils [req-58178c27-32f8-4692-b7ec-fffb4870a5dd req-356544b9-de21-41aa-bbfd-c7504c54f143 service nova] Acquiring lock "refresh_cache-0a08bc46-0eea-4802-b0be-eb24f3507f73" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 580.305592] env[62066]: DEBUG oslo_concurrency.lockutils [req-58178c27-32f8-4692-b7ec-fffb4870a5dd req-356544b9-de21-41aa-bbfd-c7504c54f143 service nova] Acquired lock "refresh_cache-0a08bc46-0eea-4802-b0be-eb24f3507f73" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 580.305754] env[62066]: DEBUG nova.network.neutron [req-58178c27-32f8-4692-b7ec-fffb4870a5dd req-356544b9-de21-41aa-bbfd-c7504c54f143 service nova] [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] Refreshing network info cache for port 19cd4055-4906-4fce-a9a8-b6d1dffd27f4 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 580.472955] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4fa8ac7b-ee00-4ef3-97aa-07a2c2ae6378 tempest-ServerExternalEventsTest-41922728 tempest-ServerExternalEventsTest-41922728-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.557s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 580.473483] env[62066]: DEBUG nova.compute.manager [None req-4fa8ac7b-ee00-4ef3-97aa-07a2c2ae6378 tempest-ServerExternalEventsTest-41922728 tempest-ServerExternalEventsTest-41922728-project-member] [instance: 39cd204a-49c2-4309-9030-555056ab125e] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 580.477274] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c4a0feae-1b12-4166-bd78-82a96fe8ac4c tempest-ServersWithSpecificFlavorTestJSON-184391817 tempest-ServersWithSpecificFlavorTestJSON-184391817-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 20.283s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 580.596781] env[62066]: DEBUG nova.network.neutron [-] [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 580.829993] env[62066]: DEBUG nova.network.neutron [req-58178c27-32f8-4692-b7ec-fffb4870a5dd req-356544b9-de21-41aa-bbfd-c7504c54f143 service nova] [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 580.981971] env[62066]: DEBUG nova.network.neutron [req-58178c27-32f8-4692-b7ec-fffb4870a5dd req-356544b9-de21-41aa-bbfd-c7504c54f143 service nova] [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 580.993022] env[62066]: DEBUG nova.compute.utils [None req-4fa8ac7b-ee00-4ef3-97aa-07a2c2ae6378 tempest-ServerExternalEventsTest-41922728 tempest-ServerExternalEventsTest-41922728-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 580.993022] env[62066]: DEBUG nova.compute.manager [None req-4fa8ac7b-ee00-4ef3-97aa-07a2c2ae6378 tempest-ServerExternalEventsTest-41922728 tempest-ServerExternalEventsTest-41922728-project-member] [instance: 39cd204a-49c2-4309-9030-555056ab125e] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 580.993022] env[62066]: DEBUG nova.network.neutron [None req-4fa8ac7b-ee00-4ef3-97aa-07a2c2ae6378 tempest-ServerExternalEventsTest-41922728 tempest-ServerExternalEventsTest-41922728-project-member] [instance: 39cd204a-49c2-4309-9030-555056ab125e] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 581.076805] env[62066]: DEBUG nova.policy [None req-4fa8ac7b-ee00-4ef3-97aa-07a2c2ae6378 tempest-ServerExternalEventsTest-41922728 tempest-ServerExternalEventsTest-41922728-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6e3462833db24cfaa8a79f3d93bbe381', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'be4b2cfed9094bfc9c6175f7f55bfd9b', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 581.100368] env[62066]: INFO nova.compute.manager [-] [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] Took 1.06 seconds to deallocate network for instance. [ 581.103084] env[62066]: DEBUG nova.compute.claims [None req-cba923da-ec24-4fd6-9695-23f810adaa32 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] Aborting claim: {{(pid=62066) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 581.103292] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cba923da-ec24-4fd6-9695-23f810adaa32 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 581.177669] env[62066]: INFO nova.scheduler.client.report [None req-d274d4c0-1d30-44ad-b71d-9f542d291186 tempest-VolumesAssistedSnapshotsTest-1416743605 tempest-VolumesAssistedSnapshotsTest-1416743605-project-member] Deleted allocations for instance 1fde7ae0-c88c-4d7f-b657-355c20e2cb10 [ 581.485133] env[62066]: DEBUG oslo_concurrency.lockutils [req-58178c27-32f8-4692-b7ec-fffb4870a5dd req-356544b9-de21-41aa-bbfd-c7504c54f143 service nova] Releasing lock "refresh_cache-0a08bc46-0eea-4802-b0be-eb24f3507f73" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 581.485438] env[62066]: DEBUG nova.compute.manager [req-58178c27-32f8-4692-b7ec-fffb4870a5dd req-356544b9-de21-41aa-bbfd-c7504c54f143 service nova] [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] Received event network-vif-deleted-19cd4055-4906-4fce-a9a8-b6d1dffd27f4 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 581.500628] env[62066]: DEBUG nova.compute.manager [None req-4fa8ac7b-ee00-4ef3-97aa-07a2c2ae6378 tempest-ServerExternalEventsTest-41922728 tempest-ServerExternalEventsTest-41922728-project-member] [instance: 39cd204a-49c2-4309-9030-555056ab125e] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 581.540539] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc6aa447-1bbc-4f82-998f-40b68560048d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.553925] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-361e9abd-1d7f-44a6-b1a6-329cdea264db {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.587849] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23d6ae49-d6ef-456c-932a-9e94cfd55cfd {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.596944] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a78f6f6b-9928-44f2-849b-db84aac0c7ef {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.611088] env[62066]: DEBUG nova.compute.provider_tree [None req-c4a0feae-1b12-4166-bd78-82a96fe8ac4c tempest-ServersWithSpecificFlavorTestJSON-184391817 tempest-ServersWithSpecificFlavorTestJSON-184391817-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 581.661308] env[62066]: DEBUG nova.network.neutron [None req-4fa8ac7b-ee00-4ef3-97aa-07a2c2ae6378 tempest-ServerExternalEventsTest-41922728 tempest-ServerExternalEventsTest-41922728-project-member] [instance: 39cd204a-49c2-4309-9030-555056ab125e] Successfully created port: 00047463-06a3-4ffe-830b-2d4ef9e64de6 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 581.692119] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d274d4c0-1d30-44ad-b71d-9f542d291186 tempest-VolumesAssistedSnapshotsTest-1416743605 tempest-VolumesAssistedSnapshotsTest-1416743605-project-member] Lock "1fde7ae0-c88c-4d7f-b657-355c20e2cb10" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 74.330s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 582.114519] env[62066]: DEBUG nova.scheduler.client.report [None req-c4a0feae-1b12-4166-bd78-82a96fe8ac4c tempest-ServersWithSpecificFlavorTestJSON-184391817 tempest-ServersWithSpecificFlavorTestJSON-184391817-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 582.194618] env[62066]: DEBUG nova.compute.manager [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 582.387874] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5bf8bcd1-f041-4648-9b3c-4febff568fa9 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Acquiring lock "3624ca43-9ee3-418a-9205-debb699295df" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 582.388143] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5bf8bcd1-f041-4648-9b3c-4febff568fa9 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Lock "3624ca43-9ee3-418a-9205-debb699295df" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 582.514777] env[62066]: DEBUG nova.compute.manager [None req-4fa8ac7b-ee00-4ef3-97aa-07a2c2ae6378 tempest-ServerExternalEventsTest-41922728 tempest-ServerExternalEventsTest-41922728-project-member] [instance: 39cd204a-49c2-4309-9030-555056ab125e] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 582.539636] env[62066]: DEBUG nova.virt.hardware [None req-4fa8ac7b-ee00-4ef3-97aa-07a2c2ae6378 tempest-ServerExternalEventsTest-41922728 tempest-ServerExternalEventsTest-41922728-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-23T13:40:41Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-23T13:40:23Z,direct_url=,disk_format='vmdk',id=50ff584c-3b50-4395-af07-3e66769bc9f7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='eb52f7069a374c61ae946f052007c6d9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-23T13:40:24Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 582.539882] env[62066]: DEBUG nova.virt.hardware [None req-4fa8ac7b-ee00-4ef3-97aa-07a2c2ae6378 tempest-ServerExternalEventsTest-41922728 tempest-ServerExternalEventsTest-41922728-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 582.540057] env[62066]: DEBUG nova.virt.hardware [None req-4fa8ac7b-ee00-4ef3-97aa-07a2c2ae6378 tempest-ServerExternalEventsTest-41922728 tempest-ServerExternalEventsTest-41922728-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 582.540225] env[62066]: DEBUG nova.virt.hardware [None req-4fa8ac7b-ee00-4ef3-97aa-07a2c2ae6378 tempest-ServerExternalEventsTest-41922728 tempest-ServerExternalEventsTest-41922728-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 582.540372] env[62066]: DEBUG nova.virt.hardware [None req-4fa8ac7b-ee00-4ef3-97aa-07a2c2ae6378 tempest-ServerExternalEventsTest-41922728 tempest-ServerExternalEventsTest-41922728-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 582.540519] env[62066]: DEBUG nova.virt.hardware [None req-4fa8ac7b-ee00-4ef3-97aa-07a2c2ae6378 tempest-ServerExternalEventsTest-41922728 tempest-ServerExternalEventsTest-41922728-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 582.540721] env[62066]: DEBUG nova.virt.hardware [None req-4fa8ac7b-ee00-4ef3-97aa-07a2c2ae6378 tempest-ServerExternalEventsTest-41922728 tempest-ServerExternalEventsTest-41922728-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 582.540874] env[62066]: DEBUG nova.virt.hardware [None req-4fa8ac7b-ee00-4ef3-97aa-07a2c2ae6378 tempest-ServerExternalEventsTest-41922728 tempest-ServerExternalEventsTest-41922728-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 582.541048] env[62066]: DEBUG nova.virt.hardware [None req-4fa8ac7b-ee00-4ef3-97aa-07a2c2ae6378 tempest-ServerExternalEventsTest-41922728 tempest-ServerExternalEventsTest-41922728-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 582.541217] env[62066]: DEBUG nova.virt.hardware [None req-4fa8ac7b-ee00-4ef3-97aa-07a2c2ae6378 tempest-ServerExternalEventsTest-41922728 tempest-ServerExternalEventsTest-41922728-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 582.541743] env[62066]: DEBUG nova.virt.hardware [None req-4fa8ac7b-ee00-4ef3-97aa-07a2c2ae6378 tempest-ServerExternalEventsTest-41922728 tempest-ServerExternalEventsTest-41922728-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 582.542249] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce734661-711e-4079-a4a8-86896aff5d49 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.550094] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cab4e74-8f6b-472a-8582-8768627cb33d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.627268] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c4a0feae-1b12-4166-bd78-82a96fe8ac4c tempest-ServersWithSpecificFlavorTestJSON-184391817 tempest-ServersWithSpecificFlavorTestJSON-184391817-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.151s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 582.627894] env[62066]: ERROR nova.compute.manager [None req-c4a0feae-1b12-4166-bd78-82a96fe8ac4c tempest-ServersWithSpecificFlavorTestJSON-184391817 tempest-ServersWithSpecificFlavorTestJSON-184391817-project-member] [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 8f4adee9-38df-4ef0-88ca-4c8c443fe300, please check neutron logs for more information. [ 582.627894] env[62066]: ERROR nova.compute.manager [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] Traceback (most recent call last): [ 582.627894] env[62066]: ERROR nova.compute.manager [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 582.627894] env[62066]: ERROR nova.compute.manager [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] self.driver.spawn(context, instance, image_meta, [ 582.627894] env[62066]: ERROR nova.compute.manager [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 582.627894] env[62066]: ERROR nova.compute.manager [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] self._vmops.spawn(context, instance, image_meta, injected_files, [ 582.627894] env[62066]: ERROR nova.compute.manager [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 582.627894] env[62066]: ERROR nova.compute.manager [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] vm_ref = self.build_virtual_machine(instance, [ 582.627894] env[62066]: ERROR nova.compute.manager [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 582.627894] env[62066]: ERROR nova.compute.manager [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] vif_infos = vmwarevif.get_vif_info(self._session, [ 582.627894] env[62066]: ERROR nova.compute.manager [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 582.628221] env[62066]: ERROR nova.compute.manager [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] for vif in network_info: [ 582.628221] env[62066]: ERROR nova.compute.manager [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 582.628221] env[62066]: ERROR nova.compute.manager [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] return self._sync_wrapper(fn, *args, **kwargs) [ 582.628221] env[62066]: ERROR nova.compute.manager [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 582.628221] env[62066]: ERROR nova.compute.manager [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] self.wait() [ 582.628221] env[62066]: ERROR nova.compute.manager [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 582.628221] env[62066]: ERROR nova.compute.manager [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] self[:] = self._gt.wait() [ 582.628221] env[62066]: ERROR nova.compute.manager [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 582.628221] env[62066]: ERROR nova.compute.manager [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] return self._exit_event.wait() [ 582.628221] env[62066]: ERROR nova.compute.manager [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 582.628221] env[62066]: ERROR nova.compute.manager [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] result = hub.switch() [ 582.628221] env[62066]: ERROR nova.compute.manager [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 582.628221] env[62066]: ERROR nova.compute.manager [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] return self.greenlet.switch() [ 582.628487] env[62066]: ERROR nova.compute.manager [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 582.628487] env[62066]: ERROR nova.compute.manager [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] result = function(*args, **kwargs) [ 582.628487] env[62066]: ERROR nova.compute.manager [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 582.628487] env[62066]: ERROR nova.compute.manager [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] return func(*args, **kwargs) [ 582.628487] env[62066]: ERROR nova.compute.manager [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 582.628487] env[62066]: ERROR nova.compute.manager [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] raise e [ 582.628487] env[62066]: ERROR nova.compute.manager [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 582.628487] env[62066]: ERROR nova.compute.manager [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] nwinfo = self.network_api.allocate_for_instance( [ 582.628487] env[62066]: ERROR nova.compute.manager [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 582.628487] env[62066]: ERROR nova.compute.manager [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] created_port_ids = self._update_ports_for_instance( [ 582.628487] env[62066]: ERROR nova.compute.manager [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 582.628487] env[62066]: ERROR nova.compute.manager [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] with excutils.save_and_reraise_exception(): [ 582.628487] env[62066]: ERROR nova.compute.manager [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 582.628750] env[62066]: ERROR nova.compute.manager [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] self.force_reraise() [ 582.628750] env[62066]: ERROR nova.compute.manager [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 582.628750] env[62066]: ERROR nova.compute.manager [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] raise self.value [ 582.628750] env[62066]: ERROR nova.compute.manager [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 582.628750] env[62066]: ERROR nova.compute.manager [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] updated_port = self._update_port( [ 582.628750] env[62066]: ERROR nova.compute.manager [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 582.628750] env[62066]: ERROR nova.compute.manager [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] _ensure_no_port_binding_failure(port) [ 582.628750] env[62066]: ERROR nova.compute.manager [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 582.628750] env[62066]: ERROR nova.compute.manager [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] raise exception.PortBindingFailed(port_id=port['id']) [ 582.628750] env[62066]: ERROR nova.compute.manager [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] nova.exception.PortBindingFailed: Binding failed for port 8f4adee9-38df-4ef0-88ca-4c8c443fe300, please check neutron logs for more information. [ 582.628750] env[62066]: ERROR nova.compute.manager [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] [ 582.628977] env[62066]: DEBUG nova.compute.utils [None req-c4a0feae-1b12-4166-bd78-82a96fe8ac4c tempest-ServersWithSpecificFlavorTestJSON-184391817 tempest-ServersWithSpecificFlavorTestJSON-184391817-project-member] [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] Binding failed for port 8f4adee9-38df-4ef0-88ca-4c8c443fe300, please check neutron logs for more information. {{(pid=62066) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 582.629863] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1e3e30d6-8bb5-4347-b6c2-c2d648c37b67 tempest-ImagesOneServerNegativeTestJSON-310236551 tempest-ImagesOneServerNegativeTestJSON-310236551-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.602s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 582.631305] env[62066]: INFO nova.compute.claims [None req-1e3e30d6-8bb5-4347-b6c2-c2d648c37b67 tempest-ImagesOneServerNegativeTestJSON-310236551 tempest-ImagesOneServerNegativeTestJSON-310236551-project-member] [instance: 194588b4-fe40-4286-8036-874a7c410327] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 582.633819] env[62066]: DEBUG nova.compute.manager [None req-c4a0feae-1b12-4166-bd78-82a96fe8ac4c tempest-ServersWithSpecificFlavorTestJSON-184391817 tempest-ServersWithSpecificFlavorTestJSON-184391817-project-member] [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] Build of instance 4c67e82d-b51b-4820-af81-e6eefc7bcb36 was re-scheduled: Binding failed for port 8f4adee9-38df-4ef0-88ca-4c8c443fe300, please check neutron logs for more information. {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 582.634291] env[62066]: DEBUG nova.compute.manager [None req-c4a0feae-1b12-4166-bd78-82a96fe8ac4c tempest-ServersWithSpecificFlavorTestJSON-184391817 tempest-ServersWithSpecificFlavorTestJSON-184391817-project-member] [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] Unplugging VIFs for instance {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 582.634529] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c4a0feae-1b12-4166-bd78-82a96fe8ac4c tempest-ServersWithSpecificFlavorTestJSON-184391817 tempest-ServersWithSpecificFlavorTestJSON-184391817-project-member] Acquiring lock "refresh_cache-4c67e82d-b51b-4820-af81-e6eefc7bcb36" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 582.634678] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c4a0feae-1b12-4166-bd78-82a96fe8ac4c tempest-ServersWithSpecificFlavorTestJSON-184391817 tempest-ServersWithSpecificFlavorTestJSON-184391817-project-member] Acquired lock "refresh_cache-4c67e82d-b51b-4820-af81-e6eefc7bcb36" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 582.634844] env[62066]: DEBUG nova.network.neutron [None req-c4a0feae-1b12-4166-bd78-82a96fe8ac4c tempest-ServersWithSpecificFlavorTestJSON-184391817 tempest-ServersWithSpecificFlavorTestJSON-184391817-project-member] [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 582.720511] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 583.159134] env[62066]: DEBUG nova.network.neutron [None req-c4a0feae-1b12-4166-bd78-82a96fe8ac4c tempest-ServersWithSpecificFlavorTestJSON-184391817 tempest-ServersWithSpecificFlavorTestJSON-184391817-project-member] [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 583.239736] env[62066]: DEBUG nova.network.neutron [None req-c4a0feae-1b12-4166-bd78-82a96fe8ac4c tempest-ServersWithSpecificFlavorTestJSON-184391817 tempest-ServersWithSpecificFlavorTestJSON-184391817-project-member] [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 583.276451] env[62066]: ERROR nova.compute.manager [None req-4fa8ac7b-ee00-4ef3-97aa-07a2c2ae6378 tempest-ServerExternalEventsTest-41922728 tempest-ServerExternalEventsTest-41922728-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 00047463-06a3-4ffe-830b-2d4ef9e64de6, please check neutron logs for more information. [ 583.276451] env[62066]: ERROR nova.compute.manager Traceback (most recent call last): [ 583.276451] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 583.276451] env[62066]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 583.276451] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 583.276451] env[62066]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 583.276451] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 583.276451] env[62066]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 583.276451] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 583.276451] env[62066]: ERROR nova.compute.manager self.force_reraise() [ 583.276451] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 583.276451] env[62066]: ERROR nova.compute.manager raise self.value [ 583.276451] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 583.276451] env[62066]: ERROR nova.compute.manager updated_port = self._update_port( [ 583.276451] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 583.276451] env[62066]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 583.276940] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 583.276940] env[62066]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 583.276940] env[62066]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 00047463-06a3-4ffe-830b-2d4ef9e64de6, please check neutron logs for more information. [ 583.276940] env[62066]: ERROR nova.compute.manager [ 583.276940] env[62066]: Traceback (most recent call last): [ 583.276940] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 583.276940] env[62066]: listener.cb(fileno) [ 583.276940] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 583.276940] env[62066]: result = function(*args, **kwargs) [ 583.276940] env[62066]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 583.276940] env[62066]: return func(*args, **kwargs) [ 583.276940] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 583.276940] env[62066]: raise e [ 583.276940] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 583.276940] env[62066]: nwinfo = self.network_api.allocate_for_instance( [ 583.276940] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 583.276940] env[62066]: created_port_ids = self._update_ports_for_instance( [ 583.276940] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 583.276940] env[62066]: with excutils.save_and_reraise_exception(): [ 583.276940] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 583.276940] env[62066]: self.force_reraise() [ 583.276940] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 583.276940] env[62066]: raise self.value [ 583.276940] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 583.276940] env[62066]: updated_port = self._update_port( [ 583.276940] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 583.276940] env[62066]: _ensure_no_port_binding_failure(port) [ 583.276940] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 583.276940] env[62066]: raise exception.PortBindingFailed(port_id=port['id']) [ 583.277700] env[62066]: nova.exception.PortBindingFailed: Binding failed for port 00047463-06a3-4ffe-830b-2d4ef9e64de6, please check neutron logs for more information. [ 583.277700] env[62066]: Removing descriptor: 20 [ 583.277700] env[62066]: ERROR nova.compute.manager [None req-4fa8ac7b-ee00-4ef3-97aa-07a2c2ae6378 tempest-ServerExternalEventsTest-41922728 tempest-ServerExternalEventsTest-41922728-project-member] [instance: 39cd204a-49c2-4309-9030-555056ab125e] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 00047463-06a3-4ffe-830b-2d4ef9e64de6, please check neutron logs for more information. [ 583.277700] env[62066]: ERROR nova.compute.manager [instance: 39cd204a-49c2-4309-9030-555056ab125e] Traceback (most recent call last): [ 583.277700] env[62066]: ERROR nova.compute.manager [instance: 39cd204a-49c2-4309-9030-555056ab125e] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 583.277700] env[62066]: ERROR nova.compute.manager [instance: 39cd204a-49c2-4309-9030-555056ab125e] yield resources [ 583.277700] env[62066]: ERROR nova.compute.manager [instance: 39cd204a-49c2-4309-9030-555056ab125e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 583.277700] env[62066]: ERROR nova.compute.manager [instance: 39cd204a-49c2-4309-9030-555056ab125e] self.driver.spawn(context, instance, image_meta, [ 583.277700] env[62066]: ERROR nova.compute.manager [instance: 39cd204a-49c2-4309-9030-555056ab125e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 583.277700] env[62066]: ERROR nova.compute.manager [instance: 39cd204a-49c2-4309-9030-555056ab125e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 583.277700] env[62066]: ERROR nova.compute.manager [instance: 39cd204a-49c2-4309-9030-555056ab125e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 583.277700] env[62066]: ERROR nova.compute.manager [instance: 39cd204a-49c2-4309-9030-555056ab125e] vm_ref = self.build_virtual_machine(instance, [ 583.278045] env[62066]: ERROR nova.compute.manager [instance: 39cd204a-49c2-4309-9030-555056ab125e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 583.278045] env[62066]: ERROR nova.compute.manager [instance: 39cd204a-49c2-4309-9030-555056ab125e] vif_infos = vmwarevif.get_vif_info(self._session, [ 583.278045] env[62066]: ERROR nova.compute.manager [instance: 39cd204a-49c2-4309-9030-555056ab125e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 583.278045] env[62066]: ERROR nova.compute.manager [instance: 39cd204a-49c2-4309-9030-555056ab125e] for vif in network_info: [ 583.278045] env[62066]: ERROR nova.compute.manager [instance: 39cd204a-49c2-4309-9030-555056ab125e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 583.278045] env[62066]: ERROR nova.compute.manager [instance: 39cd204a-49c2-4309-9030-555056ab125e] return self._sync_wrapper(fn, *args, **kwargs) [ 583.278045] env[62066]: ERROR nova.compute.manager [instance: 39cd204a-49c2-4309-9030-555056ab125e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 583.278045] env[62066]: ERROR nova.compute.manager [instance: 39cd204a-49c2-4309-9030-555056ab125e] self.wait() [ 583.278045] env[62066]: ERROR nova.compute.manager [instance: 39cd204a-49c2-4309-9030-555056ab125e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 583.278045] env[62066]: ERROR nova.compute.manager [instance: 39cd204a-49c2-4309-9030-555056ab125e] self[:] = self._gt.wait() [ 583.278045] env[62066]: ERROR nova.compute.manager [instance: 39cd204a-49c2-4309-9030-555056ab125e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 583.278045] env[62066]: ERROR nova.compute.manager [instance: 39cd204a-49c2-4309-9030-555056ab125e] return self._exit_event.wait() [ 583.278045] env[62066]: ERROR nova.compute.manager [instance: 39cd204a-49c2-4309-9030-555056ab125e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 583.278389] env[62066]: ERROR nova.compute.manager [instance: 39cd204a-49c2-4309-9030-555056ab125e] result = hub.switch() [ 583.278389] env[62066]: ERROR nova.compute.manager [instance: 39cd204a-49c2-4309-9030-555056ab125e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 583.278389] env[62066]: ERROR nova.compute.manager [instance: 39cd204a-49c2-4309-9030-555056ab125e] return self.greenlet.switch() [ 583.278389] env[62066]: ERROR nova.compute.manager [instance: 39cd204a-49c2-4309-9030-555056ab125e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 583.278389] env[62066]: ERROR nova.compute.manager [instance: 39cd204a-49c2-4309-9030-555056ab125e] result = function(*args, **kwargs) [ 583.278389] env[62066]: ERROR nova.compute.manager [instance: 39cd204a-49c2-4309-9030-555056ab125e] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 583.278389] env[62066]: ERROR nova.compute.manager [instance: 39cd204a-49c2-4309-9030-555056ab125e] return func(*args, **kwargs) [ 583.278389] env[62066]: ERROR nova.compute.manager [instance: 39cd204a-49c2-4309-9030-555056ab125e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 583.278389] env[62066]: ERROR nova.compute.manager [instance: 39cd204a-49c2-4309-9030-555056ab125e] raise e [ 583.278389] env[62066]: ERROR nova.compute.manager [instance: 39cd204a-49c2-4309-9030-555056ab125e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 583.278389] env[62066]: ERROR nova.compute.manager [instance: 39cd204a-49c2-4309-9030-555056ab125e] nwinfo = self.network_api.allocate_for_instance( [ 583.278389] env[62066]: ERROR nova.compute.manager [instance: 39cd204a-49c2-4309-9030-555056ab125e] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 583.278389] env[62066]: ERROR nova.compute.manager [instance: 39cd204a-49c2-4309-9030-555056ab125e] created_port_ids = self._update_ports_for_instance( [ 583.278715] env[62066]: ERROR nova.compute.manager [instance: 39cd204a-49c2-4309-9030-555056ab125e] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 583.278715] env[62066]: ERROR nova.compute.manager [instance: 39cd204a-49c2-4309-9030-555056ab125e] with excutils.save_and_reraise_exception(): [ 583.278715] env[62066]: ERROR nova.compute.manager [instance: 39cd204a-49c2-4309-9030-555056ab125e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 583.278715] env[62066]: ERROR nova.compute.manager [instance: 39cd204a-49c2-4309-9030-555056ab125e] self.force_reraise() [ 583.278715] env[62066]: ERROR nova.compute.manager [instance: 39cd204a-49c2-4309-9030-555056ab125e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 583.278715] env[62066]: ERROR nova.compute.manager [instance: 39cd204a-49c2-4309-9030-555056ab125e] raise self.value [ 583.278715] env[62066]: ERROR nova.compute.manager [instance: 39cd204a-49c2-4309-9030-555056ab125e] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 583.278715] env[62066]: ERROR nova.compute.manager [instance: 39cd204a-49c2-4309-9030-555056ab125e] updated_port = self._update_port( [ 583.278715] env[62066]: ERROR nova.compute.manager [instance: 39cd204a-49c2-4309-9030-555056ab125e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 583.278715] env[62066]: ERROR nova.compute.manager [instance: 39cd204a-49c2-4309-9030-555056ab125e] _ensure_no_port_binding_failure(port) [ 583.278715] env[62066]: ERROR nova.compute.manager [instance: 39cd204a-49c2-4309-9030-555056ab125e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 583.278715] env[62066]: ERROR nova.compute.manager [instance: 39cd204a-49c2-4309-9030-555056ab125e] raise exception.PortBindingFailed(port_id=port['id']) [ 583.279026] env[62066]: ERROR nova.compute.manager [instance: 39cd204a-49c2-4309-9030-555056ab125e] nova.exception.PortBindingFailed: Binding failed for port 00047463-06a3-4ffe-830b-2d4ef9e64de6, please check neutron logs for more information. [ 583.279026] env[62066]: ERROR nova.compute.manager [instance: 39cd204a-49c2-4309-9030-555056ab125e] [ 583.279026] env[62066]: INFO nova.compute.manager [None req-4fa8ac7b-ee00-4ef3-97aa-07a2c2ae6378 tempest-ServerExternalEventsTest-41922728 tempest-ServerExternalEventsTest-41922728-project-member] [instance: 39cd204a-49c2-4309-9030-555056ab125e] Terminating instance [ 583.279682] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4fa8ac7b-ee00-4ef3-97aa-07a2c2ae6378 tempest-ServerExternalEventsTest-41922728 tempest-ServerExternalEventsTest-41922728-project-member] Acquiring lock "refresh_cache-39cd204a-49c2-4309-9030-555056ab125e" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 583.279847] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4fa8ac7b-ee00-4ef3-97aa-07a2c2ae6378 tempest-ServerExternalEventsTest-41922728 tempest-ServerExternalEventsTest-41922728-project-member] Acquired lock "refresh_cache-39cd204a-49c2-4309-9030-555056ab125e" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 583.280457] env[62066]: DEBUG nova.network.neutron [None req-4fa8ac7b-ee00-4ef3-97aa-07a2c2ae6378 tempest-ServerExternalEventsTest-41922728 tempest-ServerExternalEventsTest-41922728-project-member] [instance: 39cd204a-49c2-4309-9030-555056ab125e] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 583.432794] env[62066]: DEBUG nova.compute.manager [req-99507229-41ea-43f0-97ab-4a2df71c3e41 req-c7e4c379-5caf-4623-88ad-51b8b69e97c1 service nova] [instance: 39cd204a-49c2-4309-9030-555056ab125e] Received event network-changed-00047463-06a3-4ffe-830b-2d4ef9e64de6 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 583.433027] env[62066]: DEBUG nova.compute.manager [req-99507229-41ea-43f0-97ab-4a2df71c3e41 req-c7e4c379-5caf-4623-88ad-51b8b69e97c1 service nova] [instance: 39cd204a-49c2-4309-9030-555056ab125e] Refreshing instance network info cache due to event network-changed-00047463-06a3-4ffe-830b-2d4ef9e64de6. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 583.433248] env[62066]: DEBUG oslo_concurrency.lockutils [req-99507229-41ea-43f0-97ab-4a2df71c3e41 req-c7e4c379-5caf-4623-88ad-51b8b69e97c1 service nova] Acquiring lock "refresh_cache-39cd204a-49c2-4309-9030-555056ab125e" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 583.742559] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c4a0feae-1b12-4166-bd78-82a96fe8ac4c tempest-ServersWithSpecificFlavorTestJSON-184391817 tempest-ServersWithSpecificFlavorTestJSON-184391817-project-member] Releasing lock "refresh_cache-4c67e82d-b51b-4820-af81-e6eefc7bcb36" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 583.742781] env[62066]: DEBUG nova.compute.manager [None req-c4a0feae-1b12-4166-bd78-82a96fe8ac4c tempest-ServersWithSpecificFlavorTestJSON-184391817 tempest-ServersWithSpecificFlavorTestJSON-184391817-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 583.742960] env[62066]: DEBUG nova.compute.manager [None req-c4a0feae-1b12-4166-bd78-82a96fe8ac4c tempest-ServersWithSpecificFlavorTestJSON-184391817 tempest-ServersWithSpecificFlavorTestJSON-184391817-project-member] [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 583.743144] env[62066]: DEBUG nova.network.neutron [None req-c4a0feae-1b12-4166-bd78-82a96fe8ac4c tempest-ServersWithSpecificFlavorTestJSON-184391817 tempest-ServersWithSpecificFlavorTestJSON-184391817-project-member] [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 583.759599] env[62066]: DEBUG nova.network.neutron [None req-c4a0feae-1b12-4166-bd78-82a96fe8ac4c tempest-ServersWithSpecificFlavorTestJSON-184391817 tempest-ServersWithSpecificFlavorTestJSON-184391817-project-member] [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 583.799289] env[62066]: DEBUG nova.network.neutron [None req-4fa8ac7b-ee00-4ef3-97aa-07a2c2ae6378 tempest-ServerExternalEventsTest-41922728 tempest-ServerExternalEventsTest-41922728-project-member] [instance: 39cd204a-49c2-4309-9030-555056ab125e] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 583.910106] env[62066]: DEBUG nova.network.neutron [None req-4fa8ac7b-ee00-4ef3-97aa-07a2c2ae6378 tempest-ServerExternalEventsTest-41922728 tempest-ServerExternalEventsTest-41922728-project-member] [instance: 39cd204a-49c2-4309-9030-555056ab125e] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 584.085925] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a13df006-212f-41b9-8726-7c372fc2d48f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.093886] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bfb0b1b-ee08-4ac0-a9b9-05ec0a44ac0e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.123826] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88ffcb10-38f7-4745-82d7-741565cf7f7d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.132142] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3408f7dd-7927-400d-8ba2-f43e350b5eb0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.146394] env[62066]: DEBUG nova.compute.provider_tree [None req-1e3e30d6-8bb5-4347-b6c2-c2d648c37b67 tempest-ImagesOneServerNegativeTestJSON-310236551 tempest-ImagesOneServerNegativeTestJSON-310236551-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 584.261832] env[62066]: DEBUG nova.network.neutron [None req-c4a0feae-1b12-4166-bd78-82a96fe8ac4c tempest-ServersWithSpecificFlavorTestJSON-184391817 tempest-ServersWithSpecificFlavorTestJSON-184391817-project-member] [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 584.415616] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4fa8ac7b-ee00-4ef3-97aa-07a2c2ae6378 tempest-ServerExternalEventsTest-41922728 tempest-ServerExternalEventsTest-41922728-project-member] Releasing lock "refresh_cache-39cd204a-49c2-4309-9030-555056ab125e" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 584.415616] env[62066]: DEBUG nova.compute.manager [None req-4fa8ac7b-ee00-4ef3-97aa-07a2c2ae6378 tempest-ServerExternalEventsTest-41922728 tempest-ServerExternalEventsTest-41922728-project-member] [instance: 39cd204a-49c2-4309-9030-555056ab125e] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 584.415616] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-4fa8ac7b-ee00-4ef3-97aa-07a2c2ae6378 tempest-ServerExternalEventsTest-41922728 tempest-ServerExternalEventsTest-41922728-project-member] [instance: 39cd204a-49c2-4309-9030-555056ab125e] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 584.415616] env[62066]: DEBUG oslo_concurrency.lockutils [req-99507229-41ea-43f0-97ab-4a2df71c3e41 req-c7e4c379-5caf-4623-88ad-51b8b69e97c1 service nova] Acquired lock "refresh_cache-39cd204a-49c2-4309-9030-555056ab125e" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 584.415779] env[62066]: DEBUG nova.network.neutron [req-99507229-41ea-43f0-97ab-4a2df71c3e41 req-c7e4c379-5caf-4623-88ad-51b8b69e97c1 service nova] [instance: 39cd204a-49c2-4309-9030-555056ab125e] Refreshing network info cache for port 00047463-06a3-4ffe-830b-2d4ef9e64de6 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 584.418037] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-926f484a-c07b-4163-b2fe-46aed6937b3f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.427071] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b8a6a6c-d019-4e75-958b-42c80c43cd12 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.450351] env[62066]: WARNING nova.virt.vmwareapi.vmops [None req-4fa8ac7b-ee00-4ef3-97aa-07a2c2ae6378 tempest-ServerExternalEventsTest-41922728 tempest-ServerExternalEventsTest-41922728-project-member] [instance: 39cd204a-49c2-4309-9030-555056ab125e] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 39cd204a-49c2-4309-9030-555056ab125e could not be found. [ 584.450680] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-4fa8ac7b-ee00-4ef3-97aa-07a2c2ae6378 tempest-ServerExternalEventsTest-41922728 tempest-ServerExternalEventsTest-41922728-project-member] [instance: 39cd204a-49c2-4309-9030-555056ab125e] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 584.450757] env[62066]: INFO nova.compute.manager [None req-4fa8ac7b-ee00-4ef3-97aa-07a2c2ae6378 tempest-ServerExternalEventsTest-41922728 tempest-ServerExternalEventsTest-41922728-project-member] [instance: 39cd204a-49c2-4309-9030-555056ab125e] Took 0.04 seconds to destroy the instance on the hypervisor. [ 584.450997] env[62066]: DEBUG oslo.service.loopingcall [None req-4fa8ac7b-ee00-4ef3-97aa-07a2c2ae6378 tempest-ServerExternalEventsTest-41922728 tempest-ServerExternalEventsTest-41922728-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 584.451216] env[62066]: DEBUG nova.compute.manager [-] [instance: 39cd204a-49c2-4309-9030-555056ab125e] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 584.451310] env[62066]: DEBUG nova.network.neutron [-] [instance: 39cd204a-49c2-4309-9030-555056ab125e] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 584.469464] env[62066]: DEBUG nova.network.neutron [-] [instance: 39cd204a-49c2-4309-9030-555056ab125e] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 584.650553] env[62066]: DEBUG nova.scheduler.client.report [None req-1e3e30d6-8bb5-4347-b6c2-c2d648c37b67 tempest-ImagesOneServerNegativeTestJSON-310236551 tempest-ImagesOneServerNegativeTestJSON-310236551-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 584.764674] env[62066]: INFO nova.compute.manager [None req-c4a0feae-1b12-4166-bd78-82a96fe8ac4c tempest-ServersWithSpecificFlavorTestJSON-184391817 tempest-ServersWithSpecificFlavorTestJSON-184391817-project-member] [instance: 4c67e82d-b51b-4820-af81-e6eefc7bcb36] Took 1.02 seconds to deallocate network for instance. [ 584.949205] env[62066]: DEBUG nova.network.neutron [req-99507229-41ea-43f0-97ab-4a2df71c3e41 req-c7e4c379-5caf-4623-88ad-51b8b69e97c1 service nova] [instance: 39cd204a-49c2-4309-9030-555056ab125e] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 584.972198] env[62066]: DEBUG nova.network.neutron [-] [instance: 39cd204a-49c2-4309-9030-555056ab125e] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 585.052956] env[62066]: DEBUG nova.network.neutron [req-99507229-41ea-43f0-97ab-4a2df71c3e41 req-c7e4c379-5caf-4623-88ad-51b8b69e97c1 service nova] [instance: 39cd204a-49c2-4309-9030-555056ab125e] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 585.156520] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1e3e30d6-8bb5-4347-b6c2-c2d648c37b67 tempest-ImagesOneServerNegativeTestJSON-310236551 tempest-ImagesOneServerNegativeTestJSON-310236551-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.527s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 585.157077] env[62066]: DEBUG nova.compute.manager [None req-1e3e30d6-8bb5-4347-b6c2-c2d648c37b67 tempest-ImagesOneServerNegativeTestJSON-310236551 tempest-ImagesOneServerNegativeTestJSON-310236551-project-member] [instance: 194588b4-fe40-4286-8036-874a7c410327] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 585.160975] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cb29a201-afba-4eb8-8336-fe27a681c8f8 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 20.787s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 585.475809] env[62066]: INFO nova.compute.manager [-] [instance: 39cd204a-49c2-4309-9030-555056ab125e] Took 1.02 seconds to deallocate network for instance. [ 585.478658] env[62066]: DEBUG nova.compute.claims [None req-4fa8ac7b-ee00-4ef3-97aa-07a2c2ae6378 tempest-ServerExternalEventsTest-41922728 tempest-ServerExternalEventsTest-41922728-project-member] [instance: 39cd204a-49c2-4309-9030-555056ab125e] Aborting claim: {{(pid=62066) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 585.478825] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4fa8ac7b-ee00-4ef3-97aa-07a2c2ae6378 tempest-ServerExternalEventsTest-41922728 tempest-ServerExternalEventsTest-41922728-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 585.555930] env[62066]: DEBUG nova.compute.manager [req-8e05e825-fb61-4c66-a04e-25adb1001eee req-ce1ce3e8-ec00-4bcb-825b-3ae112a2265e service nova] [instance: 39cd204a-49c2-4309-9030-555056ab125e] Received event network-vif-deleted-00047463-06a3-4ffe-830b-2d4ef9e64de6 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 585.556394] env[62066]: DEBUG oslo_concurrency.lockutils [req-99507229-41ea-43f0-97ab-4a2df71c3e41 req-c7e4c379-5caf-4623-88ad-51b8b69e97c1 service nova] Releasing lock "refresh_cache-39cd204a-49c2-4309-9030-555056ab125e" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 585.673546] env[62066]: DEBUG nova.compute.utils [None req-1e3e30d6-8bb5-4347-b6c2-c2d648c37b67 tempest-ImagesOneServerNegativeTestJSON-310236551 tempest-ImagesOneServerNegativeTestJSON-310236551-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 585.680656] env[62066]: DEBUG nova.compute.manager [None req-1e3e30d6-8bb5-4347-b6c2-c2d648c37b67 tempest-ImagesOneServerNegativeTestJSON-310236551 tempest-ImagesOneServerNegativeTestJSON-310236551-project-member] [instance: 194588b4-fe40-4286-8036-874a7c410327] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 585.681511] env[62066]: DEBUG nova.network.neutron [None req-1e3e30d6-8bb5-4347-b6c2-c2d648c37b67 tempest-ImagesOneServerNegativeTestJSON-310236551 tempest-ImagesOneServerNegativeTestJSON-310236551-project-member] [instance: 194588b4-fe40-4286-8036-874a7c410327] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 585.779349] env[62066]: DEBUG nova.policy [None req-1e3e30d6-8bb5-4347-b6c2-c2d648c37b67 tempest-ImagesOneServerNegativeTestJSON-310236551 tempest-ImagesOneServerNegativeTestJSON-310236551-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'dc35da8e28434a819636d16d4d6fb950', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a0d462917eb04c30ae8f4236464b4344', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 585.799077] env[62066]: INFO nova.scheduler.client.report [None req-c4a0feae-1b12-4166-bd78-82a96fe8ac4c tempest-ServersWithSpecificFlavorTestJSON-184391817 tempest-ServersWithSpecificFlavorTestJSON-184391817-project-member] Deleted allocations for instance 4c67e82d-b51b-4820-af81-e6eefc7bcb36 [ 586.167866] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b460d61-873b-4563-be86-5ab470d8ef1c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.175737] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f833df4-33df-413c-b873-c304485992b0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.183408] env[62066]: DEBUG nova.compute.manager [None req-1e3e30d6-8bb5-4347-b6c2-c2d648c37b67 tempest-ImagesOneServerNegativeTestJSON-310236551 tempest-ImagesOneServerNegativeTestJSON-310236551-project-member] [instance: 194588b4-fe40-4286-8036-874a7c410327] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 586.209944] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1f46139-cae6-4441-b889-a5b7c0f47339 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.217823] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7dc99551-9347-43d0-b439-08a23a2345e1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.234111] env[62066]: DEBUG nova.compute.provider_tree [None req-cb29a201-afba-4eb8-8336-fe27a681c8f8 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 586.309476] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c4a0feae-1b12-4166-bd78-82a96fe8ac4c tempest-ServersWithSpecificFlavorTestJSON-184391817 tempest-ServersWithSpecificFlavorTestJSON-184391817-project-member] Lock "4c67e82d-b51b-4820-af81-e6eefc7bcb36" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 78.263s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 586.619303] env[62066]: DEBUG nova.network.neutron [None req-1e3e30d6-8bb5-4347-b6c2-c2d648c37b67 tempest-ImagesOneServerNegativeTestJSON-310236551 tempest-ImagesOneServerNegativeTestJSON-310236551-project-member] [instance: 194588b4-fe40-4286-8036-874a7c410327] Successfully created port: 2a2c9bbf-b369-4ea3-a866-79152380ed0b {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 586.735588] env[62066]: DEBUG nova.scheduler.client.report [None req-cb29a201-afba-4eb8-8336-fe27a681c8f8 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 586.813953] env[62066]: DEBUG nova.compute.manager [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 587.220023] env[62066]: DEBUG nova.compute.manager [None req-1e3e30d6-8bb5-4347-b6c2-c2d648c37b67 tempest-ImagesOneServerNegativeTestJSON-310236551 tempest-ImagesOneServerNegativeTestJSON-310236551-project-member] [instance: 194588b4-fe40-4286-8036-874a7c410327] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 587.242342] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cb29a201-afba-4eb8-8336-fe27a681c8f8 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.081s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 587.243655] env[62066]: ERROR nova.compute.manager [None req-cb29a201-afba-4eb8-8336-fe27a681c8f8 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port c18dc899-1b90-41a9-8d63-1bf7f9d1a206, please check neutron logs for more information. [ 587.243655] env[62066]: ERROR nova.compute.manager [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] Traceback (most recent call last): [ 587.243655] env[62066]: ERROR nova.compute.manager [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 587.243655] env[62066]: ERROR nova.compute.manager [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] self.driver.spawn(context, instance, image_meta, [ 587.243655] env[62066]: ERROR nova.compute.manager [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 587.243655] env[62066]: ERROR nova.compute.manager [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] self._vmops.spawn(context, instance, image_meta, injected_files, [ 587.243655] env[62066]: ERROR nova.compute.manager [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 587.243655] env[62066]: ERROR nova.compute.manager [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] vm_ref = self.build_virtual_machine(instance, [ 587.243655] env[62066]: ERROR nova.compute.manager [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 587.243655] env[62066]: ERROR nova.compute.manager [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] vif_infos = vmwarevif.get_vif_info(self._session, [ 587.243655] env[62066]: ERROR nova.compute.manager [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 587.244172] env[62066]: ERROR nova.compute.manager [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] for vif in network_info: [ 587.244172] env[62066]: ERROR nova.compute.manager [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 587.244172] env[62066]: ERROR nova.compute.manager [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] return self._sync_wrapper(fn, *args, **kwargs) [ 587.244172] env[62066]: ERROR nova.compute.manager [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 587.244172] env[62066]: ERROR nova.compute.manager [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] self.wait() [ 587.244172] env[62066]: ERROR nova.compute.manager [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 587.244172] env[62066]: ERROR nova.compute.manager [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] self[:] = self._gt.wait() [ 587.244172] env[62066]: ERROR nova.compute.manager [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 587.244172] env[62066]: ERROR nova.compute.manager [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] return self._exit_event.wait() [ 587.244172] env[62066]: ERROR nova.compute.manager [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 587.244172] env[62066]: ERROR nova.compute.manager [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] result = hub.switch() [ 587.244172] env[62066]: ERROR nova.compute.manager [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 587.244172] env[62066]: ERROR nova.compute.manager [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] return self.greenlet.switch() [ 587.244456] env[62066]: ERROR nova.compute.manager [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 587.244456] env[62066]: ERROR nova.compute.manager [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] result = function(*args, **kwargs) [ 587.244456] env[62066]: ERROR nova.compute.manager [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 587.244456] env[62066]: ERROR nova.compute.manager [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] return func(*args, **kwargs) [ 587.244456] env[62066]: ERROR nova.compute.manager [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 587.244456] env[62066]: ERROR nova.compute.manager [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] raise e [ 587.244456] env[62066]: ERROR nova.compute.manager [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 587.244456] env[62066]: ERROR nova.compute.manager [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] nwinfo = self.network_api.allocate_for_instance( [ 587.244456] env[62066]: ERROR nova.compute.manager [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 587.244456] env[62066]: ERROR nova.compute.manager [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] created_port_ids = self._update_ports_for_instance( [ 587.244456] env[62066]: ERROR nova.compute.manager [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 587.244456] env[62066]: ERROR nova.compute.manager [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] with excutils.save_and_reraise_exception(): [ 587.244456] env[62066]: ERROR nova.compute.manager [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 587.244735] env[62066]: ERROR nova.compute.manager [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] self.force_reraise() [ 587.244735] env[62066]: ERROR nova.compute.manager [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 587.244735] env[62066]: ERROR nova.compute.manager [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] raise self.value [ 587.244735] env[62066]: ERROR nova.compute.manager [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 587.244735] env[62066]: ERROR nova.compute.manager [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] updated_port = self._update_port( [ 587.244735] env[62066]: ERROR nova.compute.manager [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 587.244735] env[62066]: ERROR nova.compute.manager [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] _ensure_no_port_binding_failure(port) [ 587.244735] env[62066]: ERROR nova.compute.manager [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 587.244735] env[62066]: ERROR nova.compute.manager [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] raise exception.PortBindingFailed(port_id=port['id']) [ 587.244735] env[62066]: ERROR nova.compute.manager [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] nova.exception.PortBindingFailed: Binding failed for port c18dc899-1b90-41a9-8d63-1bf7f9d1a206, please check neutron logs for more information. [ 587.244735] env[62066]: ERROR nova.compute.manager [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] [ 587.244980] env[62066]: DEBUG nova.compute.utils [None req-cb29a201-afba-4eb8-8336-fe27a681c8f8 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] Binding failed for port c18dc899-1b90-41a9-8d63-1bf7f9d1a206, please check neutron logs for more information. {{(pid=62066) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 587.247127] env[62066]: DEBUG oslo_concurrency.lockutils [None req-fc3cf2dd-2491-4541-a245-5eefc9a5c042 tempest-TenantUsagesTestJSON-970395509 tempest-TenantUsagesTestJSON-970395509-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.556s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 587.249431] env[62066]: INFO nova.compute.claims [None req-fc3cf2dd-2491-4541-a245-5eefc9a5c042 tempest-TenantUsagesTestJSON-970395509 tempest-TenantUsagesTestJSON-970395509-project-member] [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 587.253362] env[62066]: DEBUG nova.compute.manager [None req-cb29a201-afba-4eb8-8336-fe27a681c8f8 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] Build of instance 69d99063-a6a7-4832-a864-dc9fd6024ea8 was re-scheduled: Binding failed for port c18dc899-1b90-41a9-8d63-1bf7f9d1a206, please check neutron logs for more information. {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 587.253856] env[62066]: DEBUG nova.compute.manager [None req-cb29a201-afba-4eb8-8336-fe27a681c8f8 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] Unplugging VIFs for instance {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 587.254135] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cb29a201-afba-4eb8-8336-fe27a681c8f8 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] Acquiring lock "refresh_cache-69d99063-a6a7-4832-a864-dc9fd6024ea8" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 587.254757] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cb29a201-afba-4eb8-8336-fe27a681c8f8 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] Acquired lock "refresh_cache-69d99063-a6a7-4832-a864-dc9fd6024ea8" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 587.255335] env[62066]: DEBUG nova.network.neutron [None req-cb29a201-afba-4eb8-8336-fe27a681c8f8 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 587.261914] env[62066]: DEBUG nova.virt.hardware [None req-1e3e30d6-8bb5-4347-b6c2-c2d648c37b67 tempest-ImagesOneServerNegativeTestJSON-310236551 tempest-ImagesOneServerNegativeTestJSON-310236551-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-23T13:40:41Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-23T13:40:23Z,direct_url=,disk_format='vmdk',id=50ff584c-3b50-4395-af07-3e66769bc9f7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='eb52f7069a374c61ae946f052007c6d9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-23T13:40:24Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 587.263033] env[62066]: DEBUG nova.virt.hardware [None req-1e3e30d6-8bb5-4347-b6c2-c2d648c37b67 tempest-ImagesOneServerNegativeTestJSON-310236551 tempest-ImagesOneServerNegativeTestJSON-310236551-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 587.263033] env[62066]: DEBUG nova.virt.hardware [None req-1e3e30d6-8bb5-4347-b6c2-c2d648c37b67 tempest-ImagesOneServerNegativeTestJSON-310236551 tempest-ImagesOneServerNegativeTestJSON-310236551-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 587.263033] env[62066]: DEBUG nova.virt.hardware [None req-1e3e30d6-8bb5-4347-b6c2-c2d648c37b67 tempest-ImagesOneServerNegativeTestJSON-310236551 tempest-ImagesOneServerNegativeTestJSON-310236551-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 587.263033] env[62066]: DEBUG nova.virt.hardware [None req-1e3e30d6-8bb5-4347-b6c2-c2d648c37b67 tempest-ImagesOneServerNegativeTestJSON-310236551 tempest-ImagesOneServerNegativeTestJSON-310236551-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 587.263033] env[62066]: DEBUG nova.virt.hardware [None req-1e3e30d6-8bb5-4347-b6c2-c2d648c37b67 tempest-ImagesOneServerNegativeTestJSON-310236551 tempest-ImagesOneServerNegativeTestJSON-310236551-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 587.263241] env[62066]: DEBUG nova.virt.hardware [None req-1e3e30d6-8bb5-4347-b6c2-c2d648c37b67 tempest-ImagesOneServerNegativeTestJSON-310236551 tempest-ImagesOneServerNegativeTestJSON-310236551-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 587.263403] env[62066]: DEBUG nova.virt.hardware [None req-1e3e30d6-8bb5-4347-b6c2-c2d648c37b67 tempest-ImagesOneServerNegativeTestJSON-310236551 tempest-ImagesOneServerNegativeTestJSON-310236551-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 587.263566] env[62066]: DEBUG nova.virt.hardware [None req-1e3e30d6-8bb5-4347-b6c2-c2d648c37b67 tempest-ImagesOneServerNegativeTestJSON-310236551 tempest-ImagesOneServerNegativeTestJSON-310236551-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 587.263727] env[62066]: DEBUG nova.virt.hardware [None req-1e3e30d6-8bb5-4347-b6c2-c2d648c37b67 tempest-ImagesOneServerNegativeTestJSON-310236551 tempest-ImagesOneServerNegativeTestJSON-310236551-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 587.263895] env[62066]: DEBUG nova.virt.hardware [None req-1e3e30d6-8bb5-4347-b6c2-c2d648c37b67 tempest-ImagesOneServerNegativeTestJSON-310236551 tempest-ImagesOneServerNegativeTestJSON-310236551-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 587.264855] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98ac9673-b7c4-4eaf-9a9f-11fbccdecc6a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.278769] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4a85e71-9e27-4b36-9ab4-e6f4fca5287f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.353628] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 587.794509] env[62066]: DEBUG nova.network.neutron [None req-cb29a201-afba-4eb8-8336-fe27a681c8f8 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 588.009780] env[62066]: DEBUG nova.network.neutron [None req-cb29a201-afba-4eb8-8336-fe27a681c8f8 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 588.512954] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cb29a201-afba-4eb8-8336-fe27a681c8f8 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] Releasing lock "refresh_cache-69d99063-a6a7-4832-a864-dc9fd6024ea8" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 588.512954] env[62066]: DEBUG nova.compute.manager [None req-cb29a201-afba-4eb8-8336-fe27a681c8f8 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 588.512954] env[62066]: DEBUG nova.compute.manager [None req-cb29a201-afba-4eb8-8336-fe27a681c8f8 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 588.513132] env[62066]: DEBUG nova.network.neutron [None req-cb29a201-afba-4eb8-8336-fe27a681c8f8 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 588.543483] env[62066]: DEBUG nova.network.neutron [None req-cb29a201-afba-4eb8-8336-fe27a681c8f8 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 588.573922] env[62066]: DEBUG nova.compute.manager [req-c694cdae-aa25-48e0-8de8-bfe8bd0661e4 req-cb31e0e9-a33c-464b-9920-a34dfae482ca service nova] [instance: 194588b4-fe40-4286-8036-874a7c410327] Received event network-changed-2a2c9bbf-b369-4ea3-a866-79152380ed0b {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 588.573922] env[62066]: DEBUG nova.compute.manager [req-c694cdae-aa25-48e0-8de8-bfe8bd0661e4 req-cb31e0e9-a33c-464b-9920-a34dfae482ca service nova] [instance: 194588b4-fe40-4286-8036-874a7c410327] Refreshing instance network info cache due to event network-changed-2a2c9bbf-b369-4ea3-a866-79152380ed0b. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 588.574059] env[62066]: DEBUG oslo_concurrency.lockutils [req-c694cdae-aa25-48e0-8de8-bfe8bd0661e4 req-cb31e0e9-a33c-464b-9920-a34dfae482ca service nova] Acquiring lock "refresh_cache-194588b4-fe40-4286-8036-874a7c410327" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 588.574117] env[62066]: DEBUG oslo_concurrency.lockutils [req-c694cdae-aa25-48e0-8de8-bfe8bd0661e4 req-cb31e0e9-a33c-464b-9920-a34dfae482ca service nova] Acquired lock "refresh_cache-194588b4-fe40-4286-8036-874a7c410327" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 588.574275] env[62066]: DEBUG nova.network.neutron [req-c694cdae-aa25-48e0-8de8-bfe8bd0661e4 req-cb31e0e9-a33c-464b-9920-a34dfae482ca service nova] [instance: 194588b4-fe40-4286-8036-874a7c410327] Refreshing network info cache for port 2a2c9bbf-b369-4ea3-a866-79152380ed0b {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 588.760832] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09d3180b-9613-4d19-99bf-88387279e395 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.769317] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b30251e-10b6-461d-825e-f54b6ff90ef9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.808243] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92e548cd-1315-4992-b863-c2f57c5589b6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.815946] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aef5f1ea-029b-490a-9f09-d6b14a4af99a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.829923] env[62066]: DEBUG nova.compute.provider_tree [None req-fc3cf2dd-2491-4541-a245-5eefc9a5c042 tempest-TenantUsagesTestJSON-970395509 tempest-TenantUsagesTestJSON-970395509-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 589.047608] env[62066]: DEBUG nova.network.neutron [None req-cb29a201-afba-4eb8-8336-fe27a681c8f8 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 589.073857] env[62066]: ERROR nova.compute.manager [None req-1e3e30d6-8bb5-4347-b6c2-c2d648c37b67 tempest-ImagesOneServerNegativeTestJSON-310236551 tempest-ImagesOneServerNegativeTestJSON-310236551-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 2a2c9bbf-b369-4ea3-a866-79152380ed0b, please check neutron logs for more information. [ 589.073857] env[62066]: ERROR nova.compute.manager Traceback (most recent call last): [ 589.073857] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 589.073857] env[62066]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 589.073857] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 589.073857] env[62066]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 589.073857] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 589.073857] env[62066]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 589.073857] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 589.073857] env[62066]: ERROR nova.compute.manager self.force_reraise() [ 589.073857] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 589.073857] env[62066]: ERROR nova.compute.manager raise self.value [ 589.073857] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 589.073857] env[62066]: ERROR nova.compute.manager updated_port = self._update_port( [ 589.073857] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 589.073857] env[62066]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 589.074422] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 589.074422] env[62066]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 589.074422] env[62066]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 2a2c9bbf-b369-4ea3-a866-79152380ed0b, please check neutron logs for more information. [ 589.074422] env[62066]: ERROR nova.compute.manager [ 589.074422] env[62066]: Traceback (most recent call last): [ 589.074422] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 589.074422] env[62066]: listener.cb(fileno) [ 589.074422] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 589.074422] env[62066]: result = function(*args, **kwargs) [ 589.074422] env[62066]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 589.074422] env[62066]: return func(*args, **kwargs) [ 589.074422] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 589.074422] env[62066]: raise e [ 589.074422] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 589.074422] env[62066]: nwinfo = self.network_api.allocate_for_instance( [ 589.074422] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 589.074422] env[62066]: created_port_ids = self._update_ports_for_instance( [ 589.074422] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 589.074422] env[62066]: with excutils.save_and_reraise_exception(): [ 589.074422] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 589.074422] env[62066]: self.force_reraise() [ 589.074422] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 589.074422] env[62066]: raise self.value [ 589.074422] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 589.074422] env[62066]: updated_port = self._update_port( [ 589.074422] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 589.074422] env[62066]: _ensure_no_port_binding_failure(port) [ 589.074422] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 589.074422] env[62066]: raise exception.PortBindingFailed(port_id=port['id']) [ 589.075364] env[62066]: nova.exception.PortBindingFailed: Binding failed for port 2a2c9bbf-b369-4ea3-a866-79152380ed0b, please check neutron logs for more information. [ 589.075364] env[62066]: Removing descriptor: 20 [ 589.075364] env[62066]: ERROR nova.compute.manager [None req-1e3e30d6-8bb5-4347-b6c2-c2d648c37b67 tempest-ImagesOneServerNegativeTestJSON-310236551 tempest-ImagesOneServerNegativeTestJSON-310236551-project-member] [instance: 194588b4-fe40-4286-8036-874a7c410327] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 2a2c9bbf-b369-4ea3-a866-79152380ed0b, please check neutron logs for more information. [ 589.075364] env[62066]: ERROR nova.compute.manager [instance: 194588b4-fe40-4286-8036-874a7c410327] Traceback (most recent call last): [ 589.075364] env[62066]: ERROR nova.compute.manager [instance: 194588b4-fe40-4286-8036-874a7c410327] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 589.075364] env[62066]: ERROR nova.compute.manager [instance: 194588b4-fe40-4286-8036-874a7c410327] yield resources [ 589.075364] env[62066]: ERROR nova.compute.manager [instance: 194588b4-fe40-4286-8036-874a7c410327] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 589.075364] env[62066]: ERROR nova.compute.manager [instance: 194588b4-fe40-4286-8036-874a7c410327] self.driver.spawn(context, instance, image_meta, [ 589.075364] env[62066]: ERROR nova.compute.manager [instance: 194588b4-fe40-4286-8036-874a7c410327] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 589.075364] env[62066]: ERROR nova.compute.manager [instance: 194588b4-fe40-4286-8036-874a7c410327] self._vmops.spawn(context, instance, image_meta, injected_files, [ 589.075364] env[62066]: ERROR nova.compute.manager [instance: 194588b4-fe40-4286-8036-874a7c410327] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 589.075364] env[62066]: ERROR nova.compute.manager [instance: 194588b4-fe40-4286-8036-874a7c410327] vm_ref = self.build_virtual_machine(instance, [ 589.075691] env[62066]: ERROR nova.compute.manager [instance: 194588b4-fe40-4286-8036-874a7c410327] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 589.075691] env[62066]: ERROR nova.compute.manager [instance: 194588b4-fe40-4286-8036-874a7c410327] vif_infos = vmwarevif.get_vif_info(self._session, [ 589.075691] env[62066]: ERROR nova.compute.manager [instance: 194588b4-fe40-4286-8036-874a7c410327] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 589.075691] env[62066]: ERROR nova.compute.manager [instance: 194588b4-fe40-4286-8036-874a7c410327] for vif in network_info: [ 589.075691] env[62066]: ERROR nova.compute.manager [instance: 194588b4-fe40-4286-8036-874a7c410327] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 589.075691] env[62066]: ERROR nova.compute.manager [instance: 194588b4-fe40-4286-8036-874a7c410327] return self._sync_wrapper(fn, *args, **kwargs) [ 589.075691] env[62066]: ERROR nova.compute.manager [instance: 194588b4-fe40-4286-8036-874a7c410327] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 589.075691] env[62066]: ERROR nova.compute.manager [instance: 194588b4-fe40-4286-8036-874a7c410327] self.wait() [ 589.075691] env[62066]: ERROR nova.compute.manager [instance: 194588b4-fe40-4286-8036-874a7c410327] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 589.075691] env[62066]: ERROR nova.compute.manager [instance: 194588b4-fe40-4286-8036-874a7c410327] self[:] = self._gt.wait() [ 589.075691] env[62066]: ERROR nova.compute.manager [instance: 194588b4-fe40-4286-8036-874a7c410327] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 589.075691] env[62066]: ERROR nova.compute.manager [instance: 194588b4-fe40-4286-8036-874a7c410327] return self._exit_event.wait() [ 589.075691] env[62066]: ERROR nova.compute.manager [instance: 194588b4-fe40-4286-8036-874a7c410327] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 589.076095] env[62066]: ERROR nova.compute.manager [instance: 194588b4-fe40-4286-8036-874a7c410327] result = hub.switch() [ 589.076095] env[62066]: ERROR nova.compute.manager [instance: 194588b4-fe40-4286-8036-874a7c410327] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 589.076095] env[62066]: ERROR nova.compute.manager [instance: 194588b4-fe40-4286-8036-874a7c410327] return self.greenlet.switch() [ 589.076095] env[62066]: ERROR nova.compute.manager [instance: 194588b4-fe40-4286-8036-874a7c410327] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 589.076095] env[62066]: ERROR nova.compute.manager [instance: 194588b4-fe40-4286-8036-874a7c410327] result = function(*args, **kwargs) [ 589.076095] env[62066]: ERROR nova.compute.manager [instance: 194588b4-fe40-4286-8036-874a7c410327] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 589.076095] env[62066]: ERROR nova.compute.manager [instance: 194588b4-fe40-4286-8036-874a7c410327] return func(*args, **kwargs) [ 589.076095] env[62066]: ERROR nova.compute.manager [instance: 194588b4-fe40-4286-8036-874a7c410327] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 589.076095] env[62066]: ERROR nova.compute.manager [instance: 194588b4-fe40-4286-8036-874a7c410327] raise e [ 589.076095] env[62066]: ERROR nova.compute.manager [instance: 194588b4-fe40-4286-8036-874a7c410327] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 589.076095] env[62066]: ERROR nova.compute.manager [instance: 194588b4-fe40-4286-8036-874a7c410327] nwinfo = self.network_api.allocate_for_instance( [ 589.076095] env[62066]: ERROR nova.compute.manager [instance: 194588b4-fe40-4286-8036-874a7c410327] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 589.076095] env[62066]: ERROR nova.compute.manager [instance: 194588b4-fe40-4286-8036-874a7c410327] created_port_ids = self._update_ports_for_instance( [ 589.076454] env[62066]: ERROR nova.compute.manager [instance: 194588b4-fe40-4286-8036-874a7c410327] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 589.076454] env[62066]: ERROR nova.compute.manager [instance: 194588b4-fe40-4286-8036-874a7c410327] with excutils.save_and_reraise_exception(): [ 589.076454] env[62066]: ERROR nova.compute.manager [instance: 194588b4-fe40-4286-8036-874a7c410327] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 589.076454] env[62066]: ERROR nova.compute.manager [instance: 194588b4-fe40-4286-8036-874a7c410327] self.force_reraise() [ 589.076454] env[62066]: ERROR nova.compute.manager [instance: 194588b4-fe40-4286-8036-874a7c410327] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 589.076454] env[62066]: ERROR nova.compute.manager [instance: 194588b4-fe40-4286-8036-874a7c410327] raise self.value [ 589.076454] env[62066]: ERROR nova.compute.manager [instance: 194588b4-fe40-4286-8036-874a7c410327] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 589.076454] env[62066]: ERROR nova.compute.manager [instance: 194588b4-fe40-4286-8036-874a7c410327] updated_port = self._update_port( [ 589.076454] env[62066]: ERROR nova.compute.manager [instance: 194588b4-fe40-4286-8036-874a7c410327] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 589.076454] env[62066]: ERROR nova.compute.manager [instance: 194588b4-fe40-4286-8036-874a7c410327] _ensure_no_port_binding_failure(port) [ 589.076454] env[62066]: ERROR nova.compute.manager [instance: 194588b4-fe40-4286-8036-874a7c410327] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 589.076454] env[62066]: ERROR nova.compute.manager [instance: 194588b4-fe40-4286-8036-874a7c410327] raise exception.PortBindingFailed(port_id=port['id']) [ 589.076765] env[62066]: ERROR nova.compute.manager [instance: 194588b4-fe40-4286-8036-874a7c410327] nova.exception.PortBindingFailed: Binding failed for port 2a2c9bbf-b369-4ea3-a866-79152380ed0b, please check neutron logs for more information. [ 589.076765] env[62066]: ERROR nova.compute.manager [instance: 194588b4-fe40-4286-8036-874a7c410327] [ 589.076765] env[62066]: INFO nova.compute.manager [None req-1e3e30d6-8bb5-4347-b6c2-c2d648c37b67 tempest-ImagesOneServerNegativeTestJSON-310236551 tempest-ImagesOneServerNegativeTestJSON-310236551-project-member] [instance: 194588b4-fe40-4286-8036-874a7c410327] Terminating instance [ 589.079655] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1e3e30d6-8bb5-4347-b6c2-c2d648c37b67 tempest-ImagesOneServerNegativeTestJSON-310236551 tempest-ImagesOneServerNegativeTestJSON-310236551-project-member] Acquiring lock "refresh_cache-194588b4-fe40-4286-8036-874a7c410327" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 589.104176] env[62066]: DEBUG nova.network.neutron [req-c694cdae-aa25-48e0-8de8-bfe8bd0661e4 req-cb31e0e9-a33c-464b-9920-a34dfae482ca service nova] [instance: 194588b4-fe40-4286-8036-874a7c410327] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 589.281859] env[62066]: DEBUG nova.network.neutron [req-c694cdae-aa25-48e0-8de8-bfe8bd0661e4 req-cb31e0e9-a33c-464b-9920-a34dfae482ca service nova] [instance: 194588b4-fe40-4286-8036-874a7c410327] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 589.332717] env[62066]: DEBUG nova.scheduler.client.report [None req-fc3cf2dd-2491-4541-a245-5eefc9a5c042 tempest-TenantUsagesTestJSON-970395509 tempest-TenantUsagesTestJSON-970395509-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 589.550695] env[62066]: INFO nova.compute.manager [None req-cb29a201-afba-4eb8-8336-fe27a681c8f8 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] [instance: 69d99063-a6a7-4832-a864-dc9fd6024ea8] Took 1.04 seconds to deallocate network for instance. [ 589.784366] env[62066]: DEBUG oslo_concurrency.lockutils [req-c694cdae-aa25-48e0-8de8-bfe8bd0661e4 req-cb31e0e9-a33c-464b-9920-a34dfae482ca service nova] Releasing lock "refresh_cache-194588b4-fe40-4286-8036-874a7c410327" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 589.784892] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1e3e30d6-8bb5-4347-b6c2-c2d648c37b67 tempest-ImagesOneServerNegativeTestJSON-310236551 tempest-ImagesOneServerNegativeTestJSON-310236551-project-member] Acquired lock "refresh_cache-194588b4-fe40-4286-8036-874a7c410327" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 589.785037] env[62066]: DEBUG nova.network.neutron [None req-1e3e30d6-8bb5-4347-b6c2-c2d648c37b67 tempest-ImagesOneServerNegativeTestJSON-310236551 tempest-ImagesOneServerNegativeTestJSON-310236551-project-member] [instance: 194588b4-fe40-4286-8036-874a7c410327] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 589.841156] env[62066]: DEBUG oslo_concurrency.lockutils [None req-fc3cf2dd-2491-4541-a245-5eefc9a5c042 tempest-TenantUsagesTestJSON-970395509 tempest-TenantUsagesTestJSON-970395509-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.594s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 589.841761] env[62066]: DEBUG nova.compute.manager [None req-fc3cf2dd-2491-4541-a245-5eefc9a5c042 tempest-TenantUsagesTestJSON-970395509 tempest-TenantUsagesTestJSON-970395509-project-member] [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 589.845075] env[62066]: DEBUG oslo_concurrency.lockutils [None req-65bf05da-7d56-4074-8d13-63fde71b9619 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 20.800s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 590.329315] env[62066]: DEBUG nova.network.neutron [None req-1e3e30d6-8bb5-4347-b6c2-c2d648c37b67 tempest-ImagesOneServerNegativeTestJSON-310236551 tempest-ImagesOneServerNegativeTestJSON-310236551-project-member] [instance: 194588b4-fe40-4286-8036-874a7c410327] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 590.349652] env[62066]: DEBUG nova.compute.utils [None req-fc3cf2dd-2491-4541-a245-5eefc9a5c042 tempest-TenantUsagesTestJSON-970395509 tempest-TenantUsagesTestJSON-970395509-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 590.354189] env[62066]: DEBUG nova.compute.manager [None req-fc3cf2dd-2491-4541-a245-5eefc9a5c042 tempest-TenantUsagesTestJSON-970395509 tempest-TenantUsagesTestJSON-970395509-project-member] [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 590.354364] env[62066]: DEBUG nova.network.neutron [None req-fc3cf2dd-2491-4541-a245-5eefc9a5c042 tempest-TenantUsagesTestJSON-970395509 tempest-TenantUsagesTestJSON-970395509-project-member] [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 590.451521] env[62066]: DEBUG nova.policy [None req-fc3cf2dd-2491-4541-a245-5eefc9a5c042 tempest-TenantUsagesTestJSON-970395509 tempest-TenantUsagesTestJSON-970395509-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd6c6f1fc9fcf4d7299bb686c32256b96', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '590a223ccd1c4fdfb9fd7e0ba1408ee6', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 590.522298] env[62066]: DEBUG nova.network.neutron [None req-1e3e30d6-8bb5-4347-b6c2-c2d648c37b67 tempest-ImagesOneServerNegativeTestJSON-310236551 tempest-ImagesOneServerNegativeTestJSON-310236551-project-member] [instance: 194588b4-fe40-4286-8036-874a7c410327] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 590.605444] env[62066]: INFO nova.scheduler.client.report [None req-cb29a201-afba-4eb8-8336-fe27a681c8f8 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] Deleted allocations for instance 69d99063-a6a7-4832-a864-dc9fd6024ea8 [ 590.736351] env[62066]: DEBUG nova.compute.manager [req-3094dbfd-96b4-421b-a3a6-da55ec0aa533 req-1980b293-8783-4a35-96cb-b4a25e897609 service nova] [instance: 194588b4-fe40-4286-8036-874a7c410327] Received event network-vif-deleted-2a2c9bbf-b369-4ea3-a866-79152380ed0b {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 590.854796] env[62066]: DEBUG nova.compute.manager [None req-fc3cf2dd-2491-4541-a245-5eefc9a5c042 tempest-TenantUsagesTestJSON-970395509 tempest-TenantUsagesTestJSON-970395509-project-member] [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 590.925860] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5341cc11-7b4a-4f5f-aaf3-e68232232274 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.934527] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b38ea7a-7913-4192-af32-55a082cd0628 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.967774] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0b831cf-e16f-451f-826a-228b9fbe714b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.975990] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88e688a9-d39a-4942-a792-c15154e1555d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.990089] env[62066]: DEBUG nova.compute.provider_tree [None req-65bf05da-7d56-4074-8d13-63fde71b9619 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 591.031022] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1e3e30d6-8bb5-4347-b6c2-c2d648c37b67 tempest-ImagesOneServerNegativeTestJSON-310236551 tempest-ImagesOneServerNegativeTestJSON-310236551-project-member] Releasing lock "refresh_cache-194588b4-fe40-4286-8036-874a7c410327" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 591.031256] env[62066]: DEBUG nova.compute.manager [None req-1e3e30d6-8bb5-4347-b6c2-c2d648c37b67 tempest-ImagesOneServerNegativeTestJSON-310236551 tempest-ImagesOneServerNegativeTestJSON-310236551-project-member] [instance: 194588b4-fe40-4286-8036-874a7c410327] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 591.031450] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-1e3e30d6-8bb5-4347-b6c2-c2d648c37b67 tempest-ImagesOneServerNegativeTestJSON-310236551 tempest-ImagesOneServerNegativeTestJSON-310236551-project-member] [instance: 194588b4-fe40-4286-8036-874a7c410327] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 591.032484] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-74f30973-340b-4df4-acd8-84a9f9cdb707 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.042662] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21acf152-96eb-4881-a6a8-93782c619eb9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.068048] env[62066]: WARNING nova.virt.vmwareapi.vmops [None req-1e3e30d6-8bb5-4347-b6c2-c2d648c37b67 tempest-ImagesOneServerNegativeTestJSON-310236551 tempest-ImagesOneServerNegativeTestJSON-310236551-project-member] [instance: 194588b4-fe40-4286-8036-874a7c410327] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 194588b4-fe40-4286-8036-874a7c410327 could not be found. [ 591.068491] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-1e3e30d6-8bb5-4347-b6c2-c2d648c37b67 tempest-ImagesOneServerNegativeTestJSON-310236551 tempest-ImagesOneServerNegativeTestJSON-310236551-project-member] [instance: 194588b4-fe40-4286-8036-874a7c410327] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 591.068689] env[62066]: INFO nova.compute.manager [None req-1e3e30d6-8bb5-4347-b6c2-c2d648c37b67 tempest-ImagesOneServerNegativeTestJSON-310236551 tempest-ImagesOneServerNegativeTestJSON-310236551-project-member] [instance: 194588b4-fe40-4286-8036-874a7c410327] Took 0.04 seconds to destroy the instance on the hypervisor. [ 591.068936] env[62066]: DEBUG oslo.service.loopingcall [None req-1e3e30d6-8bb5-4347-b6c2-c2d648c37b67 tempest-ImagesOneServerNegativeTestJSON-310236551 tempest-ImagesOneServerNegativeTestJSON-310236551-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 591.069180] env[62066]: DEBUG nova.compute.manager [-] [instance: 194588b4-fe40-4286-8036-874a7c410327] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 591.069276] env[62066]: DEBUG nova.network.neutron [-] [instance: 194588b4-fe40-4286-8036-874a7c410327] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 591.115651] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cb29a201-afba-4eb8-8336-fe27a681c8f8 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] Lock "69d99063-a6a7-4832-a864-dc9fd6024ea8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 80.534s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 591.174691] env[62066]: DEBUG nova.network.neutron [None req-fc3cf2dd-2491-4541-a245-5eefc9a5c042 tempest-TenantUsagesTestJSON-970395509 tempest-TenantUsagesTestJSON-970395509-project-member] [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] Successfully created port: 601feca8-30e1-4972-a3b1-25d0817e7e50 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 591.205131] env[62066]: DEBUG nova.network.neutron [-] [instance: 194588b4-fe40-4286-8036-874a7c410327] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 591.494447] env[62066]: DEBUG nova.scheduler.client.report [None req-65bf05da-7d56-4074-8d13-63fde71b9619 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 591.618894] env[62066]: DEBUG nova.compute.manager [None req-14085d2f-7e58-404a-a2d0-4487ac658e45 tempest-FloatingIPsAssociationTestJSON-608653087 tempest-FloatingIPsAssociationTestJSON-608653087-project-member] [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 591.708476] env[62066]: DEBUG nova.network.neutron [-] [instance: 194588b4-fe40-4286-8036-874a7c410327] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 591.866030] env[62066]: DEBUG nova.compute.manager [None req-fc3cf2dd-2491-4541-a245-5eefc9a5c042 tempest-TenantUsagesTestJSON-970395509 tempest-TenantUsagesTestJSON-970395509-project-member] [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 591.901591] env[62066]: DEBUG nova.virt.hardware [None req-fc3cf2dd-2491-4541-a245-5eefc9a5c042 tempest-TenantUsagesTestJSON-970395509 tempest-TenantUsagesTestJSON-970395509-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-23T13:40:41Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-23T13:40:23Z,direct_url=,disk_format='vmdk',id=50ff584c-3b50-4395-af07-3e66769bc9f7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='eb52f7069a374c61ae946f052007c6d9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-23T13:40:24Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 591.901591] env[62066]: DEBUG nova.virt.hardware [None req-fc3cf2dd-2491-4541-a245-5eefc9a5c042 tempest-TenantUsagesTestJSON-970395509 tempest-TenantUsagesTestJSON-970395509-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 591.901591] env[62066]: DEBUG nova.virt.hardware [None req-fc3cf2dd-2491-4541-a245-5eefc9a5c042 tempest-TenantUsagesTestJSON-970395509 tempest-TenantUsagesTestJSON-970395509-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 591.901591] env[62066]: DEBUG nova.virt.hardware [None req-fc3cf2dd-2491-4541-a245-5eefc9a5c042 tempest-TenantUsagesTestJSON-970395509 tempest-TenantUsagesTestJSON-970395509-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 591.901803] env[62066]: DEBUG nova.virt.hardware [None req-fc3cf2dd-2491-4541-a245-5eefc9a5c042 tempest-TenantUsagesTestJSON-970395509 tempest-TenantUsagesTestJSON-970395509-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 591.901803] env[62066]: DEBUG nova.virt.hardware [None req-fc3cf2dd-2491-4541-a245-5eefc9a5c042 tempest-TenantUsagesTestJSON-970395509 tempest-TenantUsagesTestJSON-970395509-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 591.901803] env[62066]: DEBUG nova.virt.hardware [None req-fc3cf2dd-2491-4541-a245-5eefc9a5c042 tempest-TenantUsagesTestJSON-970395509 tempest-TenantUsagesTestJSON-970395509-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 591.901803] env[62066]: DEBUG nova.virt.hardware [None req-fc3cf2dd-2491-4541-a245-5eefc9a5c042 tempest-TenantUsagesTestJSON-970395509 tempest-TenantUsagesTestJSON-970395509-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 591.903453] env[62066]: DEBUG nova.virt.hardware [None req-fc3cf2dd-2491-4541-a245-5eefc9a5c042 tempest-TenantUsagesTestJSON-970395509 tempest-TenantUsagesTestJSON-970395509-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 591.903453] env[62066]: DEBUG nova.virt.hardware [None req-fc3cf2dd-2491-4541-a245-5eefc9a5c042 tempest-TenantUsagesTestJSON-970395509 tempest-TenantUsagesTestJSON-970395509-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 591.903453] env[62066]: DEBUG nova.virt.hardware [None req-fc3cf2dd-2491-4541-a245-5eefc9a5c042 tempest-TenantUsagesTestJSON-970395509 tempest-TenantUsagesTestJSON-970395509-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 591.903453] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a185457-2e79-4300-88dd-6805f2ed0116 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.914313] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84b4a7b0-f656-4dac-9390-a09d1d6f1dfc {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.999430] env[62066]: DEBUG oslo_concurrency.lockutils [None req-65bf05da-7d56-4074-8d13-63fde71b9619 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.154s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 592.000095] env[62066]: ERROR nova.compute.manager [None req-65bf05da-7d56-4074-8d13-63fde71b9619 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 39ab0f84-58c5-4569-bf28-6dbda905c21d, please check neutron logs for more information. [ 592.000095] env[62066]: ERROR nova.compute.manager [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] Traceback (most recent call last): [ 592.000095] env[62066]: ERROR nova.compute.manager [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 592.000095] env[62066]: ERROR nova.compute.manager [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] self.driver.spawn(context, instance, image_meta, [ 592.000095] env[62066]: ERROR nova.compute.manager [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 592.000095] env[62066]: ERROR nova.compute.manager [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] self._vmops.spawn(context, instance, image_meta, injected_files, [ 592.000095] env[62066]: ERROR nova.compute.manager [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 592.000095] env[62066]: ERROR nova.compute.manager [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] vm_ref = self.build_virtual_machine(instance, [ 592.000095] env[62066]: ERROR nova.compute.manager [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 592.000095] env[62066]: ERROR nova.compute.manager [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] vif_infos = vmwarevif.get_vif_info(self._session, [ 592.000095] env[62066]: ERROR nova.compute.manager [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 592.000359] env[62066]: ERROR nova.compute.manager [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] for vif in network_info: [ 592.000359] env[62066]: ERROR nova.compute.manager [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 592.000359] env[62066]: ERROR nova.compute.manager [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] return self._sync_wrapper(fn, *args, **kwargs) [ 592.000359] env[62066]: ERROR nova.compute.manager [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 592.000359] env[62066]: ERROR nova.compute.manager [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] self.wait() [ 592.000359] env[62066]: ERROR nova.compute.manager [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 592.000359] env[62066]: ERROR nova.compute.manager [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] self[:] = self._gt.wait() [ 592.000359] env[62066]: ERROR nova.compute.manager [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 592.000359] env[62066]: ERROR nova.compute.manager [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] return self._exit_event.wait() [ 592.000359] env[62066]: ERROR nova.compute.manager [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 592.000359] env[62066]: ERROR nova.compute.manager [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] result = hub.switch() [ 592.000359] env[62066]: ERROR nova.compute.manager [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 592.000359] env[62066]: ERROR nova.compute.manager [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] return self.greenlet.switch() [ 592.000649] env[62066]: ERROR nova.compute.manager [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 592.000649] env[62066]: ERROR nova.compute.manager [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] result = function(*args, **kwargs) [ 592.000649] env[62066]: ERROR nova.compute.manager [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 592.000649] env[62066]: ERROR nova.compute.manager [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] return func(*args, **kwargs) [ 592.000649] env[62066]: ERROR nova.compute.manager [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 592.000649] env[62066]: ERROR nova.compute.manager [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] raise e [ 592.000649] env[62066]: ERROR nova.compute.manager [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 592.000649] env[62066]: ERROR nova.compute.manager [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] nwinfo = self.network_api.allocate_for_instance( [ 592.000649] env[62066]: ERROR nova.compute.manager [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 592.000649] env[62066]: ERROR nova.compute.manager [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] created_port_ids = self._update_ports_for_instance( [ 592.000649] env[62066]: ERROR nova.compute.manager [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 592.000649] env[62066]: ERROR nova.compute.manager [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] with excutils.save_and_reraise_exception(): [ 592.000649] env[62066]: ERROR nova.compute.manager [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 592.000979] env[62066]: ERROR nova.compute.manager [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] self.force_reraise() [ 592.000979] env[62066]: ERROR nova.compute.manager [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 592.000979] env[62066]: ERROR nova.compute.manager [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] raise self.value [ 592.000979] env[62066]: ERROR nova.compute.manager [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 592.000979] env[62066]: ERROR nova.compute.manager [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] updated_port = self._update_port( [ 592.000979] env[62066]: ERROR nova.compute.manager [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 592.000979] env[62066]: ERROR nova.compute.manager [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] _ensure_no_port_binding_failure(port) [ 592.000979] env[62066]: ERROR nova.compute.manager [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 592.000979] env[62066]: ERROR nova.compute.manager [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] raise exception.PortBindingFailed(port_id=port['id']) [ 592.000979] env[62066]: ERROR nova.compute.manager [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] nova.exception.PortBindingFailed: Binding failed for port 39ab0f84-58c5-4569-bf28-6dbda905c21d, please check neutron logs for more information. [ 592.000979] env[62066]: ERROR nova.compute.manager [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] [ 592.001311] env[62066]: DEBUG nova.compute.utils [None req-65bf05da-7d56-4074-8d13-63fde71b9619 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] Binding failed for port 39ab0f84-58c5-4569-bf28-6dbda905c21d, please check neutron logs for more information. {{(pid=62066) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 592.003343] env[62066]: DEBUG nova.compute.manager [None req-65bf05da-7d56-4074-8d13-63fde71b9619 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] Build of instance eacdd472-5bb0-48bc-8603-9e1d08b41f55 was re-scheduled: Binding failed for port 39ab0f84-58c5-4569-bf28-6dbda905c21d, please check neutron logs for more information. {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 592.003637] env[62066]: DEBUG nova.compute.manager [None req-65bf05da-7d56-4074-8d13-63fde71b9619 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] Unplugging VIFs for instance {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 592.003869] env[62066]: DEBUG oslo_concurrency.lockutils [None req-65bf05da-7d56-4074-8d13-63fde71b9619 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] Acquiring lock "refresh_cache-eacdd472-5bb0-48bc-8603-9e1d08b41f55" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 592.004233] env[62066]: DEBUG oslo_concurrency.lockutils [None req-65bf05da-7d56-4074-8d13-63fde71b9619 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] Acquired lock "refresh_cache-eacdd472-5bb0-48bc-8603-9e1d08b41f55" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 592.004741] env[62066]: DEBUG nova.network.neutron [None req-65bf05da-7d56-4074-8d13-63fde71b9619 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 592.005867] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e7221104-9243-4d76-b03e-5782ae8f55d4 tempest-FloatingIPsAssociationNegativeTestJSON-961065153 tempest-FloatingIPsAssociationNegativeTestJSON-961065153-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.692s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 592.007852] env[62066]: INFO nova.compute.claims [None req-e7221104-9243-4d76-b03e-5782ae8f55d4 tempest-FloatingIPsAssociationNegativeTestJSON-961065153 tempest-FloatingIPsAssociationNegativeTestJSON-961065153-project-member] [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 592.146286] env[62066]: DEBUG oslo_concurrency.lockutils [None req-14085d2f-7e58-404a-a2d0-4487ac658e45 tempest-FloatingIPsAssociationTestJSON-608653087 tempest-FloatingIPsAssociationTestJSON-608653087-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 592.212074] env[62066]: INFO nova.compute.manager [-] [instance: 194588b4-fe40-4286-8036-874a7c410327] Took 1.14 seconds to deallocate network for instance. [ 592.214581] env[62066]: DEBUG nova.compute.claims [None req-1e3e30d6-8bb5-4347-b6c2-c2d648c37b67 tempest-ImagesOneServerNegativeTestJSON-310236551 tempest-ImagesOneServerNegativeTestJSON-310236551-project-member] [instance: 194588b4-fe40-4286-8036-874a7c410327] Aborting claim: {{(pid=62066) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 592.214737] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1e3e30d6-8bb5-4347-b6c2-c2d648c37b67 tempest-ImagesOneServerNegativeTestJSON-310236551 tempest-ImagesOneServerNegativeTestJSON-310236551-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 592.407396] env[62066]: DEBUG nova.compute.manager [req-b1d2b0af-9b92-4ae2-95b8-38dcc03c6118 req-86584cbb-6be0-466d-a444-7fc698b979a7 service nova] [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] Received event network-changed-601feca8-30e1-4972-a3b1-25d0817e7e50 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 592.407600] env[62066]: DEBUG nova.compute.manager [req-b1d2b0af-9b92-4ae2-95b8-38dcc03c6118 req-86584cbb-6be0-466d-a444-7fc698b979a7 service nova] [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] Refreshing instance network info cache due to event network-changed-601feca8-30e1-4972-a3b1-25d0817e7e50. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 592.407830] env[62066]: DEBUG oslo_concurrency.lockutils [req-b1d2b0af-9b92-4ae2-95b8-38dcc03c6118 req-86584cbb-6be0-466d-a444-7fc698b979a7 service nova] Acquiring lock "refresh_cache-a187f73b-8fb1-42b3-9a07-24d16aa0f152" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 592.407943] env[62066]: DEBUG oslo_concurrency.lockutils [req-b1d2b0af-9b92-4ae2-95b8-38dcc03c6118 req-86584cbb-6be0-466d-a444-7fc698b979a7 service nova] Acquired lock "refresh_cache-a187f73b-8fb1-42b3-9a07-24d16aa0f152" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 592.410704] env[62066]: DEBUG nova.network.neutron [req-b1d2b0af-9b92-4ae2-95b8-38dcc03c6118 req-86584cbb-6be0-466d-a444-7fc698b979a7 service nova] [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] Refreshing network info cache for port 601feca8-30e1-4972-a3b1-25d0817e7e50 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 592.552027] env[62066]: DEBUG nova.network.neutron [None req-65bf05da-7d56-4074-8d13-63fde71b9619 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 592.577046] env[62066]: ERROR nova.compute.manager [None req-fc3cf2dd-2491-4541-a245-5eefc9a5c042 tempest-TenantUsagesTestJSON-970395509 tempest-TenantUsagesTestJSON-970395509-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 601feca8-30e1-4972-a3b1-25d0817e7e50, please check neutron logs for more information. [ 592.577046] env[62066]: ERROR nova.compute.manager Traceback (most recent call last): [ 592.577046] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 592.577046] env[62066]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 592.577046] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 592.577046] env[62066]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 592.577046] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 592.577046] env[62066]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 592.577046] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 592.577046] env[62066]: ERROR nova.compute.manager self.force_reraise() [ 592.577046] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 592.577046] env[62066]: ERROR nova.compute.manager raise self.value [ 592.577046] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 592.577046] env[62066]: ERROR nova.compute.manager updated_port = self._update_port( [ 592.577046] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 592.577046] env[62066]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 592.579320] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 592.579320] env[62066]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 592.579320] env[62066]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 601feca8-30e1-4972-a3b1-25d0817e7e50, please check neutron logs for more information. [ 592.579320] env[62066]: ERROR nova.compute.manager [ 592.579320] env[62066]: Traceback (most recent call last): [ 592.579320] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 592.579320] env[62066]: listener.cb(fileno) [ 592.579320] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 592.579320] env[62066]: result = function(*args, **kwargs) [ 592.579320] env[62066]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 592.579320] env[62066]: return func(*args, **kwargs) [ 592.579320] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 592.579320] env[62066]: raise e [ 592.579320] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 592.579320] env[62066]: nwinfo = self.network_api.allocate_for_instance( [ 592.579320] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 592.579320] env[62066]: created_port_ids = self._update_ports_for_instance( [ 592.579320] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 592.579320] env[62066]: with excutils.save_and_reraise_exception(): [ 592.579320] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 592.579320] env[62066]: self.force_reraise() [ 592.579320] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 592.579320] env[62066]: raise self.value [ 592.579320] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 592.579320] env[62066]: updated_port = self._update_port( [ 592.579320] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 592.579320] env[62066]: _ensure_no_port_binding_failure(port) [ 592.579320] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 592.579320] env[62066]: raise exception.PortBindingFailed(port_id=port['id']) [ 592.580063] env[62066]: nova.exception.PortBindingFailed: Binding failed for port 601feca8-30e1-4972-a3b1-25d0817e7e50, please check neutron logs for more information. [ 592.580063] env[62066]: Removing descriptor: 15 [ 592.580063] env[62066]: ERROR nova.compute.manager [None req-fc3cf2dd-2491-4541-a245-5eefc9a5c042 tempest-TenantUsagesTestJSON-970395509 tempest-TenantUsagesTestJSON-970395509-project-member] [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 601feca8-30e1-4972-a3b1-25d0817e7e50, please check neutron logs for more information. [ 592.580063] env[62066]: ERROR nova.compute.manager [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] Traceback (most recent call last): [ 592.580063] env[62066]: ERROR nova.compute.manager [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 592.580063] env[62066]: ERROR nova.compute.manager [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] yield resources [ 592.580063] env[62066]: ERROR nova.compute.manager [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 592.580063] env[62066]: ERROR nova.compute.manager [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] self.driver.spawn(context, instance, image_meta, [ 592.580063] env[62066]: ERROR nova.compute.manager [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 592.580063] env[62066]: ERROR nova.compute.manager [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] self._vmops.spawn(context, instance, image_meta, injected_files, [ 592.580063] env[62066]: ERROR nova.compute.manager [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 592.580063] env[62066]: ERROR nova.compute.manager [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] vm_ref = self.build_virtual_machine(instance, [ 592.580360] env[62066]: ERROR nova.compute.manager [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 592.580360] env[62066]: ERROR nova.compute.manager [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] vif_infos = vmwarevif.get_vif_info(self._session, [ 592.580360] env[62066]: ERROR nova.compute.manager [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 592.580360] env[62066]: ERROR nova.compute.manager [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] for vif in network_info: [ 592.580360] env[62066]: ERROR nova.compute.manager [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 592.580360] env[62066]: ERROR nova.compute.manager [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] return self._sync_wrapper(fn, *args, **kwargs) [ 592.580360] env[62066]: ERROR nova.compute.manager [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 592.580360] env[62066]: ERROR nova.compute.manager [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] self.wait() [ 592.580360] env[62066]: ERROR nova.compute.manager [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 592.580360] env[62066]: ERROR nova.compute.manager [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] self[:] = self._gt.wait() [ 592.580360] env[62066]: ERROR nova.compute.manager [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 592.580360] env[62066]: ERROR nova.compute.manager [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] return self._exit_event.wait() [ 592.580360] env[62066]: ERROR nova.compute.manager [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 592.580667] env[62066]: ERROR nova.compute.manager [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] result = hub.switch() [ 592.580667] env[62066]: ERROR nova.compute.manager [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 592.580667] env[62066]: ERROR nova.compute.manager [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] return self.greenlet.switch() [ 592.580667] env[62066]: ERROR nova.compute.manager [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 592.580667] env[62066]: ERROR nova.compute.manager [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] result = function(*args, **kwargs) [ 592.580667] env[62066]: ERROR nova.compute.manager [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 592.580667] env[62066]: ERROR nova.compute.manager [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] return func(*args, **kwargs) [ 592.580667] env[62066]: ERROR nova.compute.manager [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 592.580667] env[62066]: ERROR nova.compute.manager [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] raise e [ 592.580667] env[62066]: ERROR nova.compute.manager [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 592.580667] env[62066]: ERROR nova.compute.manager [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] nwinfo = self.network_api.allocate_for_instance( [ 592.580667] env[62066]: ERROR nova.compute.manager [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 592.580667] env[62066]: ERROR nova.compute.manager [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] created_port_ids = self._update_ports_for_instance( [ 592.581044] env[62066]: ERROR nova.compute.manager [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 592.581044] env[62066]: ERROR nova.compute.manager [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] with excutils.save_and_reraise_exception(): [ 592.581044] env[62066]: ERROR nova.compute.manager [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 592.581044] env[62066]: ERROR nova.compute.manager [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] self.force_reraise() [ 592.581044] env[62066]: ERROR nova.compute.manager [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 592.581044] env[62066]: ERROR nova.compute.manager [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] raise self.value [ 592.581044] env[62066]: ERROR nova.compute.manager [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 592.581044] env[62066]: ERROR nova.compute.manager [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] updated_port = self._update_port( [ 592.581044] env[62066]: ERROR nova.compute.manager [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 592.581044] env[62066]: ERROR nova.compute.manager [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] _ensure_no_port_binding_failure(port) [ 592.581044] env[62066]: ERROR nova.compute.manager [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 592.581044] env[62066]: ERROR nova.compute.manager [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] raise exception.PortBindingFailed(port_id=port['id']) [ 592.581319] env[62066]: ERROR nova.compute.manager [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] nova.exception.PortBindingFailed: Binding failed for port 601feca8-30e1-4972-a3b1-25d0817e7e50, please check neutron logs for more information. [ 592.581319] env[62066]: ERROR nova.compute.manager [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] [ 592.581319] env[62066]: INFO nova.compute.manager [None req-fc3cf2dd-2491-4541-a245-5eefc9a5c042 tempest-TenantUsagesTestJSON-970395509 tempest-TenantUsagesTestJSON-970395509-project-member] [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] Terminating instance [ 592.581319] env[62066]: DEBUG oslo_concurrency.lockutils [None req-fc3cf2dd-2491-4541-a245-5eefc9a5c042 tempest-TenantUsagesTestJSON-970395509 tempest-TenantUsagesTestJSON-970395509-project-member] Acquiring lock "refresh_cache-a187f73b-8fb1-42b3-9a07-24d16aa0f152" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 592.849679] env[62066]: DEBUG nova.network.neutron [None req-65bf05da-7d56-4074-8d13-63fde71b9619 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 592.943088] env[62066]: DEBUG nova.network.neutron [req-b1d2b0af-9b92-4ae2-95b8-38dcc03c6118 req-86584cbb-6be0-466d-a444-7fc698b979a7 service nova] [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 593.083937] env[62066]: DEBUG nova.network.neutron [req-b1d2b0af-9b92-4ae2-95b8-38dcc03c6118 req-86584cbb-6be0-466d-a444-7fc698b979a7 service nova] [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 593.355518] env[62066]: DEBUG oslo_concurrency.lockutils [None req-65bf05da-7d56-4074-8d13-63fde71b9619 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] Releasing lock "refresh_cache-eacdd472-5bb0-48bc-8603-9e1d08b41f55" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 593.355759] env[62066]: DEBUG nova.compute.manager [None req-65bf05da-7d56-4074-8d13-63fde71b9619 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 593.355942] env[62066]: DEBUG nova.compute.manager [None req-65bf05da-7d56-4074-8d13-63fde71b9619 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 593.356125] env[62066]: DEBUG nova.network.neutron [None req-65bf05da-7d56-4074-8d13-63fde71b9619 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 593.378520] env[62066]: DEBUG nova.network.neutron [None req-65bf05da-7d56-4074-8d13-63fde71b9619 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 593.496698] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25ac0066-c5d1-4c8d-9fee-79ab53a0bf38 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.507183] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-506e068d-1d54-4a4a-b652-3b78ae020bed {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.542849] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c630706-88c4-4704-8d0d-84e14d73c61f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.551015] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bede787-e74f-485b-89f8-dba1eb7c5fc1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.565223] env[62066]: DEBUG nova.compute.provider_tree [None req-e7221104-9243-4d76-b03e-5782ae8f55d4 tempest-FloatingIPsAssociationNegativeTestJSON-961065153 tempest-FloatingIPsAssociationNegativeTestJSON-961065153-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 593.589825] env[62066]: DEBUG oslo_concurrency.lockutils [req-b1d2b0af-9b92-4ae2-95b8-38dcc03c6118 req-86584cbb-6be0-466d-a444-7fc698b979a7 service nova] Releasing lock "refresh_cache-a187f73b-8fb1-42b3-9a07-24d16aa0f152" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 593.590436] env[62066]: DEBUG oslo_concurrency.lockutils [None req-fc3cf2dd-2491-4541-a245-5eefc9a5c042 tempest-TenantUsagesTestJSON-970395509 tempest-TenantUsagesTestJSON-970395509-project-member] Acquired lock "refresh_cache-a187f73b-8fb1-42b3-9a07-24d16aa0f152" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 593.590620] env[62066]: DEBUG nova.network.neutron [None req-fc3cf2dd-2491-4541-a245-5eefc9a5c042 tempest-TenantUsagesTestJSON-970395509 tempest-TenantUsagesTestJSON-970395509-project-member] [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 593.881178] env[62066]: DEBUG nova.network.neutron [None req-65bf05da-7d56-4074-8d13-63fde71b9619 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 594.068232] env[62066]: DEBUG nova.scheduler.client.report [None req-e7221104-9243-4d76-b03e-5782ae8f55d4 tempest-FloatingIPsAssociationNegativeTestJSON-961065153 tempest-FloatingIPsAssociationNegativeTestJSON-961065153-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 594.291081] env[62066]: DEBUG nova.network.neutron [None req-fc3cf2dd-2491-4541-a245-5eefc9a5c042 tempest-TenantUsagesTestJSON-970395509 tempest-TenantUsagesTestJSON-970395509-project-member] [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 594.381407] env[62066]: DEBUG nova.network.neutron [None req-fc3cf2dd-2491-4541-a245-5eefc9a5c042 tempest-TenantUsagesTestJSON-970395509 tempest-TenantUsagesTestJSON-970395509-project-member] [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 594.386286] env[62066]: INFO nova.compute.manager [None req-65bf05da-7d56-4074-8d13-63fde71b9619 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] [instance: eacdd472-5bb0-48bc-8603-9e1d08b41f55] Took 1.03 seconds to deallocate network for instance. [ 594.502682] env[62066]: DEBUG nova.compute.manager [req-2ccc11ba-8201-48be-a7e6-5d5dbdc81022 req-47670186-c502-4b6f-b77b-3c6ec3507acf service nova] [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] Received event network-vif-deleted-601feca8-30e1-4972-a3b1-25d0817e7e50 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 594.578420] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e7221104-9243-4d76-b03e-5782ae8f55d4 tempest-FloatingIPsAssociationNegativeTestJSON-961065153 tempest-FloatingIPsAssociationNegativeTestJSON-961065153-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.573s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 594.578997] env[62066]: DEBUG nova.compute.manager [None req-e7221104-9243-4d76-b03e-5782ae8f55d4 tempest-FloatingIPsAssociationNegativeTestJSON-961065153 tempest-FloatingIPsAssociationNegativeTestJSON-961065153-project-member] [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 594.581963] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4ae0c9b4-adca-4b03-8ac7-409a4ee035f6 tempest-InstanceActionsTestJSON-1008103477 tempest-InstanceActionsTestJSON-1008103477-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.809s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 594.884858] env[62066]: DEBUG oslo_concurrency.lockutils [None req-fc3cf2dd-2491-4541-a245-5eefc9a5c042 tempest-TenantUsagesTestJSON-970395509 tempest-TenantUsagesTestJSON-970395509-project-member] Releasing lock "refresh_cache-a187f73b-8fb1-42b3-9a07-24d16aa0f152" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 594.885314] env[62066]: DEBUG nova.compute.manager [None req-fc3cf2dd-2491-4541-a245-5eefc9a5c042 tempest-TenantUsagesTestJSON-970395509 tempest-TenantUsagesTestJSON-970395509-project-member] [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 594.885505] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-fc3cf2dd-2491-4541-a245-5eefc9a5c042 tempest-TenantUsagesTestJSON-970395509 tempest-TenantUsagesTestJSON-970395509-project-member] [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 594.887093] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-876d86a8-8d5f-44de-ab7a-8337d1762e71 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.897192] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b846ca18-bbd6-46e1-b615-74b4dae30884 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.921523] env[62066]: WARNING nova.virt.vmwareapi.vmops [None req-fc3cf2dd-2491-4541-a245-5eefc9a5c042 tempest-TenantUsagesTestJSON-970395509 tempest-TenantUsagesTestJSON-970395509-project-member] [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance a187f73b-8fb1-42b3-9a07-24d16aa0f152 could not be found. [ 594.922906] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-fc3cf2dd-2491-4541-a245-5eefc9a5c042 tempest-TenantUsagesTestJSON-970395509 tempest-TenantUsagesTestJSON-970395509-project-member] [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 594.923131] env[62066]: INFO nova.compute.manager [None req-fc3cf2dd-2491-4541-a245-5eefc9a5c042 tempest-TenantUsagesTestJSON-970395509 tempest-TenantUsagesTestJSON-970395509-project-member] [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] Took 0.04 seconds to destroy the instance on the hypervisor. [ 594.923389] env[62066]: DEBUG oslo.service.loopingcall [None req-fc3cf2dd-2491-4541-a245-5eefc9a5c042 tempest-TenantUsagesTestJSON-970395509 tempest-TenantUsagesTestJSON-970395509-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 594.923602] env[62066]: DEBUG nova.compute.manager [-] [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 594.923696] env[62066]: DEBUG nova.network.neutron [-] [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 594.944245] env[62066]: DEBUG nova.network.neutron [-] [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 595.088601] env[62066]: DEBUG nova.compute.utils [None req-e7221104-9243-4d76-b03e-5782ae8f55d4 tempest-FloatingIPsAssociationNegativeTestJSON-961065153 tempest-FloatingIPsAssociationNegativeTestJSON-961065153-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 595.097868] env[62066]: DEBUG nova.compute.manager [None req-e7221104-9243-4d76-b03e-5782ae8f55d4 tempest-FloatingIPsAssociationNegativeTestJSON-961065153 tempest-FloatingIPsAssociationNegativeTestJSON-961065153-project-member] [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 595.098345] env[62066]: DEBUG nova.network.neutron [None req-e7221104-9243-4d76-b03e-5782ae8f55d4 tempest-FloatingIPsAssociationNegativeTestJSON-961065153 tempest-FloatingIPsAssociationNegativeTestJSON-961065153-project-member] [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 595.160452] env[62066]: DEBUG nova.policy [None req-e7221104-9243-4d76-b03e-5782ae8f55d4 tempest-FloatingIPsAssociationNegativeTestJSON-961065153 tempest-FloatingIPsAssociationNegativeTestJSON-961065153-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fb3868a8c66d4cf98b4d5488af6e91c4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6f8108178f844d5a930c4e568c65fc0c', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 595.419726] env[62066]: INFO nova.scheduler.client.report [None req-65bf05da-7d56-4074-8d13-63fde71b9619 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] Deleted allocations for instance eacdd472-5bb0-48bc-8603-9e1d08b41f55 [ 595.447930] env[62066]: DEBUG nova.network.neutron [-] [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 595.508327] env[62066]: DEBUG nova.network.neutron [None req-e7221104-9243-4d76-b03e-5782ae8f55d4 tempest-FloatingIPsAssociationNegativeTestJSON-961065153 tempest-FloatingIPsAssociationNegativeTestJSON-961065153-project-member] [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] Successfully created port: 648e8154-eb46-4556-b4ff-b8d2b55b2d10 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 595.580492] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12d2977e-3cd8-4167-96a1-d5cd575d900f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.591753] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8a7069e-4892-4352-80e0-430d92f97912 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.626115] env[62066]: DEBUG nova.compute.manager [None req-e7221104-9243-4d76-b03e-5782ae8f55d4 tempest-FloatingIPsAssociationNegativeTestJSON-961065153 tempest-FloatingIPsAssociationNegativeTestJSON-961065153-project-member] [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 595.632027] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d10d2c1-6e51-4956-9fe2-8575f920e893 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.637379] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23e843c5-118f-4f91-ad11-632103e4beb1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.654798] env[62066]: DEBUG nova.compute.provider_tree [None req-4ae0c9b4-adca-4b03-8ac7-409a4ee035f6 tempest-InstanceActionsTestJSON-1008103477 tempest-InstanceActionsTestJSON-1008103477-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 595.932641] env[62066]: DEBUG oslo_concurrency.lockutils [None req-65bf05da-7d56-4074-8d13-63fde71b9619 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] Lock "eacdd472-5bb0-48bc-8603-9e1d08b41f55" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 83.646s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 595.950042] env[62066]: INFO nova.compute.manager [-] [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] Took 1.03 seconds to deallocate network for instance. [ 595.953018] env[62066]: DEBUG nova.compute.claims [None req-fc3cf2dd-2491-4541-a245-5eefc9a5c042 tempest-TenantUsagesTestJSON-970395509 tempest-TenantUsagesTestJSON-970395509-project-member] [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] Aborting claim: {{(pid=62066) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 595.953293] env[62066]: DEBUG oslo_concurrency.lockutils [None req-fc3cf2dd-2491-4541-a245-5eefc9a5c042 tempest-TenantUsagesTestJSON-970395509 tempest-TenantUsagesTestJSON-970395509-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 596.160098] env[62066]: DEBUG nova.scheduler.client.report [None req-4ae0c9b4-adca-4b03-8ac7-409a4ee035f6 tempest-InstanceActionsTestJSON-1008103477 tempest-InstanceActionsTestJSON-1008103477-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 596.439624] env[62066]: DEBUG nova.compute.manager [None req-a9ebb4d2-ff8d-4333-a9df-86618ea2d54c tempest-ImagesOneServerTestJSON-1591819949 tempest-ImagesOneServerTestJSON-1591819949-project-member] [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 596.626728] env[62066]: DEBUG nova.compute.manager [req-596a9b65-f23f-427c-b78a-42330fe9b436 req-427e5238-3d29-4f29-94e5-68aaf4ddaf22 service nova] [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] Received event network-changed-648e8154-eb46-4556-b4ff-b8d2b55b2d10 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 596.627456] env[62066]: DEBUG nova.compute.manager [req-596a9b65-f23f-427c-b78a-42330fe9b436 req-427e5238-3d29-4f29-94e5-68aaf4ddaf22 service nova] [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] Refreshing instance network info cache due to event network-changed-648e8154-eb46-4556-b4ff-b8d2b55b2d10. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 596.627921] env[62066]: DEBUG oslo_concurrency.lockutils [req-596a9b65-f23f-427c-b78a-42330fe9b436 req-427e5238-3d29-4f29-94e5-68aaf4ddaf22 service nova] Acquiring lock "refresh_cache-eee6b9e6-2707-4a3b-925b-a7f50cdc32cb" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 596.628333] env[62066]: DEBUG oslo_concurrency.lockutils [req-596a9b65-f23f-427c-b78a-42330fe9b436 req-427e5238-3d29-4f29-94e5-68aaf4ddaf22 service nova] Acquired lock "refresh_cache-eee6b9e6-2707-4a3b-925b-a7f50cdc32cb" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 596.628776] env[62066]: DEBUG nova.network.neutron [req-596a9b65-f23f-427c-b78a-42330fe9b436 req-427e5238-3d29-4f29-94e5-68aaf4ddaf22 service nova] [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] Refreshing network info cache for port 648e8154-eb46-4556-b4ff-b8d2b55b2d10 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 596.640094] env[62066]: DEBUG nova.compute.manager [None req-e7221104-9243-4d76-b03e-5782ae8f55d4 tempest-FloatingIPsAssociationNegativeTestJSON-961065153 tempest-FloatingIPsAssociationNegativeTestJSON-961065153-project-member] [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 596.665514] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4ae0c9b4-adca-4b03-8ac7-409a4ee035f6 tempest-InstanceActionsTestJSON-1008103477 tempest-InstanceActionsTestJSON-1008103477-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.084s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 596.666130] env[62066]: ERROR nova.compute.manager [None req-4ae0c9b4-adca-4b03-8ac7-409a4ee035f6 tempest-InstanceActionsTestJSON-1008103477 tempest-InstanceActionsTestJSON-1008103477-project-member] [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 895cb3c4-534b-41ae-8230-af8fc3c155f1, please check neutron logs for more information. [ 596.666130] env[62066]: ERROR nova.compute.manager [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] Traceback (most recent call last): [ 596.666130] env[62066]: ERROR nova.compute.manager [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 596.666130] env[62066]: ERROR nova.compute.manager [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] self.driver.spawn(context, instance, image_meta, [ 596.666130] env[62066]: ERROR nova.compute.manager [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 596.666130] env[62066]: ERROR nova.compute.manager [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] self._vmops.spawn(context, instance, image_meta, injected_files, [ 596.666130] env[62066]: ERROR nova.compute.manager [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 596.666130] env[62066]: ERROR nova.compute.manager [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] vm_ref = self.build_virtual_machine(instance, [ 596.666130] env[62066]: ERROR nova.compute.manager [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 596.666130] env[62066]: ERROR nova.compute.manager [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] vif_infos = vmwarevif.get_vif_info(self._session, [ 596.666130] env[62066]: ERROR nova.compute.manager [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 596.666547] env[62066]: ERROR nova.compute.manager [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] for vif in network_info: [ 596.666547] env[62066]: ERROR nova.compute.manager [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 596.666547] env[62066]: ERROR nova.compute.manager [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] return self._sync_wrapper(fn, *args, **kwargs) [ 596.666547] env[62066]: ERROR nova.compute.manager [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 596.666547] env[62066]: ERROR nova.compute.manager [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] self.wait() [ 596.666547] env[62066]: ERROR nova.compute.manager [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 596.666547] env[62066]: ERROR nova.compute.manager [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] self[:] = self._gt.wait() [ 596.666547] env[62066]: ERROR nova.compute.manager [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 596.666547] env[62066]: ERROR nova.compute.manager [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] return self._exit_event.wait() [ 596.666547] env[62066]: ERROR nova.compute.manager [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 596.666547] env[62066]: ERROR nova.compute.manager [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] result = hub.switch() [ 596.666547] env[62066]: ERROR nova.compute.manager [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 596.666547] env[62066]: ERROR nova.compute.manager [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] return self.greenlet.switch() [ 596.667219] env[62066]: ERROR nova.compute.manager [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 596.667219] env[62066]: ERROR nova.compute.manager [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] result = function(*args, **kwargs) [ 596.667219] env[62066]: ERROR nova.compute.manager [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 596.667219] env[62066]: ERROR nova.compute.manager [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] return func(*args, **kwargs) [ 596.667219] env[62066]: ERROR nova.compute.manager [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 596.667219] env[62066]: ERROR nova.compute.manager [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] raise e [ 596.667219] env[62066]: ERROR nova.compute.manager [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 596.667219] env[62066]: ERROR nova.compute.manager [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] nwinfo = self.network_api.allocate_for_instance( [ 596.667219] env[62066]: ERROR nova.compute.manager [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 596.667219] env[62066]: ERROR nova.compute.manager [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] created_port_ids = self._update_ports_for_instance( [ 596.667219] env[62066]: ERROR nova.compute.manager [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 596.667219] env[62066]: ERROR nova.compute.manager [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] with excutils.save_and_reraise_exception(): [ 596.667219] env[62066]: ERROR nova.compute.manager [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 596.667684] env[62066]: ERROR nova.compute.manager [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] self.force_reraise() [ 596.667684] env[62066]: ERROR nova.compute.manager [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 596.667684] env[62066]: ERROR nova.compute.manager [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] raise self.value [ 596.667684] env[62066]: ERROR nova.compute.manager [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 596.667684] env[62066]: ERROR nova.compute.manager [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] updated_port = self._update_port( [ 596.667684] env[62066]: ERROR nova.compute.manager [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 596.667684] env[62066]: ERROR nova.compute.manager [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] _ensure_no_port_binding_failure(port) [ 596.667684] env[62066]: ERROR nova.compute.manager [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 596.667684] env[62066]: ERROR nova.compute.manager [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] raise exception.PortBindingFailed(port_id=port['id']) [ 596.667684] env[62066]: ERROR nova.compute.manager [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] nova.exception.PortBindingFailed: Binding failed for port 895cb3c4-534b-41ae-8230-af8fc3c155f1, please check neutron logs for more information. [ 596.667684] env[62066]: ERROR nova.compute.manager [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] [ 596.668246] env[62066]: DEBUG nova.compute.utils [None req-4ae0c9b4-adca-4b03-8ac7-409a4ee035f6 tempest-InstanceActionsTestJSON-1008103477 tempest-InstanceActionsTestJSON-1008103477-project-member] [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] Binding failed for port 895cb3c4-534b-41ae-8230-af8fc3c155f1, please check neutron logs for more information. {{(pid=62066) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 596.670734] env[62066]: DEBUG nova.virt.hardware [None req-e7221104-9243-4d76-b03e-5782ae8f55d4 tempest-FloatingIPsAssociationNegativeTestJSON-961065153 tempest-FloatingIPsAssociationNegativeTestJSON-961065153-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-23T13:40:41Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-23T13:40:23Z,direct_url=,disk_format='vmdk',id=50ff584c-3b50-4395-af07-3e66769bc9f7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='eb52f7069a374c61ae946f052007c6d9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-23T13:40:24Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 596.671390] env[62066]: DEBUG nova.virt.hardware [None req-e7221104-9243-4d76-b03e-5782ae8f55d4 tempest-FloatingIPsAssociationNegativeTestJSON-961065153 tempest-FloatingIPsAssociationNegativeTestJSON-961065153-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 596.671390] env[62066]: DEBUG nova.virt.hardware [None req-e7221104-9243-4d76-b03e-5782ae8f55d4 tempest-FloatingIPsAssociationNegativeTestJSON-961065153 tempest-FloatingIPsAssociationNegativeTestJSON-961065153-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 596.671390] env[62066]: DEBUG nova.virt.hardware [None req-e7221104-9243-4d76-b03e-5782ae8f55d4 tempest-FloatingIPsAssociationNegativeTestJSON-961065153 tempest-FloatingIPsAssociationNegativeTestJSON-961065153-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 596.671576] env[62066]: DEBUG nova.virt.hardware [None req-e7221104-9243-4d76-b03e-5782ae8f55d4 tempest-FloatingIPsAssociationNegativeTestJSON-961065153 tempest-FloatingIPsAssociationNegativeTestJSON-961065153-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 596.671576] env[62066]: DEBUG nova.virt.hardware [None req-e7221104-9243-4d76-b03e-5782ae8f55d4 tempest-FloatingIPsAssociationNegativeTestJSON-961065153 tempest-FloatingIPsAssociationNegativeTestJSON-961065153-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 596.671775] env[62066]: DEBUG nova.virt.hardware [None req-e7221104-9243-4d76-b03e-5782ae8f55d4 tempest-FloatingIPsAssociationNegativeTestJSON-961065153 tempest-FloatingIPsAssociationNegativeTestJSON-961065153-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 596.671935] env[62066]: DEBUG nova.virt.hardware [None req-e7221104-9243-4d76-b03e-5782ae8f55d4 tempest-FloatingIPsAssociationNegativeTestJSON-961065153 tempest-FloatingIPsAssociationNegativeTestJSON-961065153-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 596.672156] env[62066]: DEBUG nova.virt.hardware [None req-e7221104-9243-4d76-b03e-5782ae8f55d4 tempest-FloatingIPsAssociationNegativeTestJSON-961065153 tempest-FloatingIPsAssociationNegativeTestJSON-961065153-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 596.672301] env[62066]: DEBUG nova.virt.hardware [None req-e7221104-9243-4d76-b03e-5782ae8f55d4 tempest-FloatingIPsAssociationNegativeTestJSON-961065153 tempest-FloatingIPsAssociationNegativeTestJSON-961065153-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 596.672470] env[62066]: DEBUG nova.virt.hardware [None req-e7221104-9243-4d76-b03e-5782ae8f55d4 tempest-FloatingIPsAssociationNegativeTestJSON-961065153 tempest-FloatingIPsAssociationNegativeTestJSON-961065153-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 596.672771] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cba923da-ec24-4fd6-9695-23f810adaa32 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.570s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 596.675925] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00a6f936-f32f-40e4-8036-e42bead8a90d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.678901] env[62066]: DEBUG nova.compute.manager [None req-4ae0c9b4-adca-4b03-8ac7-409a4ee035f6 tempest-InstanceActionsTestJSON-1008103477 tempest-InstanceActionsTestJSON-1008103477-project-member] [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] Build of instance 454a790d-3cb2-4ef2-995f-97e0eb1e2469 was re-scheduled: Binding failed for port 895cb3c4-534b-41ae-8230-af8fc3c155f1, please check neutron logs for more information. {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 596.679374] env[62066]: DEBUG nova.compute.manager [None req-4ae0c9b4-adca-4b03-8ac7-409a4ee035f6 tempest-InstanceActionsTestJSON-1008103477 tempest-InstanceActionsTestJSON-1008103477-project-member] [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] Unplugging VIFs for instance {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 596.682295] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4ae0c9b4-adca-4b03-8ac7-409a4ee035f6 tempest-InstanceActionsTestJSON-1008103477 tempest-InstanceActionsTestJSON-1008103477-project-member] Acquiring lock "refresh_cache-454a790d-3cb2-4ef2-995f-97e0eb1e2469" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 596.682295] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4ae0c9b4-adca-4b03-8ac7-409a4ee035f6 tempest-InstanceActionsTestJSON-1008103477 tempest-InstanceActionsTestJSON-1008103477-project-member] Acquired lock "refresh_cache-454a790d-3cb2-4ef2-995f-97e0eb1e2469" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 596.682295] env[62066]: DEBUG nova.network.neutron [None req-4ae0c9b4-adca-4b03-8ac7-409a4ee035f6 tempest-InstanceActionsTestJSON-1008103477 tempest-InstanceActionsTestJSON-1008103477-project-member] [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 596.690752] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc2b8ecb-0618-4708-a675-8f75de27ba56 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.890686] env[62066]: ERROR nova.compute.manager [None req-e7221104-9243-4d76-b03e-5782ae8f55d4 tempest-FloatingIPsAssociationNegativeTestJSON-961065153 tempest-FloatingIPsAssociationNegativeTestJSON-961065153-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 648e8154-eb46-4556-b4ff-b8d2b55b2d10, please check neutron logs for more information. [ 596.890686] env[62066]: ERROR nova.compute.manager Traceback (most recent call last): [ 596.890686] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 596.890686] env[62066]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 596.890686] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 596.890686] env[62066]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 596.890686] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 596.890686] env[62066]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 596.890686] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 596.890686] env[62066]: ERROR nova.compute.manager self.force_reraise() [ 596.890686] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 596.890686] env[62066]: ERROR nova.compute.manager raise self.value [ 596.890686] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 596.890686] env[62066]: ERROR nova.compute.manager updated_port = self._update_port( [ 596.890686] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 596.890686] env[62066]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 596.891150] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 596.891150] env[62066]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 596.891150] env[62066]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 648e8154-eb46-4556-b4ff-b8d2b55b2d10, please check neutron logs for more information. [ 596.891150] env[62066]: ERROR nova.compute.manager [ 596.891150] env[62066]: Traceback (most recent call last): [ 596.891150] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 596.891150] env[62066]: listener.cb(fileno) [ 596.891150] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 596.891150] env[62066]: result = function(*args, **kwargs) [ 596.891150] env[62066]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 596.891150] env[62066]: return func(*args, **kwargs) [ 596.891150] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 596.891150] env[62066]: raise e [ 596.891150] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 596.891150] env[62066]: nwinfo = self.network_api.allocate_for_instance( [ 596.891150] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 596.891150] env[62066]: created_port_ids = self._update_ports_for_instance( [ 596.891150] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 596.891150] env[62066]: with excutils.save_and_reraise_exception(): [ 596.891150] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 596.891150] env[62066]: self.force_reraise() [ 596.891150] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 596.891150] env[62066]: raise self.value [ 596.891150] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 596.891150] env[62066]: updated_port = self._update_port( [ 596.891150] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 596.891150] env[62066]: _ensure_no_port_binding_failure(port) [ 596.891150] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 596.891150] env[62066]: raise exception.PortBindingFailed(port_id=port['id']) [ 596.891987] env[62066]: nova.exception.PortBindingFailed: Binding failed for port 648e8154-eb46-4556-b4ff-b8d2b55b2d10, please check neutron logs for more information. [ 596.891987] env[62066]: Removing descriptor: 20 [ 596.891987] env[62066]: ERROR nova.compute.manager [None req-e7221104-9243-4d76-b03e-5782ae8f55d4 tempest-FloatingIPsAssociationNegativeTestJSON-961065153 tempest-FloatingIPsAssociationNegativeTestJSON-961065153-project-member] [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 648e8154-eb46-4556-b4ff-b8d2b55b2d10, please check neutron logs for more information. [ 596.891987] env[62066]: ERROR nova.compute.manager [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] Traceback (most recent call last): [ 596.891987] env[62066]: ERROR nova.compute.manager [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 596.891987] env[62066]: ERROR nova.compute.manager [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] yield resources [ 596.891987] env[62066]: ERROR nova.compute.manager [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 596.891987] env[62066]: ERROR nova.compute.manager [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] self.driver.spawn(context, instance, image_meta, [ 596.891987] env[62066]: ERROR nova.compute.manager [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 596.891987] env[62066]: ERROR nova.compute.manager [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] self._vmops.spawn(context, instance, image_meta, injected_files, [ 596.891987] env[62066]: ERROR nova.compute.manager [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 596.891987] env[62066]: ERROR nova.compute.manager [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] vm_ref = self.build_virtual_machine(instance, [ 596.892335] env[62066]: ERROR nova.compute.manager [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 596.892335] env[62066]: ERROR nova.compute.manager [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] vif_infos = vmwarevif.get_vif_info(self._session, [ 596.892335] env[62066]: ERROR nova.compute.manager [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 596.892335] env[62066]: ERROR nova.compute.manager [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] for vif in network_info: [ 596.892335] env[62066]: ERROR nova.compute.manager [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 596.892335] env[62066]: ERROR nova.compute.manager [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] return self._sync_wrapper(fn, *args, **kwargs) [ 596.892335] env[62066]: ERROR nova.compute.manager [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 596.892335] env[62066]: ERROR nova.compute.manager [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] self.wait() [ 596.892335] env[62066]: ERROR nova.compute.manager [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 596.892335] env[62066]: ERROR nova.compute.manager [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] self[:] = self._gt.wait() [ 596.892335] env[62066]: ERROR nova.compute.manager [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 596.892335] env[62066]: ERROR nova.compute.manager [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] return self._exit_event.wait() [ 596.892335] env[62066]: ERROR nova.compute.manager [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 596.892669] env[62066]: ERROR nova.compute.manager [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] result = hub.switch() [ 596.892669] env[62066]: ERROR nova.compute.manager [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 596.892669] env[62066]: ERROR nova.compute.manager [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] return self.greenlet.switch() [ 596.892669] env[62066]: ERROR nova.compute.manager [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 596.892669] env[62066]: ERROR nova.compute.manager [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] result = function(*args, **kwargs) [ 596.892669] env[62066]: ERROR nova.compute.manager [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 596.892669] env[62066]: ERROR nova.compute.manager [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] return func(*args, **kwargs) [ 596.892669] env[62066]: ERROR nova.compute.manager [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 596.892669] env[62066]: ERROR nova.compute.manager [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] raise e [ 596.892669] env[62066]: ERROR nova.compute.manager [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 596.892669] env[62066]: ERROR nova.compute.manager [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] nwinfo = self.network_api.allocate_for_instance( [ 596.892669] env[62066]: ERROR nova.compute.manager [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 596.892669] env[62066]: ERROR nova.compute.manager [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] created_port_ids = self._update_ports_for_instance( [ 596.893074] env[62066]: ERROR nova.compute.manager [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 596.893074] env[62066]: ERROR nova.compute.manager [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] with excutils.save_and_reraise_exception(): [ 596.893074] env[62066]: ERROR nova.compute.manager [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 596.893074] env[62066]: ERROR nova.compute.manager [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] self.force_reraise() [ 596.893074] env[62066]: ERROR nova.compute.manager [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 596.893074] env[62066]: ERROR nova.compute.manager [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] raise self.value [ 596.893074] env[62066]: ERROR nova.compute.manager [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 596.893074] env[62066]: ERROR nova.compute.manager [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] updated_port = self._update_port( [ 596.893074] env[62066]: ERROR nova.compute.manager [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 596.893074] env[62066]: ERROR nova.compute.manager [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] _ensure_no_port_binding_failure(port) [ 596.893074] env[62066]: ERROR nova.compute.manager [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 596.893074] env[62066]: ERROR nova.compute.manager [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] raise exception.PortBindingFailed(port_id=port['id']) [ 596.893402] env[62066]: ERROR nova.compute.manager [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] nova.exception.PortBindingFailed: Binding failed for port 648e8154-eb46-4556-b4ff-b8d2b55b2d10, please check neutron logs for more information. [ 596.893402] env[62066]: ERROR nova.compute.manager [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] [ 596.893402] env[62066]: INFO nova.compute.manager [None req-e7221104-9243-4d76-b03e-5782ae8f55d4 tempest-FloatingIPsAssociationNegativeTestJSON-961065153 tempest-FloatingIPsAssociationNegativeTestJSON-961065153-project-member] [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] Terminating instance [ 596.894293] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e7221104-9243-4d76-b03e-5782ae8f55d4 tempest-FloatingIPsAssociationNegativeTestJSON-961065153 tempest-FloatingIPsAssociationNegativeTestJSON-961065153-project-member] Acquiring lock "refresh_cache-eee6b9e6-2707-4a3b-925b-a7f50cdc32cb" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 596.962601] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a9ebb4d2-ff8d-4333-a9df-86618ea2d54c tempest-ImagesOneServerTestJSON-1591819949 tempest-ImagesOneServerTestJSON-1591819949-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 597.151796] env[62066]: DEBUG nova.network.neutron [req-596a9b65-f23f-427c-b78a-42330fe9b436 req-427e5238-3d29-4f29-94e5-68aaf4ddaf22 service nova] [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 597.208142] env[62066]: DEBUG nova.network.neutron [None req-4ae0c9b4-adca-4b03-8ac7-409a4ee035f6 tempest-InstanceActionsTestJSON-1008103477 tempest-InstanceActionsTestJSON-1008103477-project-member] [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 597.282504] env[62066]: DEBUG nova.network.neutron [req-596a9b65-f23f-427c-b78a-42330fe9b436 req-427e5238-3d29-4f29-94e5-68aaf4ddaf22 service nova] [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 597.301891] env[62066]: DEBUG nova.network.neutron [None req-4ae0c9b4-adca-4b03-8ac7-409a4ee035f6 tempest-InstanceActionsTestJSON-1008103477 tempest-InstanceActionsTestJSON-1008103477-project-member] [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 597.643094] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0deef406-4e47-4e2e-b393-bb2b63603daa {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.651412] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bae0bf3-a2ed-44b4-a362-2f8843edd1e4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.684193] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c6e2f95-999c-4f54-9db7-abb80c66d60b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.691755] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73af3a1a-7ca4-4947-83af-63ed24eddcab {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.348821] env[62066]: DEBUG oslo_concurrency.lockutils [req-596a9b65-f23f-427c-b78a-42330fe9b436 req-427e5238-3d29-4f29-94e5-68aaf4ddaf22 service nova] Releasing lock "refresh_cache-eee6b9e6-2707-4a3b-925b-a7f50cdc32cb" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 598.349355] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4ae0c9b4-adca-4b03-8ac7-409a4ee035f6 tempest-InstanceActionsTestJSON-1008103477 tempest-InstanceActionsTestJSON-1008103477-project-member] Releasing lock "refresh_cache-454a790d-3cb2-4ef2-995f-97e0eb1e2469" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 598.349570] env[62066]: DEBUG nova.compute.manager [None req-4ae0c9b4-adca-4b03-8ac7-409a4ee035f6 tempest-InstanceActionsTestJSON-1008103477 tempest-InstanceActionsTestJSON-1008103477-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 598.349735] env[62066]: DEBUG nova.compute.manager [None req-4ae0c9b4-adca-4b03-8ac7-409a4ee035f6 tempest-InstanceActionsTestJSON-1008103477 tempest-InstanceActionsTestJSON-1008103477-project-member] [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 598.349914] env[62066]: DEBUG nova.network.neutron [None req-4ae0c9b4-adca-4b03-8ac7-409a4ee035f6 tempest-InstanceActionsTestJSON-1008103477 tempest-InstanceActionsTestJSON-1008103477-project-member] [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 598.352117] env[62066]: DEBUG nova.compute.provider_tree [None req-cba923da-ec24-4fd6-9695-23f810adaa32 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 598.353475] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e7221104-9243-4d76-b03e-5782ae8f55d4 tempest-FloatingIPsAssociationNegativeTestJSON-961065153 tempest-FloatingIPsAssociationNegativeTestJSON-961065153-project-member] Acquired lock "refresh_cache-eee6b9e6-2707-4a3b-925b-a7f50cdc32cb" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 598.353650] env[62066]: DEBUG nova.network.neutron [None req-e7221104-9243-4d76-b03e-5782ae8f55d4 tempest-FloatingIPsAssociationNegativeTestJSON-961065153 tempest-FloatingIPsAssociationNegativeTestJSON-961065153-project-member] [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 598.382515] env[62066]: DEBUG nova.network.neutron [None req-4ae0c9b4-adca-4b03-8ac7-409a4ee035f6 tempest-InstanceActionsTestJSON-1008103477 tempest-InstanceActionsTestJSON-1008103477-project-member] [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 598.680542] env[62066]: DEBUG nova.compute.manager [req-fc87ec67-12d6-4ddf-b4a8-90f131fbe462 req-097422b4-2471-4b64-9e66-eb21a38ae1fa service nova] [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] Received event network-vif-deleted-648e8154-eb46-4556-b4ff-b8d2b55b2d10 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 598.859077] env[62066]: DEBUG nova.scheduler.client.report [None req-cba923da-ec24-4fd6-9695-23f810adaa32 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 598.885606] env[62066]: DEBUG nova.network.neutron [None req-4ae0c9b4-adca-4b03-8ac7-409a4ee035f6 tempest-InstanceActionsTestJSON-1008103477 tempest-InstanceActionsTestJSON-1008103477-project-member] [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 598.890068] env[62066]: DEBUG nova.network.neutron [None req-e7221104-9243-4d76-b03e-5782ae8f55d4 tempest-FloatingIPsAssociationNegativeTestJSON-961065153 tempest-FloatingIPsAssociationNegativeTestJSON-961065153-project-member] [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 599.068854] env[62066]: DEBUG nova.network.neutron [None req-e7221104-9243-4d76-b03e-5782ae8f55d4 tempest-FloatingIPsAssociationNegativeTestJSON-961065153 tempest-FloatingIPsAssociationNegativeTestJSON-961065153-project-member] [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 599.368994] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cba923da-ec24-4fd6-9695-23f810adaa32 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.696s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 599.370053] env[62066]: ERROR nova.compute.manager [None req-cba923da-ec24-4fd6-9695-23f810adaa32 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 19cd4055-4906-4fce-a9a8-b6d1dffd27f4, please check neutron logs for more information. [ 599.370053] env[62066]: ERROR nova.compute.manager [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] Traceback (most recent call last): [ 599.370053] env[62066]: ERROR nova.compute.manager [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 599.370053] env[62066]: ERROR nova.compute.manager [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] self.driver.spawn(context, instance, image_meta, [ 599.370053] env[62066]: ERROR nova.compute.manager [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 599.370053] env[62066]: ERROR nova.compute.manager [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] self._vmops.spawn(context, instance, image_meta, injected_files, [ 599.370053] env[62066]: ERROR nova.compute.manager [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 599.370053] env[62066]: ERROR nova.compute.manager [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] vm_ref = self.build_virtual_machine(instance, [ 599.370053] env[62066]: ERROR nova.compute.manager [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 599.370053] env[62066]: ERROR nova.compute.manager [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] vif_infos = vmwarevif.get_vif_info(self._session, [ 599.370053] env[62066]: ERROR nova.compute.manager [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 599.370389] env[62066]: ERROR nova.compute.manager [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] for vif in network_info: [ 599.370389] env[62066]: ERROR nova.compute.manager [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 599.370389] env[62066]: ERROR nova.compute.manager [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] return self._sync_wrapper(fn, *args, **kwargs) [ 599.370389] env[62066]: ERROR nova.compute.manager [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 599.370389] env[62066]: ERROR nova.compute.manager [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] self.wait() [ 599.370389] env[62066]: ERROR nova.compute.manager [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 599.370389] env[62066]: ERROR nova.compute.manager [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] self[:] = self._gt.wait() [ 599.370389] env[62066]: ERROR nova.compute.manager [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 599.370389] env[62066]: ERROR nova.compute.manager [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] return self._exit_event.wait() [ 599.370389] env[62066]: ERROR nova.compute.manager [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 599.370389] env[62066]: ERROR nova.compute.manager [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] result = hub.switch() [ 599.370389] env[62066]: ERROR nova.compute.manager [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 599.370389] env[62066]: ERROR nova.compute.manager [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] return self.greenlet.switch() [ 599.370705] env[62066]: ERROR nova.compute.manager [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 599.370705] env[62066]: ERROR nova.compute.manager [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] result = function(*args, **kwargs) [ 599.370705] env[62066]: ERROR nova.compute.manager [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 599.370705] env[62066]: ERROR nova.compute.manager [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] return func(*args, **kwargs) [ 599.370705] env[62066]: ERROR nova.compute.manager [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 599.370705] env[62066]: ERROR nova.compute.manager [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] raise e [ 599.370705] env[62066]: ERROR nova.compute.manager [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 599.370705] env[62066]: ERROR nova.compute.manager [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] nwinfo = self.network_api.allocate_for_instance( [ 599.370705] env[62066]: ERROR nova.compute.manager [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 599.370705] env[62066]: ERROR nova.compute.manager [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] created_port_ids = self._update_ports_for_instance( [ 599.370705] env[62066]: ERROR nova.compute.manager [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 599.370705] env[62066]: ERROR nova.compute.manager [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] with excutils.save_and_reraise_exception(): [ 599.370705] env[62066]: ERROR nova.compute.manager [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 599.371016] env[62066]: ERROR nova.compute.manager [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] self.force_reraise() [ 599.371016] env[62066]: ERROR nova.compute.manager [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 599.371016] env[62066]: ERROR nova.compute.manager [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] raise self.value [ 599.371016] env[62066]: ERROR nova.compute.manager [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 599.371016] env[62066]: ERROR nova.compute.manager [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] updated_port = self._update_port( [ 599.371016] env[62066]: ERROR nova.compute.manager [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 599.371016] env[62066]: ERROR nova.compute.manager [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] _ensure_no_port_binding_failure(port) [ 599.371016] env[62066]: ERROR nova.compute.manager [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 599.371016] env[62066]: ERROR nova.compute.manager [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] raise exception.PortBindingFailed(port_id=port['id']) [ 599.371016] env[62066]: ERROR nova.compute.manager [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] nova.exception.PortBindingFailed: Binding failed for port 19cd4055-4906-4fce-a9a8-b6d1dffd27f4, please check neutron logs for more information. [ 599.371016] env[62066]: ERROR nova.compute.manager [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] [ 599.371299] env[62066]: DEBUG nova.compute.utils [None req-cba923da-ec24-4fd6-9695-23f810adaa32 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] Binding failed for port 19cd4055-4906-4fce-a9a8-b6d1dffd27f4, please check neutron logs for more information. {{(pid=62066) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 599.372349] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.651s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 599.376029] env[62066]: INFO nova.compute.claims [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 599.380906] env[62066]: DEBUG nova.compute.manager [None req-cba923da-ec24-4fd6-9695-23f810adaa32 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] Build of instance 0a08bc46-0eea-4802-b0be-eb24f3507f73 was re-scheduled: Binding failed for port 19cd4055-4906-4fce-a9a8-b6d1dffd27f4, please check neutron logs for more information. {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 599.381380] env[62066]: DEBUG nova.compute.manager [None req-cba923da-ec24-4fd6-9695-23f810adaa32 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] Unplugging VIFs for instance {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 599.381608] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cba923da-ec24-4fd6-9695-23f810adaa32 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] Acquiring lock "refresh_cache-0a08bc46-0eea-4802-b0be-eb24f3507f73" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 599.381750] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cba923da-ec24-4fd6-9695-23f810adaa32 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] Acquired lock "refresh_cache-0a08bc46-0eea-4802-b0be-eb24f3507f73" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 599.381904] env[62066]: DEBUG nova.network.neutron [None req-cba923da-ec24-4fd6-9695-23f810adaa32 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 599.392063] env[62066]: INFO nova.compute.manager [None req-4ae0c9b4-adca-4b03-8ac7-409a4ee035f6 tempest-InstanceActionsTestJSON-1008103477 tempest-InstanceActionsTestJSON-1008103477-project-member] [instance: 454a790d-3cb2-4ef2-995f-97e0eb1e2469] Took 1.04 seconds to deallocate network for instance. [ 599.571888] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e7221104-9243-4d76-b03e-5782ae8f55d4 tempest-FloatingIPsAssociationNegativeTestJSON-961065153 tempest-FloatingIPsAssociationNegativeTestJSON-961065153-project-member] Releasing lock "refresh_cache-eee6b9e6-2707-4a3b-925b-a7f50cdc32cb" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 599.573580] env[62066]: DEBUG nova.compute.manager [None req-e7221104-9243-4d76-b03e-5782ae8f55d4 tempest-FloatingIPsAssociationNegativeTestJSON-961065153 tempest-FloatingIPsAssociationNegativeTestJSON-961065153-project-member] [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 599.573580] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-e7221104-9243-4d76-b03e-5782ae8f55d4 tempest-FloatingIPsAssociationNegativeTestJSON-961065153 tempest-FloatingIPsAssociationNegativeTestJSON-961065153-project-member] [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 599.573580] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c5c7c99d-60f0-434d-b2b9-3b774c8c2d41 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.585882] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-409969a2-8f77-406a-a857-f9f983837c9a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.606763] env[62066]: WARNING nova.virt.vmwareapi.vmops [None req-e7221104-9243-4d76-b03e-5782ae8f55d4 tempest-FloatingIPsAssociationNegativeTestJSON-961065153 tempest-FloatingIPsAssociationNegativeTestJSON-961065153-project-member] [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance eee6b9e6-2707-4a3b-925b-a7f50cdc32cb could not be found. [ 599.607034] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-e7221104-9243-4d76-b03e-5782ae8f55d4 tempest-FloatingIPsAssociationNegativeTestJSON-961065153 tempest-FloatingIPsAssociationNegativeTestJSON-961065153-project-member] [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 599.607262] env[62066]: INFO nova.compute.manager [None req-e7221104-9243-4d76-b03e-5782ae8f55d4 tempest-FloatingIPsAssociationNegativeTestJSON-961065153 tempest-FloatingIPsAssociationNegativeTestJSON-961065153-project-member] [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] Took 0.03 seconds to destroy the instance on the hypervisor. [ 599.607655] env[62066]: DEBUG oslo.service.loopingcall [None req-e7221104-9243-4d76-b03e-5782ae8f55d4 tempest-FloatingIPsAssociationNegativeTestJSON-961065153 tempest-FloatingIPsAssociationNegativeTestJSON-961065153-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 599.607887] env[62066]: DEBUG nova.compute.manager [-] [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 599.607981] env[62066]: DEBUG nova.network.neutron [-] [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 599.630017] env[62066]: DEBUG nova.network.neutron [-] [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 599.912083] env[62066]: DEBUG nova.network.neutron [None req-cba923da-ec24-4fd6-9695-23f810adaa32 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 600.020423] env[62066]: DEBUG nova.network.neutron [None req-cba923da-ec24-4fd6-9695-23f810adaa32 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 600.131417] env[62066]: DEBUG nova.network.neutron [-] [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 600.432533] env[62066]: INFO nova.scheduler.client.report [None req-4ae0c9b4-adca-4b03-8ac7-409a4ee035f6 tempest-InstanceActionsTestJSON-1008103477 tempest-InstanceActionsTestJSON-1008103477-project-member] Deleted allocations for instance 454a790d-3cb2-4ef2-995f-97e0eb1e2469 [ 600.528861] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cba923da-ec24-4fd6-9695-23f810adaa32 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] Releasing lock "refresh_cache-0a08bc46-0eea-4802-b0be-eb24f3507f73" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 600.529932] env[62066]: DEBUG nova.compute.manager [None req-cba923da-ec24-4fd6-9695-23f810adaa32 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 600.530694] env[62066]: DEBUG nova.compute.manager [None req-cba923da-ec24-4fd6-9695-23f810adaa32 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 600.531340] env[62066]: DEBUG nova.network.neutron [None req-cba923da-ec24-4fd6-9695-23f810adaa32 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 600.556846] env[62066]: DEBUG nova.network.neutron [None req-cba923da-ec24-4fd6-9695-23f810adaa32 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 600.635519] env[62066]: INFO nova.compute.manager [-] [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] Took 1.03 seconds to deallocate network for instance. [ 600.640075] env[62066]: DEBUG nova.compute.claims [None req-e7221104-9243-4d76-b03e-5782ae8f55d4 tempest-FloatingIPsAssociationNegativeTestJSON-961065153 tempest-FloatingIPsAssociationNegativeTestJSON-961065153-project-member] [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] Aborting claim: {{(pid=62066) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 600.640270] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e7221104-9243-4d76-b03e-5782ae8f55d4 tempest-FloatingIPsAssociationNegativeTestJSON-961065153 tempest-FloatingIPsAssociationNegativeTestJSON-961065153-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 600.898915] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-caa978d2-a511-4d96-b567-bf3a1857788b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.907611] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93ad7c0b-0780-4155-9eb5-d8d5001998bc {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.936814] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fb1d1f7-ff65-433e-9486-d2d48eef7e9d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.942958] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ddc6759-b0f9-4bb5-a935-f6c9dc1b9db7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.948892] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4ae0c9b4-adca-4b03-8ac7-409a4ee035f6 tempest-InstanceActionsTestJSON-1008103477 tempest-InstanceActionsTestJSON-1008103477-project-member] Lock "454a790d-3cb2-4ef2-995f-97e0eb1e2469" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 86.964s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 600.958830] env[62066]: DEBUG nova.compute.provider_tree [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 601.062015] env[62066]: DEBUG nova.network.neutron [None req-cba923da-ec24-4fd6-9695-23f810adaa32 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 601.183790] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7d1d419b-6e61-4f85-9698-fe0c6e4f3988 tempest-AttachInterfacesV270Test-236309384 tempest-AttachInterfacesV270Test-236309384-project-member] Acquiring lock "c908a3ed-88a6-4e79-8c27-2e16bf2bdca4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 601.184096] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7d1d419b-6e61-4f85-9698-fe0c6e4f3988 tempest-AttachInterfacesV270Test-236309384 tempest-AttachInterfacesV270Test-236309384-project-member] Lock "c908a3ed-88a6-4e79-8c27-2e16bf2bdca4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.002s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 601.234787] env[62066]: DEBUG oslo_service.periodic_task [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 601.235010] env[62066]: DEBUG oslo_service.periodic_task [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 601.462312] env[62066]: DEBUG nova.compute.manager [None req-9371bde4-46e3-4272-a1fe-5125bccbd2c7 tempest-ServersTestJSON-680983284 tempest-ServersTestJSON-680983284-project-member] [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 601.466681] env[62066]: DEBUG nova.scheduler.client.report [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 601.565404] env[62066]: INFO nova.compute.manager [None req-cba923da-ec24-4fd6-9695-23f810adaa32 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] [instance: 0a08bc46-0eea-4802-b0be-eb24f3507f73] Took 1.03 seconds to deallocate network for instance. [ 601.740747] env[62066]: DEBUG oslo_service.periodic_task [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 601.741021] env[62066]: DEBUG nova.compute.manager [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Starting heal instance info cache {{(pid=62066) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 601.741106] env[62066]: DEBUG nova.compute.manager [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Rebuilding the list of instances to heal {{(pid=62066) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 601.973829] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.602s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 601.974733] env[62066]: DEBUG nova.compute.manager [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 601.979829] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4fa8ac7b-ee00-4ef3-97aa-07a2c2ae6378 tempest-ServerExternalEventsTest-41922728 tempest-ServerExternalEventsTest-41922728-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.501s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 602.001632] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9371bde4-46e3-4272-a1fe-5125bccbd2c7 tempest-ServersTestJSON-680983284 tempest-ServersTestJSON-680983284-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 602.247885] env[62066]: DEBUG nova.compute.manager [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] [instance: 39cd204a-49c2-4309-9030-555056ab125e] Skipping network cache update for instance because it is Building. {{(pid=62066) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 602.248064] env[62066]: DEBUG nova.compute.manager [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] [instance: 194588b4-fe40-4286-8036-874a7c410327] Skipping network cache update for instance because it is Building. {{(pid=62066) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 602.248197] env[62066]: DEBUG nova.compute.manager [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] Skipping network cache update for instance because it is Building. {{(pid=62066) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 602.248333] env[62066]: DEBUG nova.compute.manager [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] Skipping network cache update for instance because it is Building. {{(pid=62066) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 602.248480] env[62066]: DEBUG nova.compute.manager [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] Skipping network cache update for instance because it is Building. {{(pid=62066) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 602.248696] env[62066]: DEBUG nova.compute.manager [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Didn't find any instances for network info cache update. {{(pid=62066) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 602.248913] env[62066]: DEBUG oslo_service.periodic_task [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 602.249106] env[62066]: DEBUG oslo_service.periodic_task [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 602.249327] env[62066]: DEBUG oslo_service.periodic_task [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 602.249454] env[62066]: DEBUG oslo_service.periodic_task [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 602.249606] env[62066]: DEBUG oslo_service.periodic_task [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 602.249772] env[62066]: DEBUG oslo_service.periodic_task [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 602.249911] env[62066]: DEBUG nova.compute.manager [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62066) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 602.250076] env[62066]: DEBUG oslo_service.periodic_task [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Running periodic task ComputeManager.update_available_resource {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 602.480661] env[62066]: DEBUG nova.compute.utils [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 602.480874] env[62066]: DEBUG nova.compute.manager [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 602.481116] env[62066]: DEBUG nova.network.neutron [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 602.526601] env[62066]: DEBUG nova.policy [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '236eede7716d4664a75f5df76aa403ce', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7664e11610664ce5a85a2ad8758986f8', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 602.600865] env[62066]: INFO nova.scheduler.client.report [None req-cba923da-ec24-4fd6-9695-23f810adaa32 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] Deleted allocations for instance 0a08bc46-0eea-4802-b0be-eb24f3507f73 [ 602.754310] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 602.838100] env[62066]: DEBUG nova.network.neutron [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] Successfully created port: afd58cf2-370e-45b9-bc5a-7bea27a39c71 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 602.928245] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18298581-e6ff-4951-ae0d-d11fb1cbb301 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.937199] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-779ec79d-55fb-4277-92f7-ff0982b5d96a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.974080] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a333073a-a4be-48ea-9013-d86621226856 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.982507] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69ba05bf-a5ee-43ac-b47c-5fb8e303ea7e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.986482] env[62066]: DEBUG nova.compute.manager [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 602.999697] env[62066]: DEBUG nova.compute.provider_tree [None req-4fa8ac7b-ee00-4ef3-97aa-07a2c2ae6378 tempest-ServerExternalEventsTest-41922728 tempest-ServerExternalEventsTest-41922728-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 603.109369] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cba923da-ec24-4fd6-9695-23f810adaa32 tempest-ListServerFiltersTestJSON-2134928980 tempest-ListServerFiltersTestJSON-2134928980-project-member] Lock "0a08bc46-0eea-4802-b0be-eb24f3507f73" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 88.596s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 603.505908] env[62066]: DEBUG nova.scheduler.client.report [None req-4fa8ac7b-ee00-4ef3-97aa-07a2c2ae6378 tempest-ServerExternalEventsTest-41922728 tempest-ServerExternalEventsTest-41922728-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 603.612690] env[62066]: DEBUG nova.compute.manager [None req-7abf52e5-ee7c-4f92-a757-ffa8ca735a57 tempest-ServerActionsTestJSON-164481335 tempest-ServerActionsTestJSON-164481335-project-member] [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 603.770698] env[62066]: ERROR nova.compute.manager [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port afd58cf2-370e-45b9-bc5a-7bea27a39c71, please check neutron logs for more information. [ 603.770698] env[62066]: ERROR nova.compute.manager Traceback (most recent call last): [ 603.770698] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 603.770698] env[62066]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 603.770698] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 603.770698] env[62066]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 603.770698] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 603.770698] env[62066]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 603.770698] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 603.770698] env[62066]: ERROR nova.compute.manager self.force_reraise() [ 603.770698] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 603.770698] env[62066]: ERROR nova.compute.manager raise self.value [ 603.770698] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 603.770698] env[62066]: ERROR nova.compute.manager updated_port = self._update_port( [ 603.770698] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 603.770698] env[62066]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 603.771191] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 603.771191] env[62066]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 603.771191] env[62066]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port afd58cf2-370e-45b9-bc5a-7bea27a39c71, please check neutron logs for more information. [ 603.771191] env[62066]: ERROR nova.compute.manager [ 603.771191] env[62066]: Traceback (most recent call last): [ 603.771191] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 603.771191] env[62066]: listener.cb(fileno) [ 603.771191] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 603.771191] env[62066]: result = function(*args, **kwargs) [ 603.771191] env[62066]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 603.771191] env[62066]: return func(*args, **kwargs) [ 603.771191] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 603.771191] env[62066]: raise e [ 603.771191] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 603.771191] env[62066]: nwinfo = self.network_api.allocate_for_instance( [ 603.771191] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 603.771191] env[62066]: created_port_ids = self._update_ports_for_instance( [ 603.771191] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 603.771191] env[62066]: with excutils.save_and_reraise_exception(): [ 603.771191] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 603.771191] env[62066]: self.force_reraise() [ 603.771191] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 603.771191] env[62066]: raise self.value [ 603.771191] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 603.771191] env[62066]: updated_port = self._update_port( [ 603.771191] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 603.771191] env[62066]: _ensure_no_port_binding_failure(port) [ 603.771191] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 603.771191] env[62066]: raise exception.PortBindingFailed(port_id=port['id']) [ 603.771989] env[62066]: nova.exception.PortBindingFailed: Binding failed for port afd58cf2-370e-45b9-bc5a-7bea27a39c71, please check neutron logs for more information. [ 603.771989] env[62066]: Removing descriptor: 15 [ 603.957695] env[62066]: DEBUG nova.compute.manager [req-7d9a7094-5f3a-4cc5-b664-3e39a43d3390 req-21ee7ac9-2613-444c-ac93-6beb183a71f1 service nova] [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] Received event network-changed-afd58cf2-370e-45b9-bc5a-7bea27a39c71 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 603.957999] env[62066]: DEBUG nova.compute.manager [req-7d9a7094-5f3a-4cc5-b664-3e39a43d3390 req-21ee7ac9-2613-444c-ac93-6beb183a71f1 service nova] [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] Refreshing instance network info cache due to event network-changed-afd58cf2-370e-45b9-bc5a-7bea27a39c71. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 603.958607] env[62066]: DEBUG oslo_concurrency.lockutils [req-7d9a7094-5f3a-4cc5-b664-3e39a43d3390 req-21ee7ac9-2613-444c-ac93-6beb183a71f1 service nova] Acquiring lock "refresh_cache-5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 603.958783] env[62066]: DEBUG oslo_concurrency.lockutils [req-7d9a7094-5f3a-4cc5-b664-3e39a43d3390 req-21ee7ac9-2613-444c-ac93-6beb183a71f1 service nova] Acquired lock "refresh_cache-5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 603.958952] env[62066]: DEBUG nova.network.neutron [req-7d9a7094-5f3a-4cc5-b664-3e39a43d3390 req-21ee7ac9-2613-444c-ac93-6beb183a71f1 service nova] [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] Refreshing network info cache for port afd58cf2-370e-45b9-bc5a-7bea27a39c71 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 604.012049] env[62066]: DEBUG nova.compute.manager [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 604.014590] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4fa8ac7b-ee00-4ef3-97aa-07a2c2ae6378 tempest-ServerExternalEventsTest-41922728 tempest-ServerExternalEventsTest-41922728-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.035s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 604.015282] env[62066]: ERROR nova.compute.manager [None req-4fa8ac7b-ee00-4ef3-97aa-07a2c2ae6378 tempest-ServerExternalEventsTest-41922728 tempest-ServerExternalEventsTest-41922728-project-member] [instance: 39cd204a-49c2-4309-9030-555056ab125e] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 00047463-06a3-4ffe-830b-2d4ef9e64de6, please check neutron logs for more information. [ 604.015282] env[62066]: ERROR nova.compute.manager [instance: 39cd204a-49c2-4309-9030-555056ab125e] Traceback (most recent call last): [ 604.015282] env[62066]: ERROR nova.compute.manager [instance: 39cd204a-49c2-4309-9030-555056ab125e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 604.015282] env[62066]: ERROR nova.compute.manager [instance: 39cd204a-49c2-4309-9030-555056ab125e] self.driver.spawn(context, instance, image_meta, [ 604.015282] env[62066]: ERROR nova.compute.manager [instance: 39cd204a-49c2-4309-9030-555056ab125e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 604.015282] env[62066]: ERROR nova.compute.manager [instance: 39cd204a-49c2-4309-9030-555056ab125e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 604.015282] env[62066]: ERROR nova.compute.manager [instance: 39cd204a-49c2-4309-9030-555056ab125e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 604.015282] env[62066]: ERROR nova.compute.manager [instance: 39cd204a-49c2-4309-9030-555056ab125e] vm_ref = self.build_virtual_machine(instance, [ 604.015282] env[62066]: ERROR nova.compute.manager [instance: 39cd204a-49c2-4309-9030-555056ab125e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 604.015282] env[62066]: ERROR nova.compute.manager [instance: 39cd204a-49c2-4309-9030-555056ab125e] vif_infos = vmwarevif.get_vif_info(self._session, [ 604.015282] env[62066]: ERROR nova.compute.manager [instance: 39cd204a-49c2-4309-9030-555056ab125e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 604.015634] env[62066]: ERROR nova.compute.manager [instance: 39cd204a-49c2-4309-9030-555056ab125e] for vif in network_info: [ 604.015634] env[62066]: ERROR nova.compute.manager [instance: 39cd204a-49c2-4309-9030-555056ab125e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 604.015634] env[62066]: ERROR nova.compute.manager [instance: 39cd204a-49c2-4309-9030-555056ab125e] return self._sync_wrapper(fn, *args, **kwargs) [ 604.015634] env[62066]: ERROR nova.compute.manager [instance: 39cd204a-49c2-4309-9030-555056ab125e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 604.015634] env[62066]: ERROR nova.compute.manager [instance: 39cd204a-49c2-4309-9030-555056ab125e] self.wait() [ 604.015634] env[62066]: ERROR nova.compute.manager [instance: 39cd204a-49c2-4309-9030-555056ab125e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 604.015634] env[62066]: ERROR nova.compute.manager [instance: 39cd204a-49c2-4309-9030-555056ab125e] self[:] = self._gt.wait() [ 604.015634] env[62066]: ERROR nova.compute.manager [instance: 39cd204a-49c2-4309-9030-555056ab125e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 604.015634] env[62066]: ERROR nova.compute.manager [instance: 39cd204a-49c2-4309-9030-555056ab125e] return self._exit_event.wait() [ 604.015634] env[62066]: ERROR nova.compute.manager [instance: 39cd204a-49c2-4309-9030-555056ab125e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 604.015634] env[62066]: ERROR nova.compute.manager [instance: 39cd204a-49c2-4309-9030-555056ab125e] result = hub.switch() [ 604.015634] env[62066]: ERROR nova.compute.manager [instance: 39cd204a-49c2-4309-9030-555056ab125e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 604.015634] env[62066]: ERROR nova.compute.manager [instance: 39cd204a-49c2-4309-9030-555056ab125e] return self.greenlet.switch() [ 604.015937] env[62066]: ERROR nova.compute.manager [instance: 39cd204a-49c2-4309-9030-555056ab125e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 604.015937] env[62066]: ERROR nova.compute.manager [instance: 39cd204a-49c2-4309-9030-555056ab125e] result = function(*args, **kwargs) [ 604.015937] env[62066]: ERROR nova.compute.manager [instance: 39cd204a-49c2-4309-9030-555056ab125e] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 604.015937] env[62066]: ERROR nova.compute.manager [instance: 39cd204a-49c2-4309-9030-555056ab125e] return func(*args, **kwargs) [ 604.015937] env[62066]: ERROR nova.compute.manager [instance: 39cd204a-49c2-4309-9030-555056ab125e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 604.015937] env[62066]: ERROR nova.compute.manager [instance: 39cd204a-49c2-4309-9030-555056ab125e] raise e [ 604.015937] env[62066]: ERROR nova.compute.manager [instance: 39cd204a-49c2-4309-9030-555056ab125e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 604.015937] env[62066]: ERROR nova.compute.manager [instance: 39cd204a-49c2-4309-9030-555056ab125e] nwinfo = self.network_api.allocate_for_instance( [ 604.015937] env[62066]: ERROR nova.compute.manager [instance: 39cd204a-49c2-4309-9030-555056ab125e] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 604.015937] env[62066]: ERROR nova.compute.manager [instance: 39cd204a-49c2-4309-9030-555056ab125e] created_port_ids = self._update_ports_for_instance( [ 604.015937] env[62066]: ERROR nova.compute.manager [instance: 39cd204a-49c2-4309-9030-555056ab125e] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 604.015937] env[62066]: ERROR nova.compute.manager [instance: 39cd204a-49c2-4309-9030-555056ab125e] with excutils.save_and_reraise_exception(): [ 604.015937] env[62066]: ERROR nova.compute.manager [instance: 39cd204a-49c2-4309-9030-555056ab125e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 604.016285] env[62066]: ERROR nova.compute.manager [instance: 39cd204a-49c2-4309-9030-555056ab125e] self.force_reraise() [ 604.016285] env[62066]: ERROR nova.compute.manager [instance: 39cd204a-49c2-4309-9030-555056ab125e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 604.016285] env[62066]: ERROR nova.compute.manager [instance: 39cd204a-49c2-4309-9030-555056ab125e] raise self.value [ 604.016285] env[62066]: ERROR nova.compute.manager [instance: 39cd204a-49c2-4309-9030-555056ab125e] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 604.016285] env[62066]: ERROR nova.compute.manager [instance: 39cd204a-49c2-4309-9030-555056ab125e] updated_port = self._update_port( [ 604.016285] env[62066]: ERROR nova.compute.manager [instance: 39cd204a-49c2-4309-9030-555056ab125e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 604.016285] env[62066]: ERROR nova.compute.manager [instance: 39cd204a-49c2-4309-9030-555056ab125e] _ensure_no_port_binding_failure(port) [ 604.016285] env[62066]: ERROR nova.compute.manager [instance: 39cd204a-49c2-4309-9030-555056ab125e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 604.016285] env[62066]: ERROR nova.compute.manager [instance: 39cd204a-49c2-4309-9030-555056ab125e] raise exception.PortBindingFailed(port_id=port['id']) [ 604.016285] env[62066]: ERROR nova.compute.manager [instance: 39cd204a-49c2-4309-9030-555056ab125e] nova.exception.PortBindingFailed: Binding failed for port 00047463-06a3-4ffe-830b-2d4ef9e64de6, please check neutron logs for more information. [ 604.016285] env[62066]: ERROR nova.compute.manager [instance: 39cd204a-49c2-4309-9030-555056ab125e] [ 604.016620] env[62066]: DEBUG nova.compute.utils [None req-4fa8ac7b-ee00-4ef3-97aa-07a2c2ae6378 tempest-ServerExternalEventsTest-41922728 tempest-ServerExternalEventsTest-41922728-project-member] [instance: 39cd204a-49c2-4309-9030-555056ab125e] Binding failed for port 00047463-06a3-4ffe-830b-2d4ef9e64de6, please check neutron logs for more information. {{(pid=62066) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 604.018223] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.664s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 604.019636] env[62066]: INFO nova.compute.claims [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 604.022725] env[62066]: DEBUG nova.compute.manager [None req-4fa8ac7b-ee00-4ef3-97aa-07a2c2ae6378 tempest-ServerExternalEventsTest-41922728 tempest-ServerExternalEventsTest-41922728-project-member] [instance: 39cd204a-49c2-4309-9030-555056ab125e] Build of instance 39cd204a-49c2-4309-9030-555056ab125e was re-scheduled: Binding failed for port 00047463-06a3-4ffe-830b-2d4ef9e64de6, please check neutron logs for more information. {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 604.022911] env[62066]: DEBUG nova.compute.manager [None req-4fa8ac7b-ee00-4ef3-97aa-07a2c2ae6378 tempest-ServerExternalEventsTest-41922728 tempest-ServerExternalEventsTest-41922728-project-member] [instance: 39cd204a-49c2-4309-9030-555056ab125e] Unplugging VIFs for instance {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 604.023237] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4fa8ac7b-ee00-4ef3-97aa-07a2c2ae6378 tempest-ServerExternalEventsTest-41922728 tempest-ServerExternalEventsTest-41922728-project-member] Acquiring lock "refresh_cache-39cd204a-49c2-4309-9030-555056ab125e" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 604.023396] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4fa8ac7b-ee00-4ef3-97aa-07a2c2ae6378 tempest-ServerExternalEventsTest-41922728 tempest-ServerExternalEventsTest-41922728-project-member] Acquired lock "refresh_cache-39cd204a-49c2-4309-9030-555056ab125e" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 604.023574] env[62066]: DEBUG nova.network.neutron [None req-4fa8ac7b-ee00-4ef3-97aa-07a2c2ae6378 tempest-ServerExternalEventsTest-41922728 tempest-ServerExternalEventsTest-41922728-project-member] [instance: 39cd204a-49c2-4309-9030-555056ab125e] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 604.043266] env[62066]: DEBUG nova.virt.hardware [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-23T13:40:41Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-23T13:40:23Z,direct_url=,disk_format='vmdk',id=50ff584c-3b50-4395-af07-3e66769bc9f7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='eb52f7069a374c61ae946f052007c6d9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-23T13:40:24Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 604.043505] env[62066]: DEBUG nova.virt.hardware [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 604.043658] env[62066]: DEBUG nova.virt.hardware [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 604.043834] env[62066]: DEBUG nova.virt.hardware [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 604.043976] env[62066]: DEBUG nova.virt.hardware [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 604.044139] env[62066]: DEBUG nova.virt.hardware [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 604.044341] env[62066]: DEBUG nova.virt.hardware [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 604.044501] env[62066]: DEBUG nova.virt.hardware [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 604.044664] env[62066]: DEBUG nova.virt.hardware [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 604.044824] env[62066]: DEBUG nova.virt.hardware [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 604.045527] env[62066]: DEBUG nova.virt.hardware [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 604.045922] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9ed23d0-958d-40d0-9284-203bf7a2893d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.054139] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5438658-d7c8-4f6b-a940-45840b73ba31 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.070614] env[62066]: ERROR nova.compute.manager [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port afd58cf2-370e-45b9-bc5a-7bea27a39c71, please check neutron logs for more information. [ 604.070614] env[62066]: ERROR nova.compute.manager [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] Traceback (most recent call last): [ 604.070614] env[62066]: ERROR nova.compute.manager [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 604.070614] env[62066]: ERROR nova.compute.manager [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] yield resources [ 604.070614] env[62066]: ERROR nova.compute.manager [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 604.070614] env[62066]: ERROR nova.compute.manager [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] self.driver.spawn(context, instance, image_meta, [ 604.070614] env[62066]: ERROR nova.compute.manager [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 604.070614] env[62066]: ERROR nova.compute.manager [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] self._vmops.spawn(context, instance, image_meta, injected_files, [ 604.070614] env[62066]: ERROR nova.compute.manager [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 604.070614] env[62066]: ERROR nova.compute.manager [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] vm_ref = self.build_virtual_machine(instance, [ 604.070614] env[62066]: ERROR nova.compute.manager [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 604.070952] env[62066]: ERROR nova.compute.manager [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] vif_infos = vmwarevif.get_vif_info(self._session, [ 604.070952] env[62066]: ERROR nova.compute.manager [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 604.070952] env[62066]: ERROR nova.compute.manager [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] for vif in network_info: [ 604.070952] env[62066]: ERROR nova.compute.manager [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 604.070952] env[62066]: ERROR nova.compute.manager [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] return self._sync_wrapper(fn, *args, **kwargs) [ 604.070952] env[62066]: ERROR nova.compute.manager [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 604.070952] env[62066]: ERROR nova.compute.manager [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] self.wait() [ 604.070952] env[62066]: ERROR nova.compute.manager [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 604.070952] env[62066]: ERROR nova.compute.manager [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] self[:] = self._gt.wait() [ 604.070952] env[62066]: ERROR nova.compute.manager [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 604.070952] env[62066]: ERROR nova.compute.manager [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] return self._exit_event.wait() [ 604.070952] env[62066]: ERROR nova.compute.manager [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 604.070952] env[62066]: ERROR nova.compute.manager [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] current.throw(*self._exc) [ 604.071289] env[62066]: ERROR nova.compute.manager [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 604.071289] env[62066]: ERROR nova.compute.manager [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] result = function(*args, **kwargs) [ 604.071289] env[62066]: ERROR nova.compute.manager [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 604.071289] env[62066]: ERROR nova.compute.manager [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] return func(*args, **kwargs) [ 604.071289] env[62066]: ERROR nova.compute.manager [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 604.071289] env[62066]: ERROR nova.compute.manager [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] raise e [ 604.071289] env[62066]: ERROR nova.compute.manager [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 604.071289] env[62066]: ERROR nova.compute.manager [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] nwinfo = self.network_api.allocate_for_instance( [ 604.071289] env[62066]: ERROR nova.compute.manager [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 604.071289] env[62066]: ERROR nova.compute.manager [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] created_port_ids = self._update_ports_for_instance( [ 604.071289] env[62066]: ERROR nova.compute.manager [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 604.071289] env[62066]: ERROR nova.compute.manager [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] with excutils.save_and_reraise_exception(): [ 604.071289] env[62066]: ERROR nova.compute.manager [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 604.071595] env[62066]: ERROR nova.compute.manager [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] self.force_reraise() [ 604.071595] env[62066]: ERROR nova.compute.manager [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 604.071595] env[62066]: ERROR nova.compute.manager [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] raise self.value [ 604.071595] env[62066]: ERROR nova.compute.manager [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 604.071595] env[62066]: ERROR nova.compute.manager [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] updated_port = self._update_port( [ 604.071595] env[62066]: ERROR nova.compute.manager [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 604.071595] env[62066]: ERROR nova.compute.manager [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] _ensure_no_port_binding_failure(port) [ 604.071595] env[62066]: ERROR nova.compute.manager [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 604.071595] env[62066]: ERROR nova.compute.manager [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] raise exception.PortBindingFailed(port_id=port['id']) [ 604.071595] env[62066]: ERROR nova.compute.manager [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] nova.exception.PortBindingFailed: Binding failed for port afd58cf2-370e-45b9-bc5a-7bea27a39c71, please check neutron logs for more information. [ 604.071595] env[62066]: ERROR nova.compute.manager [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] [ 604.071595] env[62066]: INFO nova.compute.manager [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] Terminating instance [ 604.072868] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Acquiring lock "refresh_cache-5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 604.136075] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7abf52e5-ee7c-4f92-a757-ffa8ca735a57 tempest-ServerActionsTestJSON-164481335 tempest-ServerActionsTestJSON-164481335-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 604.481745] env[62066]: DEBUG nova.network.neutron [req-7d9a7094-5f3a-4cc5-b664-3e39a43d3390 req-21ee7ac9-2613-444c-ac93-6beb183a71f1 service nova] [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 604.561214] env[62066]: DEBUG nova.network.neutron [None req-4fa8ac7b-ee00-4ef3-97aa-07a2c2ae6378 tempest-ServerExternalEventsTest-41922728 tempest-ServerExternalEventsTest-41922728-project-member] [instance: 39cd204a-49c2-4309-9030-555056ab125e] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 604.583666] env[62066]: DEBUG nova.network.neutron [req-7d9a7094-5f3a-4cc5-b664-3e39a43d3390 req-21ee7ac9-2613-444c-ac93-6beb183a71f1 service nova] [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 604.668963] env[62066]: DEBUG nova.network.neutron [None req-4fa8ac7b-ee00-4ef3-97aa-07a2c2ae6378 tempest-ServerExternalEventsTest-41922728 tempest-ServerExternalEventsTest-41922728-project-member] [instance: 39cd204a-49c2-4309-9030-555056ab125e] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 605.086021] env[62066]: DEBUG oslo_concurrency.lockutils [req-7d9a7094-5f3a-4cc5-b664-3e39a43d3390 req-21ee7ac9-2613-444c-ac93-6beb183a71f1 service nova] Releasing lock "refresh_cache-5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 605.087615] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Acquired lock "refresh_cache-5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 605.087615] env[62066]: DEBUG nova.network.neutron [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 605.172090] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4fa8ac7b-ee00-4ef3-97aa-07a2c2ae6378 tempest-ServerExternalEventsTest-41922728 tempest-ServerExternalEventsTest-41922728-project-member] Releasing lock "refresh_cache-39cd204a-49c2-4309-9030-555056ab125e" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 605.172090] env[62066]: DEBUG nova.compute.manager [None req-4fa8ac7b-ee00-4ef3-97aa-07a2c2ae6378 tempest-ServerExternalEventsTest-41922728 tempest-ServerExternalEventsTest-41922728-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 605.172090] env[62066]: DEBUG nova.compute.manager [None req-4fa8ac7b-ee00-4ef3-97aa-07a2c2ae6378 tempest-ServerExternalEventsTest-41922728 tempest-ServerExternalEventsTest-41922728-project-member] [instance: 39cd204a-49c2-4309-9030-555056ab125e] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 605.172090] env[62066]: DEBUG nova.network.neutron [None req-4fa8ac7b-ee00-4ef3-97aa-07a2c2ae6378 tempest-ServerExternalEventsTest-41922728 tempest-ServerExternalEventsTest-41922728-project-member] [instance: 39cd204a-49c2-4309-9030-555056ab125e] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 605.202295] env[62066]: DEBUG nova.network.neutron [None req-4fa8ac7b-ee00-4ef3-97aa-07a2c2ae6378 tempest-ServerExternalEventsTest-41922728 tempest-ServerExternalEventsTest-41922728-project-member] [instance: 39cd204a-49c2-4309-9030-555056ab125e] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 605.431326] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-741eade1-a4c8-4c34-9c9c-b8df9d540786 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.439208] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ade1fdb-4aa2-46ca-81c5-6de19a0048d5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.469034] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36406ab2-c0d5-4f10-8ac0-24ee7eec7e3b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.476364] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c2ed081-bab9-40fb-babe-6f1655eb4513 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.489399] env[62066]: DEBUG nova.compute.provider_tree [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 605.611732] env[62066]: DEBUG nova.network.neutron [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 605.704773] env[62066]: DEBUG nova.network.neutron [None req-4fa8ac7b-ee00-4ef3-97aa-07a2c2ae6378 tempest-ServerExternalEventsTest-41922728 tempest-ServerExternalEventsTest-41922728-project-member] [instance: 39cd204a-49c2-4309-9030-555056ab125e] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 605.807567] env[62066]: DEBUG nova.network.neutron [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 605.984174] env[62066]: DEBUG nova.compute.manager [req-d0022732-42be-4d17-a920-a9e6226e27e4 req-9d021299-a7ed-4a52-bf46-ef3d67476201 service nova] [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] Received event network-vif-deleted-afd58cf2-370e-45b9-bc5a-7bea27a39c71 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 605.992180] env[62066]: DEBUG nova.scheduler.client.report [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 606.210613] env[62066]: INFO nova.compute.manager [None req-4fa8ac7b-ee00-4ef3-97aa-07a2c2ae6378 tempest-ServerExternalEventsTest-41922728 tempest-ServerExternalEventsTest-41922728-project-member] [instance: 39cd204a-49c2-4309-9030-555056ab125e] Took 1.04 seconds to deallocate network for instance. [ 606.314349] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Releasing lock "refresh_cache-5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 606.314862] env[62066]: DEBUG nova.compute.manager [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 606.315139] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 606.315492] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6bcf6d48-db91-4382-a12e-a680a028edb8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.325370] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0239bac0-e612-4333-ade1-96bf0491855d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.349713] env[62066]: WARNING nova.virt.vmwareapi.vmops [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962 could not be found. [ 606.349954] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 606.350160] env[62066]: INFO nova.compute.manager [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] Took 0.04 seconds to destroy the instance on the hypervisor. [ 606.350439] env[62066]: DEBUG oslo.service.loopingcall [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 606.350668] env[62066]: DEBUG nova.compute.manager [-] [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 606.350758] env[62066]: DEBUG nova.network.neutron [-] [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 606.376071] env[62066]: DEBUG nova.network.neutron [-] [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 606.497820] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.480s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 606.498400] env[62066]: DEBUG nova.compute.manager [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 606.501141] env[62066]: DEBUG oslo_concurrency.lockutils [None req-14085d2f-7e58-404a-a2d0-4487ac658e45 tempest-FloatingIPsAssociationTestJSON-608653087 tempest-FloatingIPsAssociationTestJSON-608653087-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.355s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 606.503278] env[62066]: INFO nova.compute.claims [None req-14085d2f-7e58-404a-a2d0-4487ac658e45 tempest-FloatingIPsAssociationTestJSON-608653087 tempest-FloatingIPsAssociationTestJSON-608653087-project-member] [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 606.880423] env[62066]: DEBUG nova.network.neutron [-] [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 607.009942] env[62066]: DEBUG nova.compute.utils [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 607.015204] env[62066]: DEBUG nova.compute.manager [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 607.015460] env[62066]: DEBUG nova.network.neutron [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 607.081565] env[62066]: DEBUG nova.policy [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '236eede7716d4664a75f5df76aa403ce', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7664e11610664ce5a85a2ad8758986f8', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 607.255069] env[62066]: INFO nova.scheduler.client.report [None req-4fa8ac7b-ee00-4ef3-97aa-07a2c2ae6378 tempest-ServerExternalEventsTest-41922728 tempest-ServerExternalEventsTest-41922728-project-member] Deleted allocations for instance 39cd204a-49c2-4309-9030-555056ab125e [ 607.387171] env[62066]: INFO nova.compute.manager [-] [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] Took 1.04 seconds to deallocate network for instance. [ 607.391073] env[62066]: DEBUG nova.compute.claims [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] Aborting claim: {{(pid=62066) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 607.391261] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 607.522323] env[62066]: DEBUG nova.compute.manager [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 607.735118] env[62066]: DEBUG nova.network.neutron [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] Successfully created port: fd8616a9-d9b8-42e8-b0d3-4ca9121e33e7 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 607.772221] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4fa8ac7b-ee00-4ef3-97aa-07a2c2ae6378 tempest-ServerExternalEventsTest-41922728 tempest-ServerExternalEventsTest-41922728-project-member] Lock "39cd204a-49c2-4309-9030-555056ab125e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 92.991s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 607.900983] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56115483-eb5b-4533-a58b-914b62a57ce0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.908899] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a43789fc-bc59-4e57-bfbb-1b50e0e8103c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.939293] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8227e811-8698-4f58-8ee0-489dbfd55af2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.949883] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76ce9d68-39d1-4a81-951b-4b445b672707 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.968947] env[62066]: DEBUG nova.compute.provider_tree [None req-14085d2f-7e58-404a-a2d0-4487ac658e45 tempest-FloatingIPsAssociationTestJSON-608653087 tempest-FloatingIPsAssociationTestJSON-608653087-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 608.279078] env[62066]: DEBUG nova.compute.manager [None req-29878964-568e-4c92-80e6-64ee3e8e11bb tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 608.475361] env[62066]: DEBUG nova.scheduler.client.report [None req-14085d2f-7e58-404a-a2d0-4487ac658e45 tempest-FloatingIPsAssociationTestJSON-608653087 tempest-FloatingIPsAssociationTestJSON-608653087-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 608.544578] env[62066]: DEBUG nova.compute.manager [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 608.580361] env[62066]: DEBUG nova.virt.hardware [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-23T13:40:41Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-23T13:40:23Z,direct_url=,disk_format='vmdk',id=50ff584c-3b50-4395-af07-3e66769bc9f7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='eb52f7069a374c61ae946f052007c6d9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-23T13:40:24Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 608.580675] env[62066]: DEBUG nova.virt.hardware [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 608.581049] env[62066]: DEBUG nova.virt.hardware [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 608.581184] env[62066]: DEBUG nova.virt.hardware [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 608.581935] env[62066]: DEBUG nova.virt.hardware [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 608.581935] env[62066]: DEBUG nova.virt.hardware [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 608.581935] env[62066]: DEBUG nova.virt.hardware [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 608.581935] env[62066]: DEBUG nova.virt.hardware [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 608.582158] env[62066]: DEBUG nova.virt.hardware [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 608.582360] env[62066]: DEBUG nova.virt.hardware [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 608.582576] env[62066]: DEBUG nova.virt.hardware [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 608.584078] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75b9ab1d-9877-495f-8fb4-b85005f7d5a2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.591977] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de9979f1-06ec-4a65-a987-47987dab1376 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.816862] env[62066]: DEBUG oslo_concurrency.lockutils [None req-29878964-568e-4c92-80e6-64ee3e8e11bb tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 608.981239] env[62066]: DEBUG oslo_concurrency.lockutils [None req-14085d2f-7e58-404a-a2d0-4487ac658e45 tempest-FloatingIPsAssociationTestJSON-608653087 tempest-FloatingIPsAssociationTestJSON-608653087-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.480s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 608.981845] env[62066]: DEBUG nova.compute.manager [None req-14085d2f-7e58-404a-a2d0-4487ac658e45 tempest-FloatingIPsAssociationTestJSON-608653087 tempest-FloatingIPsAssociationTestJSON-608653087-project-member] [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 608.985275] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1e3e30d6-8bb5-4347-b6c2-c2d648c37b67 tempest-ImagesOneServerNegativeTestJSON-310236551 tempest-ImagesOneServerNegativeTestJSON-310236551-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.770s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 609.494516] env[62066]: DEBUG nova.compute.utils [None req-14085d2f-7e58-404a-a2d0-4487ac658e45 tempest-FloatingIPsAssociationTestJSON-608653087 tempest-FloatingIPsAssociationTestJSON-608653087-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 609.496153] env[62066]: DEBUG nova.compute.manager [None req-14085d2f-7e58-404a-a2d0-4487ac658e45 tempest-FloatingIPsAssociationTestJSON-608653087 tempest-FloatingIPsAssociationTestJSON-608653087-project-member] [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 609.498034] env[62066]: DEBUG nova.network.neutron [None req-14085d2f-7e58-404a-a2d0-4487ac658e45 tempest-FloatingIPsAssociationTestJSON-608653087 tempest-FloatingIPsAssociationTestJSON-608653087-project-member] [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 609.561316] env[62066]: DEBUG nova.policy [None req-14085d2f-7e58-404a-a2d0-4487ac658e45 tempest-FloatingIPsAssociationTestJSON-608653087 tempest-FloatingIPsAssociationTestJSON-608653087-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '792647f6c95c460e81d6febe03e72cbb', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4c3896cd773e4ebaa5c9d3d4495c080e', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 609.612611] env[62066]: DEBUG nova.compute.manager [req-7e4221af-065a-4bcc-a0f1-5b4271286754 req-82e30490-3ae1-467e-b0a9-a78e3d88d109 service nova] [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] Received event network-changed-fd8616a9-d9b8-42e8-b0d3-4ca9121e33e7 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 609.612611] env[62066]: DEBUG nova.compute.manager [req-7e4221af-065a-4bcc-a0f1-5b4271286754 req-82e30490-3ae1-467e-b0a9-a78e3d88d109 service nova] [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] Refreshing instance network info cache due to event network-changed-fd8616a9-d9b8-42e8-b0d3-4ca9121e33e7. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 609.612611] env[62066]: DEBUG oslo_concurrency.lockutils [req-7e4221af-065a-4bcc-a0f1-5b4271286754 req-82e30490-3ae1-467e-b0a9-a78e3d88d109 service nova] Acquiring lock "refresh_cache-72c2dbe0-78de-4a7d-98df-cf405a5f20e2" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 609.612611] env[62066]: DEBUG oslo_concurrency.lockutils [req-7e4221af-065a-4bcc-a0f1-5b4271286754 req-82e30490-3ae1-467e-b0a9-a78e3d88d109 service nova] Acquired lock "refresh_cache-72c2dbe0-78de-4a7d-98df-cf405a5f20e2" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 609.615028] env[62066]: DEBUG nova.network.neutron [req-7e4221af-065a-4bcc-a0f1-5b4271286754 req-82e30490-3ae1-467e-b0a9-a78e3d88d109 service nova] [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] Refreshing network info cache for port fd8616a9-d9b8-42e8-b0d3-4ca9121e33e7 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 609.844938] env[62066]: ERROR nova.compute.manager [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port fd8616a9-d9b8-42e8-b0d3-4ca9121e33e7, please check neutron logs for more information. [ 609.844938] env[62066]: ERROR nova.compute.manager Traceback (most recent call last): [ 609.844938] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 609.844938] env[62066]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 609.844938] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 609.844938] env[62066]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 609.844938] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 609.844938] env[62066]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 609.844938] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 609.844938] env[62066]: ERROR nova.compute.manager self.force_reraise() [ 609.844938] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 609.844938] env[62066]: ERROR nova.compute.manager raise self.value [ 609.844938] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 609.844938] env[62066]: ERROR nova.compute.manager updated_port = self._update_port( [ 609.844938] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 609.844938] env[62066]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 609.845424] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 609.845424] env[62066]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 609.845424] env[62066]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port fd8616a9-d9b8-42e8-b0d3-4ca9121e33e7, please check neutron logs for more information. [ 609.845424] env[62066]: ERROR nova.compute.manager [ 609.845424] env[62066]: Traceback (most recent call last): [ 609.845424] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 609.845424] env[62066]: listener.cb(fileno) [ 609.845424] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 609.845424] env[62066]: result = function(*args, **kwargs) [ 609.845424] env[62066]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 609.845424] env[62066]: return func(*args, **kwargs) [ 609.845424] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 609.845424] env[62066]: raise e [ 609.845424] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 609.845424] env[62066]: nwinfo = self.network_api.allocate_for_instance( [ 609.845424] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 609.845424] env[62066]: created_port_ids = self._update_ports_for_instance( [ 609.845424] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 609.845424] env[62066]: with excutils.save_and_reraise_exception(): [ 609.845424] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 609.845424] env[62066]: self.force_reraise() [ 609.845424] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 609.845424] env[62066]: raise self.value [ 609.845424] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 609.845424] env[62066]: updated_port = self._update_port( [ 609.845424] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 609.845424] env[62066]: _ensure_no_port_binding_failure(port) [ 609.845424] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 609.845424] env[62066]: raise exception.PortBindingFailed(port_id=port['id']) [ 609.848022] env[62066]: nova.exception.PortBindingFailed: Binding failed for port fd8616a9-d9b8-42e8-b0d3-4ca9121e33e7, please check neutron logs for more information. [ 609.848022] env[62066]: Removing descriptor: 20 [ 609.848022] env[62066]: ERROR nova.compute.manager [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port fd8616a9-d9b8-42e8-b0d3-4ca9121e33e7, please check neutron logs for more information. [ 609.848022] env[62066]: ERROR nova.compute.manager [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] Traceback (most recent call last): [ 609.848022] env[62066]: ERROR nova.compute.manager [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 609.848022] env[62066]: ERROR nova.compute.manager [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] yield resources [ 609.848022] env[62066]: ERROR nova.compute.manager [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 609.848022] env[62066]: ERROR nova.compute.manager [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] self.driver.spawn(context, instance, image_meta, [ 609.848022] env[62066]: ERROR nova.compute.manager [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 609.848022] env[62066]: ERROR nova.compute.manager [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] self._vmops.spawn(context, instance, image_meta, injected_files, [ 609.848022] env[62066]: ERROR nova.compute.manager [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 609.848022] env[62066]: ERROR nova.compute.manager [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] vm_ref = self.build_virtual_machine(instance, [ 609.848478] env[62066]: ERROR nova.compute.manager [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 609.848478] env[62066]: ERROR nova.compute.manager [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] vif_infos = vmwarevif.get_vif_info(self._session, [ 609.848478] env[62066]: ERROR nova.compute.manager [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 609.848478] env[62066]: ERROR nova.compute.manager [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] for vif in network_info: [ 609.848478] env[62066]: ERROR nova.compute.manager [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 609.848478] env[62066]: ERROR nova.compute.manager [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] return self._sync_wrapper(fn, *args, **kwargs) [ 609.848478] env[62066]: ERROR nova.compute.manager [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 609.848478] env[62066]: ERROR nova.compute.manager [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] self.wait() [ 609.848478] env[62066]: ERROR nova.compute.manager [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 609.848478] env[62066]: ERROR nova.compute.manager [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] self[:] = self._gt.wait() [ 609.848478] env[62066]: ERROR nova.compute.manager [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 609.848478] env[62066]: ERROR nova.compute.manager [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] return self._exit_event.wait() [ 609.848478] env[62066]: ERROR nova.compute.manager [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 609.848816] env[62066]: ERROR nova.compute.manager [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] result = hub.switch() [ 609.848816] env[62066]: ERROR nova.compute.manager [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 609.848816] env[62066]: ERROR nova.compute.manager [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] return self.greenlet.switch() [ 609.848816] env[62066]: ERROR nova.compute.manager [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 609.848816] env[62066]: ERROR nova.compute.manager [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] result = function(*args, **kwargs) [ 609.848816] env[62066]: ERROR nova.compute.manager [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 609.848816] env[62066]: ERROR nova.compute.manager [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] return func(*args, **kwargs) [ 609.848816] env[62066]: ERROR nova.compute.manager [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 609.848816] env[62066]: ERROR nova.compute.manager [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] raise e [ 609.848816] env[62066]: ERROR nova.compute.manager [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 609.848816] env[62066]: ERROR nova.compute.manager [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] nwinfo = self.network_api.allocate_for_instance( [ 609.848816] env[62066]: ERROR nova.compute.manager [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 609.848816] env[62066]: ERROR nova.compute.manager [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] created_port_ids = self._update_ports_for_instance( [ 609.849149] env[62066]: ERROR nova.compute.manager [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 609.849149] env[62066]: ERROR nova.compute.manager [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] with excutils.save_and_reraise_exception(): [ 609.849149] env[62066]: ERROR nova.compute.manager [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 609.849149] env[62066]: ERROR nova.compute.manager [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] self.force_reraise() [ 609.849149] env[62066]: ERROR nova.compute.manager [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 609.849149] env[62066]: ERROR nova.compute.manager [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] raise self.value [ 609.849149] env[62066]: ERROR nova.compute.manager [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 609.849149] env[62066]: ERROR nova.compute.manager [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] updated_port = self._update_port( [ 609.849149] env[62066]: ERROR nova.compute.manager [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 609.849149] env[62066]: ERROR nova.compute.manager [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] _ensure_no_port_binding_failure(port) [ 609.849149] env[62066]: ERROR nova.compute.manager [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 609.849149] env[62066]: ERROR nova.compute.manager [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] raise exception.PortBindingFailed(port_id=port['id']) [ 609.849471] env[62066]: ERROR nova.compute.manager [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] nova.exception.PortBindingFailed: Binding failed for port fd8616a9-d9b8-42e8-b0d3-4ca9121e33e7, please check neutron logs for more information. [ 609.849471] env[62066]: ERROR nova.compute.manager [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] [ 609.849471] env[62066]: INFO nova.compute.manager [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] Terminating instance [ 609.849471] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Acquiring lock "refresh_cache-72c2dbe0-78de-4a7d-98df-cf405a5f20e2" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 609.922486] env[62066]: DEBUG nova.network.neutron [None req-14085d2f-7e58-404a-a2d0-4487ac658e45 tempest-FloatingIPsAssociationTestJSON-608653087 tempest-FloatingIPsAssociationTestJSON-608653087-project-member] [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] Successfully created port: d641e931-fa53-4d0a-ad67-1b5304b8fbee {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 610.000222] env[62066]: DEBUG nova.compute.manager [None req-14085d2f-7e58-404a-a2d0-4487ac658e45 tempest-FloatingIPsAssociationTestJSON-608653087 tempest-FloatingIPsAssociationTestJSON-608653087-project-member] [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 610.051043] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a75f4dc-ef2f-419b-8155-cdcb27133a34 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.059835] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0be5f60-dfa2-49be-9a76-3b3c85499458 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.095794] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d34f7a91-a089-4d9c-8916-c690db4305ee {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.101190] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a4d93091-69b6-4b35-887c-e7b927c2300a tempest-ServerMetadataTestJSON-1590103242 tempest-ServerMetadataTestJSON-1590103242-project-member] Acquiring lock "0ed07533-9e79-416d-a495-a8e20b97e108" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 610.101416] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a4d93091-69b6-4b35-887c-e7b927c2300a tempest-ServerMetadataTestJSON-1590103242 tempest-ServerMetadataTestJSON-1590103242-project-member] Lock "0ed07533-9e79-416d-a495-a8e20b97e108" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 610.107456] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f92059b-0f90-43ee-b7b3-7386d6ada5a7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.124504] env[62066]: DEBUG nova.compute.provider_tree [None req-1e3e30d6-8bb5-4347-b6c2-c2d648c37b67 tempest-ImagesOneServerNegativeTestJSON-310236551 tempest-ImagesOneServerNegativeTestJSON-310236551-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 610.155415] env[62066]: DEBUG nova.network.neutron [req-7e4221af-065a-4bcc-a0f1-5b4271286754 req-82e30490-3ae1-467e-b0a9-a78e3d88d109 service nova] [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 610.288926] env[62066]: DEBUG nova.network.neutron [req-7e4221af-065a-4bcc-a0f1-5b4271286754 req-82e30490-3ae1-467e-b0a9-a78e3d88d109 service nova] [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 610.628740] env[62066]: DEBUG nova.scheduler.client.report [None req-1e3e30d6-8bb5-4347-b6c2-c2d648c37b67 tempest-ImagesOneServerNegativeTestJSON-310236551 tempest-ImagesOneServerNegativeTestJSON-310236551-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 610.793531] env[62066]: DEBUG oslo_concurrency.lockutils [req-7e4221af-065a-4bcc-a0f1-5b4271286754 req-82e30490-3ae1-467e-b0a9-a78e3d88d109 service nova] Releasing lock "refresh_cache-72c2dbe0-78de-4a7d-98df-cf405a5f20e2" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 610.793958] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Acquired lock "refresh_cache-72c2dbe0-78de-4a7d-98df-cf405a5f20e2" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 610.794161] env[62066]: DEBUG nova.network.neutron [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 610.893498] env[62066]: DEBUG nova.compute.manager [req-14aaf419-0bbe-4bbc-b05a-c505245add01 req-3c1510af-04c1-48f8-b8ff-2e0c5b220b07 service nova] [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] Received event network-changed-d641e931-fa53-4d0a-ad67-1b5304b8fbee {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 610.893498] env[62066]: DEBUG nova.compute.manager [req-14aaf419-0bbe-4bbc-b05a-c505245add01 req-3c1510af-04c1-48f8-b8ff-2e0c5b220b07 service nova] [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] Refreshing instance network info cache due to event network-changed-d641e931-fa53-4d0a-ad67-1b5304b8fbee. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 610.893498] env[62066]: DEBUG oslo_concurrency.lockutils [req-14aaf419-0bbe-4bbc-b05a-c505245add01 req-3c1510af-04c1-48f8-b8ff-2e0c5b220b07 service nova] Acquiring lock "refresh_cache-08d41411-7928-4379-9f2e-c6ce00843d82" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 610.893498] env[62066]: DEBUG oslo_concurrency.lockutils [req-14aaf419-0bbe-4bbc-b05a-c505245add01 req-3c1510af-04c1-48f8-b8ff-2e0c5b220b07 service nova] Acquired lock "refresh_cache-08d41411-7928-4379-9f2e-c6ce00843d82" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 610.893498] env[62066]: DEBUG nova.network.neutron [req-14aaf419-0bbe-4bbc-b05a-c505245add01 req-3c1510af-04c1-48f8-b8ff-2e0c5b220b07 service nova] [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] Refreshing network info cache for port d641e931-fa53-4d0a-ad67-1b5304b8fbee {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 611.013016] env[62066]: DEBUG nova.compute.manager [None req-14085d2f-7e58-404a-a2d0-4487ac658e45 tempest-FloatingIPsAssociationTestJSON-608653087 tempest-FloatingIPsAssociationTestJSON-608653087-project-member] [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 611.039021] env[62066]: DEBUG nova.virt.hardware [None req-14085d2f-7e58-404a-a2d0-4487ac658e45 tempest-FloatingIPsAssociationTestJSON-608653087 tempest-FloatingIPsAssociationTestJSON-608653087-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-23T13:40:41Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-23T13:40:23Z,direct_url=,disk_format='vmdk',id=50ff584c-3b50-4395-af07-3e66769bc9f7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='eb52f7069a374c61ae946f052007c6d9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-23T13:40:24Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 611.039021] env[62066]: DEBUG nova.virt.hardware [None req-14085d2f-7e58-404a-a2d0-4487ac658e45 tempest-FloatingIPsAssociationTestJSON-608653087 tempest-FloatingIPsAssociationTestJSON-608653087-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 611.039021] env[62066]: DEBUG nova.virt.hardware [None req-14085d2f-7e58-404a-a2d0-4487ac658e45 tempest-FloatingIPsAssociationTestJSON-608653087 tempest-FloatingIPsAssociationTestJSON-608653087-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 611.039212] env[62066]: DEBUG nova.virt.hardware [None req-14085d2f-7e58-404a-a2d0-4487ac658e45 tempest-FloatingIPsAssociationTestJSON-608653087 tempest-FloatingIPsAssociationTestJSON-608653087-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 611.039212] env[62066]: DEBUG nova.virt.hardware [None req-14085d2f-7e58-404a-a2d0-4487ac658e45 tempest-FloatingIPsAssociationTestJSON-608653087 tempest-FloatingIPsAssociationTestJSON-608653087-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 611.039212] env[62066]: DEBUG nova.virt.hardware [None req-14085d2f-7e58-404a-a2d0-4487ac658e45 tempest-FloatingIPsAssociationTestJSON-608653087 tempest-FloatingIPsAssociationTestJSON-608653087-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 611.039212] env[62066]: DEBUG nova.virt.hardware [None req-14085d2f-7e58-404a-a2d0-4487ac658e45 tempest-FloatingIPsAssociationTestJSON-608653087 tempest-FloatingIPsAssociationTestJSON-608653087-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 611.039212] env[62066]: DEBUG nova.virt.hardware [None req-14085d2f-7e58-404a-a2d0-4487ac658e45 tempest-FloatingIPsAssociationTestJSON-608653087 tempest-FloatingIPsAssociationTestJSON-608653087-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 611.039450] env[62066]: DEBUG nova.virt.hardware [None req-14085d2f-7e58-404a-a2d0-4487ac658e45 tempest-FloatingIPsAssociationTestJSON-608653087 tempest-FloatingIPsAssociationTestJSON-608653087-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 611.043017] env[62066]: DEBUG nova.virt.hardware [None req-14085d2f-7e58-404a-a2d0-4487ac658e45 tempest-FloatingIPsAssociationTestJSON-608653087 tempest-FloatingIPsAssociationTestJSON-608653087-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 611.043017] env[62066]: DEBUG nova.virt.hardware [None req-14085d2f-7e58-404a-a2d0-4487ac658e45 tempest-FloatingIPsAssociationTestJSON-608653087 tempest-FloatingIPsAssociationTestJSON-608653087-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 611.043017] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3292cc5-aaa4-40cc-8daa-b2e2ebd42de2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.050134] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb8f55bd-fdcf-4ab2-8248-51989696ba38 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.135132] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1e3e30d6-8bb5-4347-b6c2-c2d648c37b67 tempest-ImagesOneServerNegativeTestJSON-310236551 tempest-ImagesOneServerNegativeTestJSON-310236551-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.150s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 611.135654] env[62066]: ERROR nova.compute.manager [None req-1e3e30d6-8bb5-4347-b6c2-c2d648c37b67 tempest-ImagesOneServerNegativeTestJSON-310236551 tempest-ImagesOneServerNegativeTestJSON-310236551-project-member] [instance: 194588b4-fe40-4286-8036-874a7c410327] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 2a2c9bbf-b369-4ea3-a866-79152380ed0b, please check neutron logs for more information. [ 611.135654] env[62066]: ERROR nova.compute.manager [instance: 194588b4-fe40-4286-8036-874a7c410327] Traceback (most recent call last): [ 611.135654] env[62066]: ERROR nova.compute.manager [instance: 194588b4-fe40-4286-8036-874a7c410327] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 611.135654] env[62066]: ERROR nova.compute.manager [instance: 194588b4-fe40-4286-8036-874a7c410327] self.driver.spawn(context, instance, image_meta, [ 611.135654] env[62066]: ERROR nova.compute.manager [instance: 194588b4-fe40-4286-8036-874a7c410327] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 611.135654] env[62066]: ERROR nova.compute.manager [instance: 194588b4-fe40-4286-8036-874a7c410327] self._vmops.spawn(context, instance, image_meta, injected_files, [ 611.135654] env[62066]: ERROR nova.compute.manager [instance: 194588b4-fe40-4286-8036-874a7c410327] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 611.135654] env[62066]: ERROR nova.compute.manager [instance: 194588b4-fe40-4286-8036-874a7c410327] vm_ref = self.build_virtual_machine(instance, [ 611.135654] env[62066]: ERROR nova.compute.manager [instance: 194588b4-fe40-4286-8036-874a7c410327] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 611.135654] env[62066]: ERROR nova.compute.manager [instance: 194588b4-fe40-4286-8036-874a7c410327] vif_infos = vmwarevif.get_vif_info(self._session, [ 611.135654] env[62066]: ERROR nova.compute.manager [instance: 194588b4-fe40-4286-8036-874a7c410327] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 611.136460] env[62066]: ERROR nova.compute.manager [instance: 194588b4-fe40-4286-8036-874a7c410327] for vif in network_info: [ 611.136460] env[62066]: ERROR nova.compute.manager [instance: 194588b4-fe40-4286-8036-874a7c410327] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 611.136460] env[62066]: ERROR nova.compute.manager [instance: 194588b4-fe40-4286-8036-874a7c410327] return self._sync_wrapper(fn, *args, **kwargs) [ 611.136460] env[62066]: ERROR nova.compute.manager [instance: 194588b4-fe40-4286-8036-874a7c410327] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 611.136460] env[62066]: ERROR nova.compute.manager [instance: 194588b4-fe40-4286-8036-874a7c410327] self.wait() [ 611.136460] env[62066]: ERROR nova.compute.manager [instance: 194588b4-fe40-4286-8036-874a7c410327] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 611.136460] env[62066]: ERROR nova.compute.manager [instance: 194588b4-fe40-4286-8036-874a7c410327] self[:] = self._gt.wait() [ 611.136460] env[62066]: ERROR nova.compute.manager [instance: 194588b4-fe40-4286-8036-874a7c410327] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 611.136460] env[62066]: ERROR nova.compute.manager [instance: 194588b4-fe40-4286-8036-874a7c410327] return self._exit_event.wait() [ 611.136460] env[62066]: ERROR nova.compute.manager [instance: 194588b4-fe40-4286-8036-874a7c410327] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 611.136460] env[62066]: ERROR nova.compute.manager [instance: 194588b4-fe40-4286-8036-874a7c410327] result = hub.switch() [ 611.136460] env[62066]: ERROR nova.compute.manager [instance: 194588b4-fe40-4286-8036-874a7c410327] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 611.136460] env[62066]: ERROR nova.compute.manager [instance: 194588b4-fe40-4286-8036-874a7c410327] return self.greenlet.switch() [ 611.139685] env[62066]: ERROR nova.compute.manager [instance: 194588b4-fe40-4286-8036-874a7c410327] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 611.139685] env[62066]: ERROR nova.compute.manager [instance: 194588b4-fe40-4286-8036-874a7c410327] result = function(*args, **kwargs) [ 611.139685] env[62066]: ERROR nova.compute.manager [instance: 194588b4-fe40-4286-8036-874a7c410327] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 611.139685] env[62066]: ERROR nova.compute.manager [instance: 194588b4-fe40-4286-8036-874a7c410327] return func(*args, **kwargs) [ 611.139685] env[62066]: ERROR nova.compute.manager [instance: 194588b4-fe40-4286-8036-874a7c410327] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 611.139685] env[62066]: ERROR nova.compute.manager [instance: 194588b4-fe40-4286-8036-874a7c410327] raise e [ 611.139685] env[62066]: ERROR nova.compute.manager [instance: 194588b4-fe40-4286-8036-874a7c410327] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 611.139685] env[62066]: ERROR nova.compute.manager [instance: 194588b4-fe40-4286-8036-874a7c410327] nwinfo = self.network_api.allocate_for_instance( [ 611.139685] env[62066]: ERROR nova.compute.manager [instance: 194588b4-fe40-4286-8036-874a7c410327] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 611.139685] env[62066]: ERROR nova.compute.manager [instance: 194588b4-fe40-4286-8036-874a7c410327] created_port_ids = self._update_ports_for_instance( [ 611.139685] env[62066]: ERROR nova.compute.manager [instance: 194588b4-fe40-4286-8036-874a7c410327] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 611.139685] env[62066]: ERROR nova.compute.manager [instance: 194588b4-fe40-4286-8036-874a7c410327] with excutils.save_and_reraise_exception(): [ 611.139685] env[62066]: ERROR nova.compute.manager [instance: 194588b4-fe40-4286-8036-874a7c410327] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 611.140267] env[62066]: ERROR nova.compute.manager [instance: 194588b4-fe40-4286-8036-874a7c410327] self.force_reraise() [ 611.140267] env[62066]: ERROR nova.compute.manager [instance: 194588b4-fe40-4286-8036-874a7c410327] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 611.140267] env[62066]: ERROR nova.compute.manager [instance: 194588b4-fe40-4286-8036-874a7c410327] raise self.value [ 611.140267] env[62066]: ERROR nova.compute.manager [instance: 194588b4-fe40-4286-8036-874a7c410327] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 611.140267] env[62066]: ERROR nova.compute.manager [instance: 194588b4-fe40-4286-8036-874a7c410327] updated_port = self._update_port( [ 611.140267] env[62066]: ERROR nova.compute.manager [instance: 194588b4-fe40-4286-8036-874a7c410327] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 611.140267] env[62066]: ERROR nova.compute.manager [instance: 194588b4-fe40-4286-8036-874a7c410327] _ensure_no_port_binding_failure(port) [ 611.140267] env[62066]: ERROR nova.compute.manager [instance: 194588b4-fe40-4286-8036-874a7c410327] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 611.140267] env[62066]: ERROR nova.compute.manager [instance: 194588b4-fe40-4286-8036-874a7c410327] raise exception.PortBindingFailed(port_id=port['id']) [ 611.140267] env[62066]: ERROR nova.compute.manager [instance: 194588b4-fe40-4286-8036-874a7c410327] nova.exception.PortBindingFailed: Binding failed for port 2a2c9bbf-b369-4ea3-a866-79152380ed0b, please check neutron logs for more information. [ 611.140267] env[62066]: ERROR nova.compute.manager [instance: 194588b4-fe40-4286-8036-874a7c410327] [ 611.140536] env[62066]: DEBUG nova.compute.utils [None req-1e3e30d6-8bb5-4347-b6c2-c2d648c37b67 tempest-ImagesOneServerNegativeTestJSON-310236551 tempest-ImagesOneServerNegativeTestJSON-310236551-project-member] [instance: 194588b4-fe40-4286-8036-874a7c410327] Binding failed for port 2a2c9bbf-b369-4ea3-a866-79152380ed0b, please check neutron logs for more information. {{(pid=62066) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 611.140536] env[62066]: DEBUG oslo_concurrency.lockutils [None req-fc3cf2dd-2491-4541-a245-5eefc9a5c042 tempest-TenantUsagesTestJSON-970395509 tempest-TenantUsagesTestJSON-970395509-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.184s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 611.140933] env[62066]: DEBUG nova.compute.manager [None req-1e3e30d6-8bb5-4347-b6c2-c2d648c37b67 tempest-ImagesOneServerNegativeTestJSON-310236551 tempest-ImagesOneServerNegativeTestJSON-310236551-project-member] [instance: 194588b4-fe40-4286-8036-874a7c410327] Build of instance 194588b4-fe40-4286-8036-874a7c410327 was re-scheduled: Binding failed for port 2a2c9bbf-b369-4ea3-a866-79152380ed0b, please check neutron logs for more information. {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 611.141148] env[62066]: DEBUG nova.compute.manager [None req-1e3e30d6-8bb5-4347-b6c2-c2d648c37b67 tempest-ImagesOneServerNegativeTestJSON-310236551 tempest-ImagesOneServerNegativeTestJSON-310236551-project-member] [instance: 194588b4-fe40-4286-8036-874a7c410327] Unplugging VIFs for instance {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 611.141374] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1e3e30d6-8bb5-4347-b6c2-c2d648c37b67 tempest-ImagesOneServerNegativeTestJSON-310236551 tempest-ImagesOneServerNegativeTestJSON-310236551-project-member] Acquiring lock "refresh_cache-194588b4-fe40-4286-8036-874a7c410327" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 611.141521] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1e3e30d6-8bb5-4347-b6c2-c2d648c37b67 tempest-ImagesOneServerNegativeTestJSON-310236551 tempest-ImagesOneServerNegativeTestJSON-310236551-project-member] Acquired lock "refresh_cache-194588b4-fe40-4286-8036-874a7c410327" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 611.141680] env[62066]: DEBUG nova.network.neutron [None req-1e3e30d6-8bb5-4347-b6c2-c2d648c37b67 tempest-ImagesOneServerNegativeTestJSON-310236551 tempest-ImagesOneServerNegativeTestJSON-310236551-project-member] [instance: 194588b4-fe40-4286-8036-874a7c410327] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 611.155644] env[62066]: ERROR nova.compute.manager [None req-14085d2f-7e58-404a-a2d0-4487ac658e45 tempest-FloatingIPsAssociationTestJSON-608653087 tempest-FloatingIPsAssociationTestJSON-608653087-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port d641e931-fa53-4d0a-ad67-1b5304b8fbee, please check neutron logs for more information. [ 611.155644] env[62066]: ERROR nova.compute.manager Traceback (most recent call last): [ 611.155644] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 611.155644] env[62066]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 611.155644] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 611.155644] env[62066]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 611.155644] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 611.155644] env[62066]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 611.155644] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 611.155644] env[62066]: ERROR nova.compute.manager self.force_reraise() [ 611.155644] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 611.155644] env[62066]: ERROR nova.compute.manager raise self.value [ 611.155644] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 611.155644] env[62066]: ERROR nova.compute.manager updated_port = self._update_port( [ 611.155644] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 611.155644] env[62066]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 611.156329] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 611.156329] env[62066]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 611.156329] env[62066]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port d641e931-fa53-4d0a-ad67-1b5304b8fbee, please check neutron logs for more information. [ 611.156329] env[62066]: ERROR nova.compute.manager [ 611.156329] env[62066]: Traceback (most recent call last): [ 611.156329] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 611.156329] env[62066]: listener.cb(fileno) [ 611.156329] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 611.156329] env[62066]: result = function(*args, **kwargs) [ 611.156329] env[62066]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 611.156329] env[62066]: return func(*args, **kwargs) [ 611.156329] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 611.156329] env[62066]: raise e [ 611.156329] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 611.156329] env[62066]: nwinfo = self.network_api.allocate_for_instance( [ 611.156329] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 611.156329] env[62066]: created_port_ids = self._update_ports_for_instance( [ 611.156329] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 611.156329] env[62066]: with excutils.save_and_reraise_exception(): [ 611.156329] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 611.156329] env[62066]: self.force_reraise() [ 611.156329] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 611.156329] env[62066]: raise self.value [ 611.156329] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 611.156329] env[62066]: updated_port = self._update_port( [ 611.156329] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 611.156329] env[62066]: _ensure_no_port_binding_failure(port) [ 611.156329] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 611.156329] env[62066]: raise exception.PortBindingFailed(port_id=port['id']) [ 611.157314] env[62066]: nova.exception.PortBindingFailed: Binding failed for port d641e931-fa53-4d0a-ad67-1b5304b8fbee, please check neutron logs for more information. [ 611.157314] env[62066]: Removing descriptor: 15 [ 611.157314] env[62066]: ERROR nova.compute.manager [None req-14085d2f-7e58-404a-a2d0-4487ac658e45 tempest-FloatingIPsAssociationTestJSON-608653087 tempest-FloatingIPsAssociationTestJSON-608653087-project-member] [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port d641e931-fa53-4d0a-ad67-1b5304b8fbee, please check neutron logs for more information. [ 611.157314] env[62066]: ERROR nova.compute.manager [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] Traceback (most recent call last): [ 611.157314] env[62066]: ERROR nova.compute.manager [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 611.157314] env[62066]: ERROR nova.compute.manager [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] yield resources [ 611.157314] env[62066]: ERROR nova.compute.manager [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 611.157314] env[62066]: ERROR nova.compute.manager [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] self.driver.spawn(context, instance, image_meta, [ 611.157314] env[62066]: ERROR nova.compute.manager [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 611.157314] env[62066]: ERROR nova.compute.manager [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] self._vmops.spawn(context, instance, image_meta, injected_files, [ 611.157314] env[62066]: ERROR nova.compute.manager [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 611.157314] env[62066]: ERROR nova.compute.manager [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] vm_ref = self.build_virtual_machine(instance, [ 611.157622] env[62066]: ERROR nova.compute.manager [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 611.157622] env[62066]: ERROR nova.compute.manager [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] vif_infos = vmwarevif.get_vif_info(self._session, [ 611.157622] env[62066]: ERROR nova.compute.manager [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 611.157622] env[62066]: ERROR nova.compute.manager [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] for vif in network_info: [ 611.157622] env[62066]: ERROR nova.compute.manager [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 611.157622] env[62066]: ERROR nova.compute.manager [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] return self._sync_wrapper(fn, *args, **kwargs) [ 611.157622] env[62066]: ERROR nova.compute.manager [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 611.157622] env[62066]: ERROR nova.compute.manager [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] self.wait() [ 611.157622] env[62066]: ERROR nova.compute.manager [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 611.157622] env[62066]: ERROR nova.compute.manager [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] self[:] = self._gt.wait() [ 611.157622] env[62066]: ERROR nova.compute.manager [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 611.157622] env[62066]: ERROR nova.compute.manager [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] return self._exit_event.wait() [ 611.157622] env[62066]: ERROR nova.compute.manager [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 611.157932] env[62066]: ERROR nova.compute.manager [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] result = hub.switch() [ 611.157932] env[62066]: ERROR nova.compute.manager [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 611.157932] env[62066]: ERROR nova.compute.manager [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] return self.greenlet.switch() [ 611.157932] env[62066]: ERROR nova.compute.manager [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 611.157932] env[62066]: ERROR nova.compute.manager [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] result = function(*args, **kwargs) [ 611.157932] env[62066]: ERROR nova.compute.manager [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 611.157932] env[62066]: ERROR nova.compute.manager [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] return func(*args, **kwargs) [ 611.157932] env[62066]: ERROR nova.compute.manager [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 611.157932] env[62066]: ERROR nova.compute.manager [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] raise e [ 611.157932] env[62066]: ERROR nova.compute.manager [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 611.157932] env[62066]: ERROR nova.compute.manager [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] nwinfo = self.network_api.allocate_for_instance( [ 611.157932] env[62066]: ERROR nova.compute.manager [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 611.157932] env[62066]: ERROR nova.compute.manager [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] created_port_ids = self._update_ports_for_instance( [ 611.158249] env[62066]: ERROR nova.compute.manager [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 611.158249] env[62066]: ERROR nova.compute.manager [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] with excutils.save_and_reraise_exception(): [ 611.158249] env[62066]: ERROR nova.compute.manager [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 611.158249] env[62066]: ERROR nova.compute.manager [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] self.force_reraise() [ 611.158249] env[62066]: ERROR nova.compute.manager [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 611.158249] env[62066]: ERROR nova.compute.manager [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] raise self.value [ 611.158249] env[62066]: ERROR nova.compute.manager [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 611.158249] env[62066]: ERROR nova.compute.manager [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] updated_port = self._update_port( [ 611.158249] env[62066]: ERROR nova.compute.manager [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 611.158249] env[62066]: ERROR nova.compute.manager [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] _ensure_no_port_binding_failure(port) [ 611.158249] env[62066]: ERROR nova.compute.manager [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 611.158249] env[62066]: ERROR nova.compute.manager [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] raise exception.PortBindingFailed(port_id=port['id']) [ 611.158586] env[62066]: ERROR nova.compute.manager [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] nova.exception.PortBindingFailed: Binding failed for port d641e931-fa53-4d0a-ad67-1b5304b8fbee, please check neutron logs for more information. [ 611.158586] env[62066]: ERROR nova.compute.manager [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] [ 611.158586] env[62066]: INFO nova.compute.manager [None req-14085d2f-7e58-404a-a2d0-4487ac658e45 tempest-FloatingIPsAssociationTestJSON-608653087 tempest-FloatingIPsAssociationTestJSON-608653087-project-member] [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] Terminating instance [ 611.159683] env[62066]: DEBUG oslo_concurrency.lockutils [None req-14085d2f-7e58-404a-a2d0-4487ac658e45 tempest-FloatingIPsAssociationTestJSON-608653087 tempest-FloatingIPsAssociationTestJSON-608653087-project-member] Acquiring lock "refresh_cache-08d41411-7928-4379-9f2e-c6ce00843d82" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 611.311101] env[62066]: DEBUG nova.network.neutron [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 611.387357] env[62066]: DEBUG nova.network.neutron [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 611.409865] env[62066]: DEBUG nova.network.neutron [req-14aaf419-0bbe-4bbc-b05a-c505245add01 req-3c1510af-04c1-48f8-b8ff-2e0c5b220b07 service nova] [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 611.490055] env[62066]: DEBUG nova.network.neutron [req-14aaf419-0bbe-4bbc-b05a-c505245add01 req-3c1510af-04c1-48f8-b8ff-2e0c5b220b07 service nova] [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 611.670833] env[62066]: DEBUG nova.network.neutron [None req-1e3e30d6-8bb5-4347-b6c2-c2d648c37b67 tempest-ImagesOneServerNegativeTestJSON-310236551 tempest-ImagesOneServerNegativeTestJSON-310236551-project-member] [instance: 194588b4-fe40-4286-8036-874a7c410327] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 611.733552] env[62066]: DEBUG nova.compute.manager [req-00685ee1-e3a5-47f0-b2f4-853f60b9dcbf req-5f983be0-874e-40d4-9615-94631eeae2ff service nova] [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] Received event network-vif-deleted-fd8616a9-d9b8-42e8-b0d3-4ca9121e33e7 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 611.834627] env[62066]: DEBUG nova.network.neutron [None req-1e3e30d6-8bb5-4347-b6c2-c2d648c37b67 tempest-ImagesOneServerNegativeTestJSON-310236551 tempest-ImagesOneServerNegativeTestJSON-310236551-project-member] [instance: 194588b4-fe40-4286-8036-874a7c410327] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 611.890050] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Releasing lock "refresh_cache-72c2dbe0-78de-4a7d-98df-cf405a5f20e2" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 611.890468] env[62066]: DEBUG nova.compute.manager [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 611.890661] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 611.893034] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e595bace-0a35-4e48-b99b-6a2167190a61 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.900956] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1746d0e-07ec-4dfa-af84-a2fbf40095ee {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.929000] env[62066]: WARNING nova.virt.vmwareapi.vmops [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 72c2dbe0-78de-4a7d-98df-cf405a5f20e2 could not be found. [ 611.929000] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 611.929000] env[62066]: INFO nova.compute.manager [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] Took 0.04 seconds to destroy the instance on the hypervisor. [ 611.929000] env[62066]: DEBUG oslo.service.loopingcall [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 611.929000] env[62066]: DEBUG nova.compute.manager [-] [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 611.929000] env[62066]: DEBUG nova.network.neutron [-] [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 611.953813] env[62066]: DEBUG nova.network.neutron [-] [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 611.993771] env[62066]: DEBUG oslo_concurrency.lockutils [req-14aaf419-0bbe-4bbc-b05a-c505245add01 req-3c1510af-04c1-48f8-b8ff-2e0c5b220b07 service nova] Releasing lock "refresh_cache-08d41411-7928-4379-9f2e-c6ce00843d82" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 611.993771] env[62066]: DEBUG oslo_concurrency.lockutils [None req-14085d2f-7e58-404a-a2d0-4487ac658e45 tempest-FloatingIPsAssociationTestJSON-608653087 tempest-FloatingIPsAssociationTestJSON-608653087-project-member] Acquired lock "refresh_cache-08d41411-7928-4379-9f2e-c6ce00843d82" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 611.993771] env[62066]: DEBUG nova.network.neutron [None req-14085d2f-7e58-404a-a2d0-4487ac658e45 tempest-FloatingIPsAssociationTestJSON-608653087 tempest-FloatingIPsAssociationTestJSON-608653087-project-member] [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 612.136148] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c175a2b-2af9-4acb-a8a6-61d98ed820b1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.147011] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e75ebe88-4f7f-43e5-8889-f8ac1ee608f4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.180291] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f25834b7-eb07-42f5-8d0f-18b8ab29bc15 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.187547] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8acf39a-4bd5-43c4-a09c-e675d89841a2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.200534] env[62066]: DEBUG nova.compute.provider_tree [None req-fc3cf2dd-2491-4541-a245-5eefc9a5c042 tempest-TenantUsagesTestJSON-970395509 tempest-TenantUsagesTestJSON-970395509-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 612.338193] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1e3e30d6-8bb5-4347-b6c2-c2d648c37b67 tempest-ImagesOneServerNegativeTestJSON-310236551 tempest-ImagesOneServerNegativeTestJSON-310236551-project-member] Releasing lock "refresh_cache-194588b4-fe40-4286-8036-874a7c410327" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 612.338486] env[62066]: DEBUG nova.compute.manager [None req-1e3e30d6-8bb5-4347-b6c2-c2d648c37b67 tempest-ImagesOneServerNegativeTestJSON-310236551 tempest-ImagesOneServerNegativeTestJSON-310236551-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 612.338688] env[62066]: DEBUG nova.compute.manager [None req-1e3e30d6-8bb5-4347-b6c2-c2d648c37b67 tempest-ImagesOneServerNegativeTestJSON-310236551 tempest-ImagesOneServerNegativeTestJSON-310236551-project-member] [instance: 194588b4-fe40-4286-8036-874a7c410327] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 612.338859] env[62066]: DEBUG nova.network.neutron [None req-1e3e30d6-8bb5-4347-b6c2-c2d648c37b67 tempest-ImagesOneServerNegativeTestJSON-310236551 tempest-ImagesOneServerNegativeTestJSON-310236551-project-member] [instance: 194588b4-fe40-4286-8036-874a7c410327] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 612.359101] env[62066]: DEBUG nova.network.neutron [None req-1e3e30d6-8bb5-4347-b6c2-c2d648c37b67 tempest-ImagesOneServerNegativeTestJSON-310236551 tempest-ImagesOneServerNegativeTestJSON-310236551-project-member] [instance: 194588b4-fe40-4286-8036-874a7c410327] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 612.455852] env[62066]: DEBUG nova.network.neutron [-] [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 612.514773] env[62066]: DEBUG nova.network.neutron [None req-14085d2f-7e58-404a-a2d0-4487ac658e45 tempest-FloatingIPsAssociationTestJSON-608653087 tempest-FloatingIPsAssociationTestJSON-608653087-project-member] [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 612.621209] env[62066]: DEBUG nova.network.neutron [None req-14085d2f-7e58-404a-a2d0-4487ac658e45 tempest-FloatingIPsAssociationTestJSON-608653087 tempest-FloatingIPsAssociationTestJSON-608653087-project-member] [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 612.704104] env[62066]: DEBUG nova.scheduler.client.report [None req-fc3cf2dd-2491-4541-a245-5eefc9a5c042 tempest-TenantUsagesTestJSON-970395509 tempest-TenantUsagesTestJSON-970395509-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 612.862184] env[62066]: DEBUG nova.network.neutron [None req-1e3e30d6-8bb5-4347-b6c2-c2d648c37b67 tempest-ImagesOneServerNegativeTestJSON-310236551 tempest-ImagesOneServerNegativeTestJSON-310236551-project-member] [instance: 194588b4-fe40-4286-8036-874a7c410327] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 612.959742] env[62066]: INFO nova.compute.manager [-] [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] Took 1.03 seconds to deallocate network for instance. [ 612.961928] env[62066]: DEBUG nova.compute.claims [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] Aborting claim: {{(pid=62066) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 612.963036] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 612.991881] env[62066]: DEBUG nova.compute.manager [req-4040e0be-337e-4948-a2b7-7eff2e6c7063 req-f7fb60ba-644d-4306-8b95-06bf8439bb36 service nova] [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] Received event network-vif-deleted-d641e931-fa53-4d0a-ad67-1b5304b8fbee {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 613.124025] env[62066]: DEBUG oslo_concurrency.lockutils [None req-14085d2f-7e58-404a-a2d0-4487ac658e45 tempest-FloatingIPsAssociationTestJSON-608653087 tempest-FloatingIPsAssociationTestJSON-608653087-project-member] Releasing lock "refresh_cache-08d41411-7928-4379-9f2e-c6ce00843d82" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 613.124476] env[62066]: DEBUG nova.compute.manager [None req-14085d2f-7e58-404a-a2d0-4487ac658e45 tempest-FloatingIPsAssociationTestJSON-608653087 tempest-FloatingIPsAssociationTestJSON-608653087-project-member] [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 613.124676] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-14085d2f-7e58-404a-a2d0-4487ac658e45 tempest-FloatingIPsAssociationTestJSON-608653087 tempest-FloatingIPsAssociationTestJSON-608653087-project-member] [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 613.124971] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-22ee489d-d29d-4b50-91f4-a1bb9ff083d1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.135599] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26d456dc-18d5-4ecf-a058-96e2df388985 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.158036] env[62066]: WARNING nova.virt.vmwareapi.vmops [None req-14085d2f-7e58-404a-a2d0-4487ac658e45 tempest-FloatingIPsAssociationTestJSON-608653087 tempest-FloatingIPsAssociationTestJSON-608653087-project-member] [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 08d41411-7928-4379-9f2e-c6ce00843d82 could not be found. [ 613.158293] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-14085d2f-7e58-404a-a2d0-4487ac658e45 tempest-FloatingIPsAssociationTestJSON-608653087 tempest-FloatingIPsAssociationTestJSON-608653087-project-member] [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 613.158528] env[62066]: INFO nova.compute.manager [None req-14085d2f-7e58-404a-a2d0-4487ac658e45 tempest-FloatingIPsAssociationTestJSON-608653087 tempest-FloatingIPsAssociationTestJSON-608653087-project-member] [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] Took 0.03 seconds to destroy the instance on the hypervisor. [ 613.158711] env[62066]: DEBUG oslo.service.loopingcall [None req-14085d2f-7e58-404a-a2d0-4487ac658e45 tempest-FloatingIPsAssociationTestJSON-608653087 tempest-FloatingIPsAssociationTestJSON-608653087-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 613.158937] env[62066]: DEBUG nova.compute.manager [-] [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 613.159045] env[62066]: DEBUG nova.network.neutron [-] [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 613.174253] env[62066]: DEBUG nova.network.neutron [-] [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 613.212042] env[62066]: DEBUG oslo_concurrency.lockutils [None req-fc3cf2dd-2491-4541-a245-5eefc9a5c042 tempest-TenantUsagesTestJSON-970395509 tempest-TenantUsagesTestJSON-970395509-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.074s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 613.212708] env[62066]: ERROR nova.compute.manager [None req-fc3cf2dd-2491-4541-a245-5eefc9a5c042 tempest-TenantUsagesTestJSON-970395509 tempest-TenantUsagesTestJSON-970395509-project-member] [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 601feca8-30e1-4972-a3b1-25d0817e7e50, please check neutron logs for more information. [ 613.212708] env[62066]: ERROR nova.compute.manager [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] Traceback (most recent call last): [ 613.212708] env[62066]: ERROR nova.compute.manager [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 613.212708] env[62066]: ERROR nova.compute.manager [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] self.driver.spawn(context, instance, image_meta, [ 613.212708] env[62066]: ERROR nova.compute.manager [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 613.212708] env[62066]: ERROR nova.compute.manager [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] self._vmops.spawn(context, instance, image_meta, injected_files, [ 613.212708] env[62066]: ERROR nova.compute.manager [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 613.212708] env[62066]: ERROR nova.compute.manager [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] vm_ref = self.build_virtual_machine(instance, [ 613.212708] env[62066]: ERROR nova.compute.manager [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 613.212708] env[62066]: ERROR nova.compute.manager [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] vif_infos = vmwarevif.get_vif_info(self._session, [ 613.212708] env[62066]: ERROR nova.compute.manager [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 613.213193] env[62066]: ERROR nova.compute.manager [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] for vif in network_info: [ 613.213193] env[62066]: ERROR nova.compute.manager [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 613.213193] env[62066]: ERROR nova.compute.manager [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] return self._sync_wrapper(fn, *args, **kwargs) [ 613.213193] env[62066]: ERROR nova.compute.manager [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 613.213193] env[62066]: ERROR nova.compute.manager [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] self.wait() [ 613.213193] env[62066]: ERROR nova.compute.manager [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 613.213193] env[62066]: ERROR nova.compute.manager [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] self[:] = self._gt.wait() [ 613.213193] env[62066]: ERROR nova.compute.manager [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 613.213193] env[62066]: ERROR nova.compute.manager [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] return self._exit_event.wait() [ 613.213193] env[62066]: ERROR nova.compute.manager [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 613.213193] env[62066]: ERROR nova.compute.manager [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] result = hub.switch() [ 613.213193] env[62066]: ERROR nova.compute.manager [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 613.213193] env[62066]: ERROR nova.compute.manager [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] return self.greenlet.switch() [ 613.213708] env[62066]: ERROR nova.compute.manager [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 613.213708] env[62066]: ERROR nova.compute.manager [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] result = function(*args, **kwargs) [ 613.213708] env[62066]: ERROR nova.compute.manager [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 613.213708] env[62066]: ERROR nova.compute.manager [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] return func(*args, **kwargs) [ 613.213708] env[62066]: ERROR nova.compute.manager [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 613.213708] env[62066]: ERROR nova.compute.manager [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] raise e [ 613.213708] env[62066]: ERROR nova.compute.manager [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 613.213708] env[62066]: ERROR nova.compute.manager [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] nwinfo = self.network_api.allocate_for_instance( [ 613.213708] env[62066]: ERROR nova.compute.manager [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 613.213708] env[62066]: ERROR nova.compute.manager [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] created_port_ids = self._update_ports_for_instance( [ 613.213708] env[62066]: ERROR nova.compute.manager [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 613.213708] env[62066]: ERROR nova.compute.manager [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] with excutils.save_and_reraise_exception(): [ 613.213708] env[62066]: ERROR nova.compute.manager [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 613.214227] env[62066]: ERROR nova.compute.manager [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] self.force_reraise() [ 613.214227] env[62066]: ERROR nova.compute.manager [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 613.214227] env[62066]: ERROR nova.compute.manager [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] raise self.value [ 613.214227] env[62066]: ERROR nova.compute.manager [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 613.214227] env[62066]: ERROR nova.compute.manager [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] updated_port = self._update_port( [ 613.214227] env[62066]: ERROR nova.compute.manager [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 613.214227] env[62066]: ERROR nova.compute.manager [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] _ensure_no_port_binding_failure(port) [ 613.214227] env[62066]: ERROR nova.compute.manager [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 613.214227] env[62066]: ERROR nova.compute.manager [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] raise exception.PortBindingFailed(port_id=port['id']) [ 613.214227] env[62066]: ERROR nova.compute.manager [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] nova.exception.PortBindingFailed: Binding failed for port 601feca8-30e1-4972-a3b1-25d0817e7e50, please check neutron logs for more information. [ 613.214227] env[62066]: ERROR nova.compute.manager [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] [ 613.214667] env[62066]: DEBUG nova.compute.utils [None req-fc3cf2dd-2491-4541-a245-5eefc9a5c042 tempest-TenantUsagesTestJSON-970395509 tempest-TenantUsagesTestJSON-970395509-project-member] [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] Binding failed for port 601feca8-30e1-4972-a3b1-25d0817e7e50, please check neutron logs for more information. {{(pid=62066) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 613.215131] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a9ebb4d2-ff8d-4333-a9df-86618ea2d54c tempest-ImagesOneServerTestJSON-1591819949 tempest-ImagesOneServerTestJSON-1591819949-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.252s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 613.218705] env[62066]: INFO nova.compute.claims [None req-a9ebb4d2-ff8d-4333-a9df-86618ea2d54c tempest-ImagesOneServerTestJSON-1591819949 tempest-ImagesOneServerTestJSON-1591819949-project-member] [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 613.219927] env[62066]: DEBUG nova.compute.manager [None req-fc3cf2dd-2491-4541-a245-5eefc9a5c042 tempest-TenantUsagesTestJSON-970395509 tempest-TenantUsagesTestJSON-970395509-project-member] [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] Build of instance a187f73b-8fb1-42b3-9a07-24d16aa0f152 was re-scheduled: Binding failed for port 601feca8-30e1-4972-a3b1-25d0817e7e50, please check neutron logs for more information. {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 613.220409] env[62066]: DEBUG nova.compute.manager [None req-fc3cf2dd-2491-4541-a245-5eefc9a5c042 tempest-TenantUsagesTestJSON-970395509 tempest-TenantUsagesTestJSON-970395509-project-member] [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] Unplugging VIFs for instance {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 613.220637] env[62066]: DEBUG oslo_concurrency.lockutils [None req-fc3cf2dd-2491-4541-a245-5eefc9a5c042 tempest-TenantUsagesTestJSON-970395509 tempest-TenantUsagesTestJSON-970395509-project-member] Acquiring lock "refresh_cache-a187f73b-8fb1-42b3-9a07-24d16aa0f152" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 613.220784] env[62066]: DEBUG oslo_concurrency.lockutils [None req-fc3cf2dd-2491-4541-a245-5eefc9a5c042 tempest-TenantUsagesTestJSON-970395509 tempest-TenantUsagesTestJSON-970395509-project-member] Acquired lock "refresh_cache-a187f73b-8fb1-42b3-9a07-24d16aa0f152" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 613.220942] env[62066]: DEBUG nova.network.neutron [None req-fc3cf2dd-2491-4541-a245-5eefc9a5c042 tempest-TenantUsagesTestJSON-970395509 tempest-TenantUsagesTestJSON-970395509-project-member] [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 613.368778] env[62066]: INFO nova.compute.manager [None req-1e3e30d6-8bb5-4347-b6c2-c2d648c37b67 tempest-ImagesOneServerNegativeTestJSON-310236551 tempest-ImagesOneServerNegativeTestJSON-310236551-project-member] [instance: 194588b4-fe40-4286-8036-874a7c410327] Took 1.03 seconds to deallocate network for instance. [ 613.679232] env[62066]: DEBUG nova.network.neutron [-] [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 613.751155] env[62066]: DEBUG nova.network.neutron [None req-fc3cf2dd-2491-4541-a245-5eefc9a5c042 tempest-TenantUsagesTestJSON-970395509 tempest-TenantUsagesTestJSON-970395509-project-member] [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 613.846106] env[62066]: DEBUG nova.network.neutron [None req-fc3cf2dd-2491-4541-a245-5eefc9a5c042 tempest-TenantUsagesTestJSON-970395509 tempest-TenantUsagesTestJSON-970395509-project-member] [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 614.182350] env[62066]: INFO nova.compute.manager [-] [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] Took 1.02 seconds to deallocate network for instance. [ 614.187388] env[62066]: DEBUG nova.compute.claims [None req-14085d2f-7e58-404a-a2d0-4487ac658e45 tempest-FloatingIPsAssociationTestJSON-608653087 tempest-FloatingIPsAssociationTestJSON-608653087-project-member] [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] Aborting claim: {{(pid=62066) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 614.187470] env[62066]: DEBUG oslo_concurrency.lockutils [None req-14085d2f-7e58-404a-a2d0-4487ac658e45 tempest-FloatingIPsAssociationTestJSON-608653087 tempest-FloatingIPsAssociationTestJSON-608653087-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 614.229409] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e9793203-98b3-4ca0-8a49-79767b1f52f7 tempest-ServersTestBootFromVolume-1535558002 tempest-ServersTestBootFromVolume-1535558002-project-member] Acquiring lock "2b81d060-ecb7-4ae5-b51d-33197937a76f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 614.229657] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e9793203-98b3-4ca0-8a49-79767b1f52f7 tempest-ServersTestBootFromVolume-1535558002 tempest-ServersTestBootFromVolume-1535558002-project-member] Lock "2b81d060-ecb7-4ae5-b51d-33197937a76f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 614.349838] env[62066]: DEBUG oslo_concurrency.lockutils [None req-fc3cf2dd-2491-4541-a245-5eefc9a5c042 tempest-TenantUsagesTestJSON-970395509 tempest-TenantUsagesTestJSON-970395509-project-member] Releasing lock "refresh_cache-a187f73b-8fb1-42b3-9a07-24d16aa0f152" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 614.350101] env[62066]: DEBUG nova.compute.manager [None req-fc3cf2dd-2491-4541-a245-5eefc9a5c042 tempest-TenantUsagesTestJSON-970395509 tempest-TenantUsagesTestJSON-970395509-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 614.350274] env[62066]: DEBUG nova.compute.manager [None req-fc3cf2dd-2491-4541-a245-5eefc9a5c042 tempest-TenantUsagesTestJSON-970395509 tempest-TenantUsagesTestJSON-970395509-project-member] [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 614.350441] env[62066]: DEBUG nova.network.neutron [None req-fc3cf2dd-2491-4541-a245-5eefc9a5c042 tempest-TenantUsagesTestJSON-970395509 tempest-TenantUsagesTestJSON-970395509-project-member] [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 614.369533] env[62066]: DEBUG nova.network.neutron [None req-fc3cf2dd-2491-4541-a245-5eefc9a5c042 tempest-TenantUsagesTestJSON-970395509 tempest-TenantUsagesTestJSON-970395509-project-member] [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 614.409166] env[62066]: INFO nova.scheduler.client.report [None req-1e3e30d6-8bb5-4347-b6c2-c2d648c37b67 tempest-ImagesOneServerNegativeTestJSON-310236551 tempest-ImagesOneServerNegativeTestJSON-310236551-project-member] Deleted allocations for instance 194588b4-fe40-4286-8036-874a7c410327 [ 614.584901] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c43ec0ba-b44f-4c46-9718-82269c41be56 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.593754] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-994e6c28-e642-49be-a406-416cd52ac675 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.626038] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48fc78d1-f266-4f9a-b28b-7178259492e6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.634597] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee643925-bbce-454e-b52f-07734d110e46 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.647836] env[62066]: DEBUG nova.compute.provider_tree [None req-a9ebb4d2-ff8d-4333-a9df-86618ea2d54c tempest-ImagesOneServerTestJSON-1591819949 tempest-ImagesOneServerTestJSON-1591819949-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 614.872756] env[62066]: DEBUG nova.network.neutron [None req-fc3cf2dd-2491-4541-a245-5eefc9a5c042 tempest-TenantUsagesTestJSON-970395509 tempest-TenantUsagesTestJSON-970395509-project-member] [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 614.921832] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1e3e30d6-8bb5-4347-b6c2-c2d648c37b67 tempest-ImagesOneServerNegativeTestJSON-310236551 tempest-ImagesOneServerNegativeTestJSON-310236551-project-member] Lock "194588b4-fe40-4286-8036-874a7c410327" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 99.238s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 615.152504] env[62066]: DEBUG nova.scheduler.client.report [None req-a9ebb4d2-ff8d-4333-a9df-86618ea2d54c tempest-ImagesOneServerTestJSON-1591819949 tempest-ImagesOneServerTestJSON-1591819949-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 615.374321] env[62066]: INFO nova.compute.manager [None req-fc3cf2dd-2491-4541-a245-5eefc9a5c042 tempest-TenantUsagesTestJSON-970395509 tempest-TenantUsagesTestJSON-970395509-project-member] [instance: a187f73b-8fb1-42b3-9a07-24d16aa0f152] Took 1.02 seconds to deallocate network for instance. [ 615.424364] env[62066]: DEBUG nova.compute.manager [None req-9fdb5839-7f19-4e7f-8faa-852358a443a9 tempest-ServersListShow296Test-1792957209 tempest-ServersListShow296Test-1792957209-project-member] [instance: b68fe0ac-732c-448c-8bae-2dcd0ce2dc73] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 615.663886] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a9ebb4d2-ff8d-4333-a9df-86618ea2d54c tempest-ImagesOneServerTestJSON-1591819949 tempest-ImagesOneServerTestJSON-1591819949-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.449s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 615.664478] env[62066]: DEBUG nova.compute.manager [None req-a9ebb4d2-ff8d-4333-a9df-86618ea2d54c tempest-ImagesOneServerTestJSON-1591819949 tempest-ImagesOneServerTestJSON-1591819949-project-member] [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 615.667635] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e7221104-9243-4d76-b03e-5782ae8f55d4 tempest-FloatingIPsAssociationNegativeTestJSON-961065153 tempest-FloatingIPsAssociationNegativeTestJSON-961065153-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.027s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 615.933018] env[62066]: DEBUG nova.compute.manager [None req-9fdb5839-7f19-4e7f-8faa-852358a443a9 tempest-ServersListShow296Test-1792957209 tempest-ServersListShow296Test-1792957209-project-member] [instance: b68fe0ac-732c-448c-8bae-2dcd0ce2dc73] Instance disappeared before build. {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 616.172958] env[62066]: DEBUG nova.compute.utils [None req-a9ebb4d2-ff8d-4333-a9df-86618ea2d54c tempest-ImagesOneServerTestJSON-1591819949 tempest-ImagesOneServerTestJSON-1591819949-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 616.179994] env[62066]: DEBUG nova.compute.manager [None req-a9ebb4d2-ff8d-4333-a9df-86618ea2d54c tempest-ImagesOneServerTestJSON-1591819949 tempest-ImagesOneServerTestJSON-1591819949-project-member] [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 616.181105] env[62066]: DEBUG nova.network.neutron [None req-a9ebb4d2-ff8d-4333-a9df-86618ea2d54c tempest-ImagesOneServerTestJSON-1591819949 tempest-ImagesOneServerTestJSON-1591819949-project-member] [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 616.254476] env[62066]: DEBUG nova.policy [None req-a9ebb4d2-ff8d-4333-a9df-86618ea2d54c tempest-ImagesOneServerTestJSON-1591819949 tempest-ImagesOneServerTestJSON-1591819949-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '068eb8e7562f49dd97647e553e2e362d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '774e816358d24b2aad74a476ab83f95d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 616.408342] env[62066]: INFO nova.scheduler.client.report [None req-fc3cf2dd-2491-4541-a245-5eefc9a5c042 tempest-TenantUsagesTestJSON-970395509 tempest-TenantUsagesTestJSON-970395509-project-member] Deleted allocations for instance a187f73b-8fb1-42b3-9a07-24d16aa0f152 [ 616.450218] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9fdb5839-7f19-4e7f-8faa-852358a443a9 tempest-ServersListShow296Test-1792957209 tempest-ServersListShow296Test-1792957209-project-member] Lock "b68fe0ac-732c-448c-8bae-2dcd0ce2dc73" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 81.421s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 616.601154] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10b277e1-2a9c-462e-9196-e7945a6851a6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.611097] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26260e10-c422-4459-afc5-72f9b8a8fc60 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.645600] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f827076-e895-4356-bc6f-d78204cfd8d9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.651735] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31739600-d733-4592-9119-c458bc212c4e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.667936] env[62066]: DEBUG nova.compute.provider_tree [None req-e7221104-9243-4d76-b03e-5782ae8f55d4 tempest-FloatingIPsAssociationNegativeTestJSON-961065153 tempest-FloatingIPsAssociationNegativeTestJSON-961065153-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 616.686730] env[62066]: DEBUG nova.compute.manager [None req-a9ebb4d2-ff8d-4333-a9df-86618ea2d54c tempest-ImagesOneServerTestJSON-1591819949 tempest-ImagesOneServerTestJSON-1591819949-project-member] [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 616.742487] env[62066]: DEBUG nova.network.neutron [None req-a9ebb4d2-ff8d-4333-a9df-86618ea2d54c tempest-ImagesOneServerTestJSON-1591819949 tempest-ImagesOneServerTestJSON-1591819949-project-member] [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] Successfully created port: e09cd9e1-44fe-4858-bf9a-0afd7b681ca4 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 616.919250] env[62066]: DEBUG oslo_concurrency.lockutils [None req-fc3cf2dd-2491-4541-a245-5eefc9a5c042 tempest-TenantUsagesTestJSON-970395509 tempest-TenantUsagesTestJSON-970395509-project-member] Lock "a187f73b-8fb1-42b3-9a07-24d16aa0f152" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 94.870s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 616.955286] env[62066]: DEBUG nova.compute.manager [None req-5e085536-daab-46ff-8eb8-69c1da687135 tempest-ServersAaction247Test-1482016775 tempest-ServersAaction247Test-1482016775-project-member] [instance: 5d54c3fd-b239-4965-b7a3-5909e8de8bc0] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 617.171581] env[62066]: DEBUG nova.scheduler.client.report [None req-e7221104-9243-4d76-b03e-5782ae8f55d4 tempest-FloatingIPsAssociationNegativeTestJSON-961065153 tempest-FloatingIPsAssociationNegativeTestJSON-961065153-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 617.423936] env[62066]: DEBUG nova.compute.manager [None req-2c107b45-3cbf-4be9-af5a-796fbdb9b5fe tempest-ImagesNegativeTestJSON-881899573 tempest-ImagesNegativeTestJSON-881899573-project-member] [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 617.481856] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5e085536-daab-46ff-8eb8-69c1da687135 tempest-ServersAaction247Test-1482016775 tempest-ServersAaction247Test-1482016775-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 617.678501] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e7221104-9243-4d76-b03e-5782ae8f55d4 tempest-FloatingIPsAssociationNegativeTestJSON-961065153 tempest-FloatingIPsAssociationNegativeTestJSON-961065153-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.010s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 617.678602] env[62066]: ERROR nova.compute.manager [None req-e7221104-9243-4d76-b03e-5782ae8f55d4 tempest-FloatingIPsAssociationNegativeTestJSON-961065153 tempest-FloatingIPsAssociationNegativeTestJSON-961065153-project-member] [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 648e8154-eb46-4556-b4ff-b8d2b55b2d10, please check neutron logs for more information. [ 617.678602] env[62066]: ERROR nova.compute.manager [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] Traceback (most recent call last): [ 617.678602] env[62066]: ERROR nova.compute.manager [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 617.678602] env[62066]: ERROR nova.compute.manager [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] self.driver.spawn(context, instance, image_meta, [ 617.678602] env[62066]: ERROR nova.compute.manager [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 617.678602] env[62066]: ERROR nova.compute.manager [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] self._vmops.spawn(context, instance, image_meta, injected_files, [ 617.678602] env[62066]: ERROR nova.compute.manager [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 617.678602] env[62066]: ERROR nova.compute.manager [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] vm_ref = self.build_virtual_machine(instance, [ 617.678602] env[62066]: ERROR nova.compute.manager [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 617.678602] env[62066]: ERROR nova.compute.manager [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] vif_infos = vmwarevif.get_vif_info(self._session, [ 617.678602] env[62066]: ERROR nova.compute.manager [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 617.679134] env[62066]: ERROR nova.compute.manager [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] for vif in network_info: [ 617.679134] env[62066]: ERROR nova.compute.manager [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 617.679134] env[62066]: ERROR nova.compute.manager [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] return self._sync_wrapper(fn, *args, **kwargs) [ 617.679134] env[62066]: ERROR nova.compute.manager [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 617.679134] env[62066]: ERROR nova.compute.manager [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] self.wait() [ 617.679134] env[62066]: ERROR nova.compute.manager [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 617.679134] env[62066]: ERROR nova.compute.manager [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] self[:] = self._gt.wait() [ 617.679134] env[62066]: ERROR nova.compute.manager [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 617.679134] env[62066]: ERROR nova.compute.manager [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] return self._exit_event.wait() [ 617.679134] env[62066]: ERROR nova.compute.manager [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 617.679134] env[62066]: ERROR nova.compute.manager [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] result = hub.switch() [ 617.679134] env[62066]: ERROR nova.compute.manager [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 617.679134] env[62066]: ERROR nova.compute.manager [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] return self.greenlet.switch() [ 617.680155] env[62066]: ERROR nova.compute.manager [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 617.680155] env[62066]: ERROR nova.compute.manager [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] result = function(*args, **kwargs) [ 617.680155] env[62066]: ERROR nova.compute.manager [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 617.680155] env[62066]: ERROR nova.compute.manager [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] return func(*args, **kwargs) [ 617.680155] env[62066]: ERROR nova.compute.manager [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 617.680155] env[62066]: ERROR nova.compute.manager [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] raise e [ 617.680155] env[62066]: ERROR nova.compute.manager [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 617.680155] env[62066]: ERROR nova.compute.manager [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] nwinfo = self.network_api.allocate_for_instance( [ 617.680155] env[62066]: ERROR nova.compute.manager [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 617.680155] env[62066]: ERROR nova.compute.manager [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] created_port_ids = self._update_ports_for_instance( [ 617.680155] env[62066]: ERROR nova.compute.manager [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 617.680155] env[62066]: ERROR nova.compute.manager [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] with excutils.save_and_reraise_exception(): [ 617.680155] env[62066]: ERROR nova.compute.manager [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 617.681170] env[62066]: ERROR nova.compute.manager [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] self.force_reraise() [ 617.681170] env[62066]: ERROR nova.compute.manager [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 617.681170] env[62066]: ERROR nova.compute.manager [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] raise self.value [ 617.681170] env[62066]: ERROR nova.compute.manager [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 617.681170] env[62066]: ERROR nova.compute.manager [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] updated_port = self._update_port( [ 617.681170] env[62066]: ERROR nova.compute.manager [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 617.681170] env[62066]: ERROR nova.compute.manager [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] _ensure_no_port_binding_failure(port) [ 617.681170] env[62066]: ERROR nova.compute.manager [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 617.681170] env[62066]: ERROR nova.compute.manager [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] raise exception.PortBindingFailed(port_id=port['id']) [ 617.681170] env[62066]: ERROR nova.compute.manager [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] nova.exception.PortBindingFailed: Binding failed for port 648e8154-eb46-4556-b4ff-b8d2b55b2d10, please check neutron logs for more information. [ 617.681170] env[62066]: ERROR nova.compute.manager [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] [ 617.681875] env[62066]: DEBUG nova.compute.utils [None req-e7221104-9243-4d76-b03e-5782ae8f55d4 tempest-FloatingIPsAssociationNegativeTestJSON-961065153 tempest-FloatingIPsAssociationNegativeTestJSON-961065153-project-member] [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] Binding failed for port 648e8154-eb46-4556-b4ff-b8d2b55b2d10, please check neutron logs for more information. {{(pid=62066) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 617.682269] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9371bde4-46e3-4272-a1fe-5125bccbd2c7 tempest-ServersTestJSON-680983284 tempest-ServersTestJSON-680983284-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.681s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 617.685274] env[62066]: INFO nova.compute.claims [None req-9371bde4-46e3-4272-a1fe-5125bccbd2c7 tempest-ServersTestJSON-680983284 tempest-ServersTestJSON-680983284-project-member] [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 617.688638] env[62066]: DEBUG nova.compute.manager [None req-e7221104-9243-4d76-b03e-5782ae8f55d4 tempest-FloatingIPsAssociationNegativeTestJSON-961065153 tempest-FloatingIPsAssociationNegativeTestJSON-961065153-project-member] [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] Build of instance eee6b9e6-2707-4a3b-925b-a7f50cdc32cb was re-scheduled: Binding failed for port 648e8154-eb46-4556-b4ff-b8d2b55b2d10, please check neutron logs for more information. {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 617.689411] env[62066]: DEBUG nova.compute.manager [None req-e7221104-9243-4d76-b03e-5782ae8f55d4 tempest-FloatingIPsAssociationNegativeTestJSON-961065153 tempest-FloatingIPsAssociationNegativeTestJSON-961065153-project-member] [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] Unplugging VIFs for instance {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 617.689477] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e7221104-9243-4d76-b03e-5782ae8f55d4 tempest-FloatingIPsAssociationNegativeTestJSON-961065153 tempest-FloatingIPsAssociationNegativeTestJSON-961065153-project-member] Acquiring lock "refresh_cache-eee6b9e6-2707-4a3b-925b-a7f50cdc32cb" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 617.692956] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e7221104-9243-4d76-b03e-5782ae8f55d4 tempest-FloatingIPsAssociationNegativeTestJSON-961065153 tempest-FloatingIPsAssociationNegativeTestJSON-961065153-project-member] Acquired lock "refresh_cache-eee6b9e6-2707-4a3b-925b-a7f50cdc32cb" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 617.693197] env[62066]: DEBUG nova.network.neutron [None req-e7221104-9243-4d76-b03e-5782ae8f55d4 tempest-FloatingIPsAssociationNegativeTestJSON-961065153 tempest-FloatingIPsAssociationNegativeTestJSON-961065153-project-member] [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 617.708528] env[62066]: DEBUG nova.compute.manager [None req-a9ebb4d2-ff8d-4333-a9df-86618ea2d54c tempest-ImagesOneServerTestJSON-1591819949 tempest-ImagesOneServerTestJSON-1591819949-project-member] [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 617.751062] env[62066]: DEBUG nova.virt.hardware [None req-a9ebb4d2-ff8d-4333-a9df-86618ea2d54c tempest-ImagesOneServerTestJSON-1591819949 tempest-ImagesOneServerTestJSON-1591819949-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-23T13:40:41Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-23T13:40:23Z,direct_url=,disk_format='vmdk',id=50ff584c-3b50-4395-af07-3e66769bc9f7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='eb52f7069a374c61ae946f052007c6d9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-23T13:40:24Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 617.751220] env[62066]: DEBUG nova.virt.hardware [None req-a9ebb4d2-ff8d-4333-a9df-86618ea2d54c tempest-ImagesOneServerTestJSON-1591819949 tempest-ImagesOneServerTestJSON-1591819949-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 617.751265] env[62066]: DEBUG nova.virt.hardware [None req-a9ebb4d2-ff8d-4333-a9df-86618ea2d54c tempest-ImagesOneServerTestJSON-1591819949 tempest-ImagesOneServerTestJSON-1591819949-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 617.751447] env[62066]: DEBUG nova.virt.hardware [None req-a9ebb4d2-ff8d-4333-a9df-86618ea2d54c tempest-ImagesOneServerTestJSON-1591819949 tempest-ImagesOneServerTestJSON-1591819949-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 617.751592] env[62066]: DEBUG nova.virt.hardware [None req-a9ebb4d2-ff8d-4333-a9df-86618ea2d54c tempest-ImagesOneServerTestJSON-1591819949 tempest-ImagesOneServerTestJSON-1591819949-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 617.751738] env[62066]: DEBUG nova.virt.hardware [None req-a9ebb4d2-ff8d-4333-a9df-86618ea2d54c tempest-ImagesOneServerTestJSON-1591819949 tempest-ImagesOneServerTestJSON-1591819949-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 617.751943] env[62066]: DEBUG nova.virt.hardware [None req-a9ebb4d2-ff8d-4333-a9df-86618ea2d54c tempest-ImagesOneServerTestJSON-1591819949 tempest-ImagesOneServerTestJSON-1591819949-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 617.752566] env[62066]: DEBUG nova.virt.hardware [None req-a9ebb4d2-ff8d-4333-a9df-86618ea2d54c tempest-ImagesOneServerTestJSON-1591819949 tempest-ImagesOneServerTestJSON-1591819949-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 617.752706] env[62066]: DEBUG nova.virt.hardware [None req-a9ebb4d2-ff8d-4333-a9df-86618ea2d54c tempest-ImagesOneServerTestJSON-1591819949 tempest-ImagesOneServerTestJSON-1591819949-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 617.753072] env[62066]: DEBUG nova.virt.hardware [None req-a9ebb4d2-ff8d-4333-a9df-86618ea2d54c tempest-ImagesOneServerTestJSON-1591819949 tempest-ImagesOneServerTestJSON-1591819949-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 617.753072] env[62066]: DEBUG nova.virt.hardware [None req-a9ebb4d2-ff8d-4333-a9df-86618ea2d54c tempest-ImagesOneServerTestJSON-1591819949 tempest-ImagesOneServerTestJSON-1591819949-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 617.754570] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb25a7eb-d954-4950-b03d-96efb66edf35 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.763554] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41a71ebb-3431-48cb-8a1a-6e0768303001 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.954105] env[62066]: DEBUG oslo_concurrency.lockutils [None req-2c107b45-3cbf-4be9-af5a-796fbdb9b5fe tempest-ImagesNegativeTestJSON-881899573 tempest-ImagesNegativeTestJSON-881899573-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 618.225482] env[62066]: DEBUG nova.network.neutron [None req-e7221104-9243-4d76-b03e-5782ae8f55d4 tempest-FloatingIPsAssociationNegativeTestJSON-961065153 tempest-FloatingIPsAssociationNegativeTestJSON-961065153-project-member] [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 618.282174] env[62066]: DEBUG nova.compute.manager [req-f9d54c92-c5fa-416b-b7c6-4136f944a934 req-acde99f0-cee9-4f02-8d75-87b0484b4272 service nova] [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] Received event network-changed-e09cd9e1-44fe-4858-bf9a-0afd7b681ca4 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 618.282372] env[62066]: DEBUG nova.compute.manager [req-f9d54c92-c5fa-416b-b7c6-4136f944a934 req-acde99f0-cee9-4f02-8d75-87b0484b4272 service nova] [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] Refreshing instance network info cache due to event network-changed-e09cd9e1-44fe-4858-bf9a-0afd7b681ca4. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 618.282692] env[62066]: DEBUG oslo_concurrency.lockutils [req-f9d54c92-c5fa-416b-b7c6-4136f944a934 req-acde99f0-cee9-4f02-8d75-87b0484b4272 service nova] Acquiring lock "refresh_cache-623300f7-54d2-4b0a-b356-a1ae585682cb" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 618.282832] env[62066]: DEBUG oslo_concurrency.lockutils [req-f9d54c92-c5fa-416b-b7c6-4136f944a934 req-acde99f0-cee9-4f02-8d75-87b0484b4272 service nova] Acquired lock "refresh_cache-623300f7-54d2-4b0a-b356-a1ae585682cb" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 618.282991] env[62066]: DEBUG nova.network.neutron [req-f9d54c92-c5fa-416b-b7c6-4136f944a934 req-acde99f0-cee9-4f02-8d75-87b0484b4272 service nova] [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] Refreshing network info cache for port e09cd9e1-44fe-4858-bf9a-0afd7b681ca4 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 618.355640] env[62066]: DEBUG nova.network.neutron [None req-e7221104-9243-4d76-b03e-5782ae8f55d4 tempest-FloatingIPsAssociationNegativeTestJSON-961065153 tempest-FloatingIPsAssociationNegativeTestJSON-961065153-project-member] [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 618.504619] env[62066]: ERROR nova.compute.manager [None req-a9ebb4d2-ff8d-4333-a9df-86618ea2d54c tempest-ImagesOneServerTestJSON-1591819949 tempest-ImagesOneServerTestJSON-1591819949-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port e09cd9e1-44fe-4858-bf9a-0afd7b681ca4, please check neutron logs for more information. [ 618.504619] env[62066]: ERROR nova.compute.manager Traceback (most recent call last): [ 618.504619] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 618.504619] env[62066]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 618.504619] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 618.504619] env[62066]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 618.504619] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 618.504619] env[62066]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 618.504619] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 618.504619] env[62066]: ERROR nova.compute.manager self.force_reraise() [ 618.504619] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 618.504619] env[62066]: ERROR nova.compute.manager raise self.value [ 618.504619] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 618.504619] env[62066]: ERROR nova.compute.manager updated_port = self._update_port( [ 618.504619] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 618.504619] env[62066]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 618.505202] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 618.505202] env[62066]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 618.505202] env[62066]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port e09cd9e1-44fe-4858-bf9a-0afd7b681ca4, please check neutron logs for more information. [ 618.505202] env[62066]: ERROR nova.compute.manager [ 618.505202] env[62066]: Traceback (most recent call last): [ 618.505202] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 618.505202] env[62066]: listener.cb(fileno) [ 618.505202] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 618.505202] env[62066]: result = function(*args, **kwargs) [ 618.505202] env[62066]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 618.505202] env[62066]: return func(*args, **kwargs) [ 618.505202] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 618.505202] env[62066]: raise e [ 618.505202] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 618.505202] env[62066]: nwinfo = self.network_api.allocate_for_instance( [ 618.505202] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 618.505202] env[62066]: created_port_ids = self._update_ports_for_instance( [ 618.505202] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 618.505202] env[62066]: with excutils.save_and_reraise_exception(): [ 618.505202] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 618.505202] env[62066]: self.force_reraise() [ 618.505202] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 618.505202] env[62066]: raise self.value [ 618.505202] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 618.505202] env[62066]: updated_port = self._update_port( [ 618.505202] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 618.505202] env[62066]: _ensure_no_port_binding_failure(port) [ 618.505202] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 618.505202] env[62066]: raise exception.PortBindingFailed(port_id=port['id']) [ 618.506398] env[62066]: nova.exception.PortBindingFailed: Binding failed for port e09cd9e1-44fe-4858-bf9a-0afd7b681ca4, please check neutron logs for more information. [ 618.506398] env[62066]: Removing descriptor: 15 [ 618.506398] env[62066]: ERROR nova.compute.manager [None req-a9ebb4d2-ff8d-4333-a9df-86618ea2d54c tempest-ImagesOneServerTestJSON-1591819949 tempest-ImagesOneServerTestJSON-1591819949-project-member] [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port e09cd9e1-44fe-4858-bf9a-0afd7b681ca4, please check neutron logs for more information. [ 618.506398] env[62066]: ERROR nova.compute.manager [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] Traceback (most recent call last): [ 618.506398] env[62066]: ERROR nova.compute.manager [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 618.506398] env[62066]: ERROR nova.compute.manager [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] yield resources [ 618.506398] env[62066]: ERROR nova.compute.manager [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 618.506398] env[62066]: ERROR nova.compute.manager [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] self.driver.spawn(context, instance, image_meta, [ 618.506398] env[62066]: ERROR nova.compute.manager [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 618.506398] env[62066]: ERROR nova.compute.manager [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] self._vmops.spawn(context, instance, image_meta, injected_files, [ 618.506398] env[62066]: ERROR nova.compute.manager [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 618.506398] env[62066]: ERROR nova.compute.manager [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] vm_ref = self.build_virtual_machine(instance, [ 618.506742] env[62066]: ERROR nova.compute.manager [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 618.506742] env[62066]: ERROR nova.compute.manager [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] vif_infos = vmwarevif.get_vif_info(self._session, [ 618.506742] env[62066]: ERROR nova.compute.manager [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 618.506742] env[62066]: ERROR nova.compute.manager [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] for vif in network_info: [ 618.506742] env[62066]: ERROR nova.compute.manager [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 618.506742] env[62066]: ERROR nova.compute.manager [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] return self._sync_wrapper(fn, *args, **kwargs) [ 618.506742] env[62066]: ERROR nova.compute.manager [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 618.506742] env[62066]: ERROR nova.compute.manager [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] self.wait() [ 618.506742] env[62066]: ERROR nova.compute.manager [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 618.506742] env[62066]: ERROR nova.compute.manager [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] self[:] = self._gt.wait() [ 618.506742] env[62066]: ERROR nova.compute.manager [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 618.506742] env[62066]: ERROR nova.compute.manager [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] return self._exit_event.wait() [ 618.506742] env[62066]: ERROR nova.compute.manager [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 618.507110] env[62066]: ERROR nova.compute.manager [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] result = hub.switch() [ 618.507110] env[62066]: ERROR nova.compute.manager [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 618.507110] env[62066]: ERROR nova.compute.manager [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] return self.greenlet.switch() [ 618.507110] env[62066]: ERROR nova.compute.manager [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 618.507110] env[62066]: ERROR nova.compute.manager [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] result = function(*args, **kwargs) [ 618.507110] env[62066]: ERROR nova.compute.manager [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 618.507110] env[62066]: ERROR nova.compute.manager [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] return func(*args, **kwargs) [ 618.507110] env[62066]: ERROR nova.compute.manager [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 618.507110] env[62066]: ERROR nova.compute.manager [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] raise e [ 618.507110] env[62066]: ERROR nova.compute.manager [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 618.507110] env[62066]: ERROR nova.compute.manager [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] nwinfo = self.network_api.allocate_for_instance( [ 618.507110] env[62066]: ERROR nova.compute.manager [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 618.507110] env[62066]: ERROR nova.compute.manager [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] created_port_ids = self._update_ports_for_instance( [ 618.507531] env[62066]: ERROR nova.compute.manager [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 618.507531] env[62066]: ERROR nova.compute.manager [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] with excutils.save_and_reraise_exception(): [ 618.507531] env[62066]: ERROR nova.compute.manager [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 618.507531] env[62066]: ERROR nova.compute.manager [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] self.force_reraise() [ 618.507531] env[62066]: ERROR nova.compute.manager [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 618.507531] env[62066]: ERROR nova.compute.manager [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] raise self.value [ 618.507531] env[62066]: ERROR nova.compute.manager [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 618.507531] env[62066]: ERROR nova.compute.manager [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] updated_port = self._update_port( [ 618.507531] env[62066]: ERROR nova.compute.manager [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 618.507531] env[62066]: ERROR nova.compute.manager [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] _ensure_no_port_binding_failure(port) [ 618.507531] env[62066]: ERROR nova.compute.manager [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 618.507531] env[62066]: ERROR nova.compute.manager [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] raise exception.PortBindingFailed(port_id=port['id']) [ 618.507864] env[62066]: ERROR nova.compute.manager [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] nova.exception.PortBindingFailed: Binding failed for port e09cd9e1-44fe-4858-bf9a-0afd7b681ca4, please check neutron logs for more information. [ 618.507864] env[62066]: ERROR nova.compute.manager [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] [ 618.507864] env[62066]: INFO nova.compute.manager [None req-a9ebb4d2-ff8d-4333-a9df-86618ea2d54c tempest-ImagesOneServerTestJSON-1591819949 tempest-ImagesOneServerTestJSON-1591819949-project-member] [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] Terminating instance [ 618.508208] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a9ebb4d2-ff8d-4333-a9df-86618ea2d54c tempest-ImagesOneServerTestJSON-1591819949 tempest-ImagesOneServerTestJSON-1591819949-project-member] Acquiring lock "refresh_cache-623300f7-54d2-4b0a-b356-a1ae585682cb" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 618.816294] env[62066]: DEBUG nova.network.neutron [req-f9d54c92-c5fa-416b-b7c6-4136f944a934 req-acde99f0-cee9-4f02-8d75-87b0484b4272 service nova] [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 618.863610] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e7221104-9243-4d76-b03e-5782ae8f55d4 tempest-FloatingIPsAssociationNegativeTestJSON-961065153 tempest-FloatingIPsAssociationNegativeTestJSON-961065153-project-member] Releasing lock "refresh_cache-eee6b9e6-2707-4a3b-925b-a7f50cdc32cb" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 618.863846] env[62066]: DEBUG nova.compute.manager [None req-e7221104-9243-4d76-b03e-5782ae8f55d4 tempest-FloatingIPsAssociationNegativeTestJSON-961065153 tempest-FloatingIPsAssociationNegativeTestJSON-961065153-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 618.864031] env[62066]: DEBUG nova.compute.manager [None req-e7221104-9243-4d76-b03e-5782ae8f55d4 tempest-FloatingIPsAssociationNegativeTestJSON-961065153 tempest-FloatingIPsAssociationNegativeTestJSON-961065153-project-member] [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 618.864197] env[62066]: DEBUG nova.network.neutron [None req-e7221104-9243-4d76-b03e-5782ae8f55d4 tempest-FloatingIPsAssociationNegativeTestJSON-961065153 tempest-FloatingIPsAssociationNegativeTestJSON-961065153-project-member] [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 618.893113] env[62066]: DEBUG nova.network.neutron [None req-e7221104-9243-4d76-b03e-5782ae8f55d4 tempest-FloatingIPsAssociationNegativeTestJSON-961065153 tempest-FloatingIPsAssociationNegativeTestJSON-961065153-project-member] [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 618.917926] env[62066]: DEBUG nova.network.neutron [req-f9d54c92-c5fa-416b-b7c6-4136f944a934 req-acde99f0-cee9-4f02-8d75-87b0484b4272 service nova] [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 619.205392] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a611a48f-af74-428f-92cf-9ad522569336 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.213765] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea74013f-5852-47de-b347-57a2b4db67f6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.256977] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d93b9109-ca38-4c2c-a63b-eb83552523e5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.270160] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6c83bb5-c9b5-4693-b81b-6fba5075f6d4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.285046] env[62066]: DEBUG nova.compute.provider_tree [None req-9371bde4-46e3-4272-a1fe-5125bccbd2c7 tempest-ServersTestJSON-680983284 tempest-ServersTestJSON-680983284-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 619.396722] env[62066]: DEBUG nova.network.neutron [None req-e7221104-9243-4d76-b03e-5782ae8f55d4 tempest-FloatingIPsAssociationNegativeTestJSON-961065153 tempest-FloatingIPsAssociationNegativeTestJSON-961065153-project-member] [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 619.420044] env[62066]: DEBUG oslo_concurrency.lockutils [req-f9d54c92-c5fa-416b-b7c6-4136f944a934 req-acde99f0-cee9-4f02-8d75-87b0484b4272 service nova] Releasing lock "refresh_cache-623300f7-54d2-4b0a-b356-a1ae585682cb" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 619.420506] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a9ebb4d2-ff8d-4333-a9df-86618ea2d54c tempest-ImagesOneServerTestJSON-1591819949 tempest-ImagesOneServerTestJSON-1591819949-project-member] Acquired lock "refresh_cache-623300f7-54d2-4b0a-b356-a1ae585682cb" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 619.420693] env[62066]: DEBUG nova.network.neutron [None req-a9ebb4d2-ff8d-4333-a9df-86618ea2d54c tempest-ImagesOneServerTestJSON-1591819949 tempest-ImagesOneServerTestJSON-1591819949-project-member] [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 619.754870] env[62066]: DEBUG oslo_concurrency.lockutils [None req-478d4566-e74c-4f71-b94d-419dc0a31446 tempest-ServerPasswordTestJSON-849130511 tempest-ServerPasswordTestJSON-849130511-project-member] Acquiring lock "04d25af4-0e73-4650-9c3d-85817754bac9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 619.754870] env[62066]: DEBUG oslo_concurrency.lockutils [None req-478d4566-e74c-4f71-b94d-419dc0a31446 tempest-ServerPasswordTestJSON-849130511 tempest-ServerPasswordTestJSON-849130511-project-member] Lock "04d25af4-0e73-4650-9c3d-85817754bac9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 619.793022] env[62066]: DEBUG nova.scheduler.client.report [None req-9371bde4-46e3-4272-a1fe-5125bccbd2c7 tempest-ServersTestJSON-680983284 tempest-ServersTestJSON-680983284-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 619.901511] env[62066]: INFO nova.compute.manager [None req-e7221104-9243-4d76-b03e-5782ae8f55d4 tempest-FloatingIPsAssociationNegativeTestJSON-961065153 tempest-FloatingIPsAssociationNegativeTestJSON-961065153-project-member] [instance: eee6b9e6-2707-4a3b-925b-a7f50cdc32cb] Took 1.03 seconds to deallocate network for instance. [ 619.951827] env[62066]: DEBUG nova.network.neutron [None req-a9ebb4d2-ff8d-4333-a9df-86618ea2d54c tempest-ImagesOneServerTestJSON-1591819949 tempest-ImagesOneServerTestJSON-1591819949-project-member] [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 620.101188] env[62066]: DEBUG nova.network.neutron [None req-a9ebb4d2-ff8d-4333-a9df-86618ea2d54c tempest-ImagesOneServerTestJSON-1591819949 tempest-ImagesOneServerTestJSON-1591819949-project-member] [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 620.300063] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9371bde4-46e3-4272-a1fe-5125bccbd2c7 tempest-ServersTestJSON-680983284 tempest-ServersTestJSON-680983284-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.618s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 620.300063] env[62066]: DEBUG nova.compute.manager [None req-9371bde4-46e3-4272-a1fe-5125bccbd2c7 tempest-ServersTestJSON-680983284 tempest-ServersTestJSON-680983284-project-member] [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 620.305050] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 17.549s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 620.305050] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 620.305050] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62066) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 620.305050] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7abf52e5-ee7c-4f92-a757-ffa8ca735a57 tempest-ServerActionsTestJSON-164481335 tempest-ServerActionsTestJSON-164481335-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.168s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 620.310150] env[62066]: INFO nova.compute.claims [None req-7abf52e5-ee7c-4f92-a757-ffa8ca735a57 tempest-ServerActionsTestJSON-164481335 tempest-ServerActionsTestJSON-164481335-project-member] [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 620.316929] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-254c9137-c6d3-4041-9415-5476ddf63ca8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.327758] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a79f0a24-c761-48fc-8883-0acd8f27dc01 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.344181] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e9e4b87-7029-4016-bfb9-2ce26b3bed37 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.349789] env[62066]: DEBUG nova.compute.manager [req-18291df0-13d9-48fe-8b63-073f02e92ab9 req-fd865425-0ea2-4886-9578-0998868dc1ec service nova] [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] Received event network-vif-deleted-e09cd9e1-44fe-4858-bf9a-0afd7b681ca4 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 620.356185] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0340bbc6-55a8-44a9-9b70-cb6de8a8e0a8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.389154] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181463MB free_disk=154GB free_vcpus=48 pci_devices=None {{(pid=62066) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 620.389357] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 620.610286] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a9ebb4d2-ff8d-4333-a9df-86618ea2d54c tempest-ImagesOneServerTestJSON-1591819949 tempest-ImagesOneServerTestJSON-1591819949-project-member] Releasing lock "refresh_cache-623300f7-54d2-4b0a-b356-a1ae585682cb" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 620.610286] env[62066]: DEBUG nova.compute.manager [None req-a9ebb4d2-ff8d-4333-a9df-86618ea2d54c tempest-ImagesOneServerTestJSON-1591819949 tempest-ImagesOneServerTestJSON-1591819949-project-member] [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 620.610286] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-a9ebb4d2-ff8d-4333-a9df-86618ea2d54c tempest-ImagesOneServerTestJSON-1591819949 tempest-ImagesOneServerTestJSON-1591819949-project-member] [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 620.610286] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-faf3d3e3-e723-49eb-9481-ce5351d8375e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.620437] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa7ed52a-bc0d-4acf-aad3-37fc368181b5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.648199] env[62066]: WARNING nova.virt.vmwareapi.vmops [None req-a9ebb4d2-ff8d-4333-a9df-86618ea2d54c tempest-ImagesOneServerTestJSON-1591819949 tempest-ImagesOneServerTestJSON-1591819949-project-member] [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 623300f7-54d2-4b0a-b356-a1ae585682cb could not be found. [ 620.649629] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-a9ebb4d2-ff8d-4333-a9df-86618ea2d54c tempest-ImagesOneServerTestJSON-1591819949 tempest-ImagesOneServerTestJSON-1591819949-project-member] [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 620.649629] env[62066]: INFO nova.compute.manager [None req-a9ebb4d2-ff8d-4333-a9df-86618ea2d54c tempest-ImagesOneServerTestJSON-1591819949 tempest-ImagesOneServerTestJSON-1591819949-project-member] [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] Took 0.04 seconds to destroy the instance on the hypervisor. [ 620.649629] env[62066]: DEBUG oslo.service.loopingcall [None req-a9ebb4d2-ff8d-4333-a9df-86618ea2d54c tempest-ImagesOneServerTestJSON-1591819949 tempest-ImagesOneServerTestJSON-1591819949-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 620.650302] env[62066]: DEBUG nova.compute.manager [-] [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 620.651106] env[62066]: DEBUG nova.network.neutron [-] [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 620.675204] env[62066]: DEBUG nova.network.neutron [-] [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 620.818146] env[62066]: DEBUG nova.compute.utils [None req-9371bde4-46e3-4272-a1fe-5125bccbd2c7 tempest-ServersTestJSON-680983284 tempest-ServersTestJSON-680983284-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 620.823021] env[62066]: DEBUG nova.compute.manager [None req-9371bde4-46e3-4272-a1fe-5125bccbd2c7 tempest-ServersTestJSON-680983284 tempest-ServersTestJSON-680983284-project-member] [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 620.823021] env[62066]: DEBUG nova.network.neutron [None req-9371bde4-46e3-4272-a1fe-5125bccbd2c7 tempest-ServersTestJSON-680983284 tempest-ServersTestJSON-680983284-project-member] [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 620.879545] env[62066]: DEBUG nova.policy [None req-9371bde4-46e3-4272-a1fe-5125bccbd2c7 tempest-ServersTestJSON-680983284 tempest-ServersTestJSON-680983284-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e90ac8b08c244a66b1e12ba9194ecb37', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3d5da9724a5148218b02abc160faee6d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 620.934865] env[62066]: INFO nova.scheduler.client.report [None req-e7221104-9243-4d76-b03e-5782ae8f55d4 tempest-FloatingIPsAssociationNegativeTestJSON-961065153 tempest-FloatingIPsAssociationNegativeTestJSON-961065153-project-member] Deleted allocations for instance eee6b9e6-2707-4a3b-925b-a7f50cdc32cb [ 621.181862] env[62066]: DEBUG nova.network.neutron [-] [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 621.326230] env[62066]: DEBUG nova.compute.manager [None req-9371bde4-46e3-4272-a1fe-5125bccbd2c7 tempest-ServersTestJSON-680983284 tempest-ServersTestJSON-680983284-project-member] [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 621.447318] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e7221104-9243-4d76-b03e-5782ae8f55d4 tempest-FloatingIPsAssociationNegativeTestJSON-961065153 tempest-FloatingIPsAssociationNegativeTestJSON-961065153-project-member] Lock "eee6b9e6-2707-4a3b-925b-a7f50cdc32cb" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 97.812s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 621.536123] env[62066]: DEBUG nova.network.neutron [None req-9371bde4-46e3-4272-a1fe-5125bccbd2c7 tempest-ServersTestJSON-680983284 tempest-ServersTestJSON-680983284-project-member] [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] Successfully created port: 3e3a7d95-33be-49e4-9f86-f4315cbe8f18 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 621.690124] env[62066]: INFO nova.compute.manager [-] [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] Took 1.04 seconds to deallocate network for instance. [ 621.692693] env[62066]: DEBUG nova.compute.claims [None req-a9ebb4d2-ff8d-4333-a9df-86618ea2d54c tempest-ImagesOneServerTestJSON-1591819949 tempest-ImagesOneServerTestJSON-1591819949-project-member] [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] Aborting claim: {{(pid=62066) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 621.693023] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a9ebb4d2-ff8d-4333-a9df-86618ea2d54c tempest-ImagesOneServerTestJSON-1591819949 tempest-ImagesOneServerTestJSON-1591819949-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 621.811999] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f53d400c-0bd6-4ce3-baf2-feb2ff287d56 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.820833] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9622cfe5-a5ab-4be0-8063-fbec30668a9f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.859044] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-427fbc32-eb12-41ef-92bc-bdb8024877be {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.868053] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86e65e78-7827-4a2d-9c50-12af9d7076e9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.882099] env[62066]: DEBUG nova.compute.provider_tree [None req-7abf52e5-ee7c-4f92-a757-ffa8ca735a57 tempest-ServerActionsTestJSON-164481335 tempest-ServerActionsTestJSON-164481335-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 621.949053] env[62066]: DEBUG nova.compute.manager [None req-ef1737fc-6565-4a6f-bd10-02fc3b07f848 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 622.367407] env[62066]: DEBUG nova.compute.manager [None req-9371bde4-46e3-4272-a1fe-5125bccbd2c7 tempest-ServersTestJSON-680983284 tempest-ServersTestJSON-680983284-project-member] [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 622.385659] env[62066]: DEBUG nova.scheduler.client.report [None req-7abf52e5-ee7c-4f92-a757-ffa8ca735a57 tempest-ServerActionsTestJSON-164481335 tempest-ServerActionsTestJSON-164481335-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 622.397682] env[62066]: DEBUG nova.virt.hardware [None req-9371bde4-46e3-4272-a1fe-5125bccbd2c7 tempest-ServersTestJSON-680983284 tempest-ServersTestJSON-680983284-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-23T13:40:41Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-23T13:40:23Z,direct_url=,disk_format='vmdk',id=50ff584c-3b50-4395-af07-3e66769bc9f7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='eb52f7069a374c61ae946f052007c6d9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-23T13:40:24Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 622.397814] env[62066]: DEBUG nova.virt.hardware [None req-9371bde4-46e3-4272-a1fe-5125bccbd2c7 tempest-ServersTestJSON-680983284 tempest-ServersTestJSON-680983284-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 622.397868] env[62066]: DEBUG nova.virt.hardware [None req-9371bde4-46e3-4272-a1fe-5125bccbd2c7 tempest-ServersTestJSON-680983284 tempest-ServersTestJSON-680983284-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 622.398257] env[62066]: DEBUG nova.virt.hardware [None req-9371bde4-46e3-4272-a1fe-5125bccbd2c7 tempest-ServersTestJSON-680983284 tempest-ServersTestJSON-680983284-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 622.398257] env[62066]: DEBUG nova.virt.hardware [None req-9371bde4-46e3-4272-a1fe-5125bccbd2c7 tempest-ServersTestJSON-680983284 tempest-ServersTestJSON-680983284-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 622.398352] env[62066]: DEBUG nova.virt.hardware [None req-9371bde4-46e3-4272-a1fe-5125bccbd2c7 tempest-ServersTestJSON-680983284 tempest-ServersTestJSON-680983284-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 622.398592] env[62066]: DEBUG nova.virt.hardware [None req-9371bde4-46e3-4272-a1fe-5125bccbd2c7 tempest-ServersTestJSON-680983284 tempest-ServersTestJSON-680983284-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 622.400959] env[62066]: DEBUG nova.virt.hardware [None req-9371bde4-46e3-4272-a1fe-5125bccbd2c7 tempest-ServersTestJSON-680983284 tempest-ServersTestJSON-680983284-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 622.401692] env[62066]: DEBUG nova.virt.hardware [None req-9371bde4-46e3-4272-a1fe-5125bccbd2c7 tempest-ServersTestJSON-680983284 tempest-ServersTestJSON-680983284-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 622.401931] env[62066]: DEBUG nova.virt.hardware [None req-9371bde4-46e3-4272-a1fe-5125bccbd2c7 tempest-ServersTestJSON-680983284 tempest-ServersTestJSON-680983284-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 622.402104] env[62066]: DEBUG nova.virt.hardware [None req-9371bde4-46e3-4272-a1fe-5125bccbd2c7 tempest-ServersTestJSON-680983284 tempest-ServersTestJSON-680983284-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 622.403103] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6d4d381-abb8-4523-a395-c29025e02e06 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.416432] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67d49b80-4509-408c-aa73-87fe9ce6ad6d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.478879] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ef1737fc-6565-4a6f-bd10-02fc3b07f848 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 622.891855] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7abf52e5-ee7c-4f92-a757-ffa8ca735a57 tempest-ServerActionsTestJSON-164481335 tempest-ServerActionsTestJSON-164481335-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.588s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 622.892241] env[62066]: DEBUG nova.compute.manager [None req-7abf52e5-ee7c-4f92-a757-ffa8ca735a57 tempest-ServerActionsTestJSON-164481335 tempest-ServerActionsTestJSON-164481335-project-member] [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 622.896789] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.505s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 622.981842] env[62066]: DEBUG nova.compute.manager [req-21434b54-1a1d-461f-80b6-11a4c9746edb req-52ef891f-7df2-4c4a-bf82-6f4e1d82eab6 service nova] [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] Received event network-changed-3e3a7d95-33be-49e4-9f86-f4315cbe8f18 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 622.982406] env[62066]: DEBUG nova.compute.manager [req-21434b54-1a1d-461f-80b6-11a4c9746edb req-52ef891f-7df2-4c4a-bf82-6f4e1d82eab6 service nova] [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] Refreshing instance network info cache due to event network-changed-3e3a7d95-33be-49e4-9f86-f4315cbe8f18. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 622.982625] env[62066]: DEBUG oslo_concurrency.lockutils [req-21434b54-1a1d-461f-80b6-11a4c9746edb req-52ef891f-7df2-4c4a-bf82-6f4e1d82eab6 service nova] Acquiring lock "refresh_cache-0040f46d-9a60-4fa1-8a00-c08022e17df0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 622.982765] env[62066]: DEBUG oslo_concurrency.lockutils [req-21434b54-1a1d-461f-80b6-11a4c9746edb req-52ef891f-7df2-4c4a-bf82-6f4e1d82eab6 service nova] Acquired lock "refresh_cache-0040f46d-9a60-4fa1-8a00-c08022e17df0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 622.982994] env[62066]: DEBUG nova.network.neutron [req-21434b54-1a1d-461f-80b6-11a4c9746edb req-52ef891f-7df2-4c4a-bf82-6f4e1d82eab6 service nova] [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] Refreshing network info cache for port 3e3a7d95-33be-49e4-9f86-f4315cbe8f18 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 623.053281] env[62066]: ERROR nova.compute.manager [None req-9371bde4-46e3-4272-a1fe-5125bccbd2c7 tempest-ServersTestJSON-680983284 tempest-ServersTestJSON-680983284-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 3e3a7d95-33be-49e4-9f86-f4315cbe8f18, please check neutron logs for more information. [ 623.053281] env[62066]: ERROR nova.compute.manager Traceback (most recent call last): [ 623.053281] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 623.053281] env[62066]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 623.053281] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 623.053281] env[62066]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 623.053281] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 623.053281] env[62066]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 623.053281] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 623.053281] env[62066]: ERROR nova.compute.manager self.force_reraise() [ 623.053281] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 623.053281] env[62066]: ERROR nova.compute.manager raise self.value [ 623.053281] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 623.053281] env[62066]: ERROR nova.compute.manager updated_port = self._update_port( [ 623.053281] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 623.053281] env[62066]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 623.053758] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 623.053758] env[62066]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 623.053758] env[62066]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 3e3a7d95-33be-49e4-9f86-f4315cbe8f18, please check neutron logs for more information. [ 623.053758] env[62066]: ERROR nova.compute.manager [ 623.053758] env[62066]: Traceback (most recent call last): [ 623.053758] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 623.053758] env[62066]: listener.cb(fileno) [ 623.053758] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 623.053758] env[62066]: result = function(*args, **kwargs) [ 623.053758] env[62066]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 623.053758] env[62066]: return func(*args, **kwargs) [ 623.053758] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 623.053758] env[62066]: raise e [ 623.053758] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 623.053758] env[62066]: nwinfo = self.network_api.allocate_for_instance( [ 623.053758] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 623.053758] env[62066]: created_port_ids = self._update_ports_for_instance( [ 623.053758] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 623.053758] env[62066]: with excutils.save_and_reraise_exception(): [ 623.053758] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 623.053758] env[62066]: self.force_reraise() [ 623.053758] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 623.053758] env[62066]: raise self.value [ 623.053758] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 623.053758] env[62066]: updated_port = self._update_port( [ 623.053758] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 623.053758] env[62066]: _ensure_no_port_binding_failure(port) [ 623.053758] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 623.053758] env[62066]: raise exception.PortBindingFailed(port_id=port['id']) [ 623.054509] env[62066]: nova.exception.PortBindingFailed: Binding failed for port 3e3a7d95-33be-49e4-9f86-f4315cbe8f18, please check neutron logs for more information. [ 623.054509] env[62066]: Removing descriptor: 15 [ 623.054509] env[62066]: ERROR nova.compute.manager [None req-9371bde4-46e3-4272-a1fe-5125bccbd2c7 tempest-ServersTestJSON-680983284 tempest-ServersTestJSON-680983284-project-member] [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 3e3a7d95-33be-49e4-9f86-f4315cbe8f18, please check neutron logs for more information. [ 623.054509] env[62066]: ERROR nova.compute.manager [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] Traceback (most recent call last): [ 623.054509] env[62066]: ERROR nova.compute.manager [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 623.054509] env[62066]: ERROR nova.compute.manager [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] yield resources [ 623.054509] env[62066]: ERROR nova.compute.manager [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 623.054509] env[62066]: ERROR nova.compute.manager [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] self.driver.spawn(context, instance, image_meta, [ 623.054509] env[62066]: ERROR nova.compute.manager [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 623.054509] env[62066]: ERROR nova.compute.manager [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] self._vmops.spawn(context, instance, image_meta, injected_files, [ 623.054509] env[62066]: ERROR nova.compute.manager [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 623.054509] env[62066]: ERROR nova.compute.manager [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] vm_ref = self.build_virtual_machine(instance, [ 623.054822] env[62066]: ERROR nova.compute.manager [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 623.054822] env[62066]: ERROR nova.compute.manager [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] vif_infos = vmwarevif.get_vif_info(self._session, [ 623.054822] env[62066]: ERROR nova.compute.manager [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 623.054822] env[62066]: ERROR nova.compute.manager [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] for vif in network_info: [ 623.054822] env[62066]: ERROR nova.compute.manager [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 623.054822] env[62066]: ERROR nova.compute.manager [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] return self._sync_wrapper(fn, *args, **kwargs) [ 623.054822] env[62066]: ERROR nova.compute.manager [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 623.054822] env[62066]: ERROR nova.compute.manager [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] self.wait() [ 623.054822] env[62066]: ERROR nova.compute.manager [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 623.054822] env[62066]: ERROR nova.compute.manager [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] self[:] = self._gt.wait() [ 623.054822] env[62066]: ERROR nova.compute.manager [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 623.054822] env[62066]: ERROR nova.compute.manager [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] return self._exit_event.wait() [ 623.054822] env[62066]: ERROR nova.compute.manager [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 623.055217] env[62066]: ERROR nova.compute.manager [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] result = hub.switch() [ 623.055217] env[62066]: ERROR nova.compute.manager [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 623.055217] env[62066]: ERROR nova.compute.manager [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] return self.greenlet.switch() [ 623.055217] env[62066]: ERROR nova.compute.manager [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 623.055217] env[62066]: ERROR nova.compute.manager [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] result = function(*args, **kwargs) [ 623.055217] env[62066]: ERROR nova.compute.manager [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 623.055217] env[62066]: ERROR nova.compute.manager [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] return func(*args, **kwargs) [ 623.055217] env[62066]: ERROR nova.compute.manager [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 623.055217] env[62066]: ERROR nova.compute.manager [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] raise e [ 623.055217] env[62066]: ERROR nova.compute.manager [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 623.055217] env[62066]: ERROR nova.compute.manager [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] nwinfo = self.network_api.allocate_for_instance( [ 623.055217] env[62066]: ERROR nova.compute.manager [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 623.055217] env[62066]: ERROR nova.compute.manager [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] created_port_ids = self._update_ports_for_instance( [ 623.055556] env[62066]: ERROR nova.compute.manager [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 623.055556] env[62066]: ERROR nova.compute.manager [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] with excutils.save_and_reraise_exception(): [ 623.055556] env[62066]: ERROR nova.compute.manager [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 623.055556] env[62066]: ERROR nova.compute.manager [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] self.force_reraise() [ 623.055556] env[62066]: ERROR nova.compute.manager [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 623.055556] env[62066]: ERROR nova.compute.manager [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] raise self.value [ 623.055556] env[62066]: ERROR nova.compute.manager [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 623.055556] env[62066]: ERROR nova.compute.manager [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] updated_port = self._update_port( [ 623.055556] env[62066]: ERROR nova.compute.manager [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 623.055556] env[62066]: ERROR nova.compute.manager [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] _ensure_no_port_binding_failure(port) [ 623.055556] env[62066]: ERROR nova.compute.manager [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 623.055556] env[62066]: ERROR nova.compute.manager [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] raise exception.PortBindingFailed(port_id=port['id']) [ 623.055902] env[62066]: ERROR nova.compute.manager [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] nova.exception.PortBindingFailed: Binding failed for port 3e3a7d95-33be-49e4-9f86-f4315cbe8f18, please check neutron logs for more information. [ 623.055902] env[62066]: ERROR nova.compute.manager [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] [ 623.055902] env[62066]: INFO nova.compute.manager [None req-9371bde4-46e3-4272-a1fe-5125bccbd2c7 tempest-ServersTestJSON-680983284 tempest-ServersTestJSON-680983284-project-member] [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] Terminating instance [ 623.055902] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9371bde4-46e3-4272-a1fe-5125bccbd2c7 tempest-ServersTestJSON-680983284 tempest-ServersTestJSON-680983284-project-member] Acquiring lock "refresh_cache-0040f46d-9a60-4fa1-8a00-c08022e17df0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 623.297636] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] Acquiring lock "d973b09f-7727-4055-b05e-123079df9ce4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 623.297894] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] Lock "d973b09f-7727-4055-b05e-123079df9ce4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 623.334084] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] Acquiring lock "71c44a07-fdf0-4847-868a-47f16c033c3d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 623.335120] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] Lock "71c44a07-fdf0-4847-868a-47f16c033c3d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 623.368443] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] Acquiring lock "f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 623.369439] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] Lock "f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 623.412356] env[62066]: DEBUG nova.compute.utils [None req-7abf52e5-ee7c-4f92-a757-ffa8ca735a57 tempest-ServerActionsTestJSON-164481335 tempest-ServerActionsTestJSON-164481335-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 623.413262] env[62066]: DEBUG nova.compute.manager [None req-7abf52e5-ee7c-4f92-a757-ffa8ca735a57 tempest-ServerActionsTestJSON-164481335 tempest-ServerActionsTestJSON-164481335-project-member] [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 623.413915] env[62066]: DEBUG nova.network.neutron [None req-7abf52e5-ee7c-4f92-a757-ffa8ca735a57 tempest-ServerActionsTestJSON-164481335 tempest-ServerActionsTestJSON-164481335-project-member] [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 623.494650] env[62066]: DEBUG nova.policy [None req-7abf52e5-ee7c-4f92-a757-ffa8ca735a57 tempest-ServerActionsTestJSON-164481335 tempest-ServerActionsTestJSON-164481335-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5f5c4bc4b1254f2ca1b0ded2e19c3ee4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4cf38a47ab264ee9aeee3b3375999ef1', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 623.518876] env[62066]: DEBUG nova.network.neutron [req-21434b54-1a1d-461f-80b6-11a4c9746edb req-52ef891f-7df2-4c4a-bf82-6f4e1d82eab6 service nova] [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 623.640266] env[62066]: DEBUG nova.network.neutron [req-21434b54-1a1d-461f-80b6-11a4c9746edb req-52ef891f-7df2-4c4a-bf82-6f4e1d82eab6 service nova] [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 623.868378] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7fc7916e-6e82-4e11-a2d0-03a259fc6439 tempest-ServerAddressesNegativeTestJSON-1804754632 tempest-ServerAddressesNegativeTestJSON-1804754632-project-member] Acquiring lock "239f0b4b-32cf-4f9f-a47e-bac4235fb6ef" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 623.869210] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7fc7916e-6e82-4e11-a2d0-03a259fc6439 tempest-ServerAddressesNegativeTestJSON-1804754632 tempest-ServerAddressesNegativeTestJSON-1804754632-project-member] Lock "239f0b4b-32cf-4f9f-a47e-bac4235fb6ef" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 623.923120] env[62066]: DEBUG nova.compute.manager [None req-7abf52e5-ee7c-4f92-a757-ffa8ca735a57 tempest-ServerActionsTestJSON-164481335 tempest-ServerActionsTestJSON-164481335-project-member] [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 624.017238] env[62066]: DEBUG nova.network.neutron [None req-7abf52e5-ee7c-4f92-a757-ffa8ca735a57 tempest-ServerActionsTestJSON-164481335 tempest-ServerActionsTestJSON-164481335-project-member] [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] Successfully created port: c80661ff-38e1-4e3e-9647-43d6309380db {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 624.040800] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69015c36-1e65-4f31-a3d0-05bfb590eda5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.049789] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5af30727-441c-4fb5-b1f2-4dfbf3b3d6f4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.080323] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d24f8115-4b2c-479b-a164-126f5b145978 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.088256] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d342d5ea-ed6a-4896-b56b-2d6e44796a12 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.102771] env[62066]: DEBUG nova.compute.provider_tree [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 624.146607] env[62066]: DEBUG oslo_concurrency.lockutils [req-21434b54-1a1d-461f-80b6-11a4c9746edb req-52ef891f-7df2-4c4a-bf82-6f4e1d82eab6 service nova] Releasing lock "refresh_cache-0040f46d-9a60-4fa1-8a00-c08022e17df0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 624.147225] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9371bde4-46e3-4272-a1fe-5125bccbd2c7 tempest-ServersTestJSON-680983284 tempest-ServersTestJSON-680983284-project-member] Acquired lock "refresh_cache-0040f46d-9a60-4fa1-8a00-c08022e17df0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 624.147545] env[62066]: DEBUG nova.network.neutron [None req-9371bde4-46e3-4272-a1fe-5125bccbd2c7 tempest-ServersTestJSON-680983284 tempest-ServersTestJSON-680983284-project-member] [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 624.606719] env[62066]: DEBUG nova.scheduler.client.report [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 624.671959] env[62066]: DEBUG nova.network.neutron [None req-9371bde4-46e3-4272-a1fe-5125bccbd2c7 tempest-ServersTestJSON-680983284 tempest-ServersTestJSON-680983284-project-member] [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 624.870994] env[62066]: DEBUG nova.network.neutron [None req-9371bde4-46e3-4272-a1fe-5125bccbd2c7 tempest-ServersTestJSON-680983284 tempest-ServersTestJSON-680983284-project-member] [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 624.939028] env[62066]: DEBUG nova.compute.manager [None req-7abf52e5-ee7c-4f92-a757-ffa8ca735a57 tempest-ServerActionsTestJSON-164481335 tempest-ServerActionsTestJSON-164481335-project-member] [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 624.967170] env[62066]: DEBUG nova.virt.hardware [None req-7abf52e5-ee7c-4f92-a757-ffa8ca735a57 tempest-ServerActionsTestJSON-164481335 tempest-ServerActionsTestJSON-164481335-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-23T13:40:41Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-23T13:40:23Z,direct_url=,disk_format='vmdk',id=50ff584c-3b50-4395-af07-3e66769bc9f7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='eb52f7069a374c61ae946f052007c6d9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-23T13:40:24Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 624.967170] env[62066]: DEBUG nova.virt.hardware [None req-7abf52e5-ee7c-4f92-a757-ffa8ca735a57 tempest-ServerActionsTestJSON-164481335 tempest-ServerActionsTestJSON-164481335-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 624.967368] env[62066]: DEBUG nova.virt.hardware [None req-7abf52e5-ee7c-4f92-a757-ffa8ca735a57 tempest-ServerActionsTestJSON-164481335 tempest-ServerActionsTestJSON-164481335-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 624.967625] env[62066]: DEBUG nova.virt.hardware [None req-7abf52e5-ee7c-4f92-a757-ffa8ca735a57 tempest-ServerActionsTestJSON-164481335 tempest-ServerActionsTestJSON-164481335-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 624.967874] env[62066]: DEBUG nova.virt.hardware [None req-7abf52e5-ee7c-4f92-a757-ffa8ca735a57 tempest-ServerActionsTestJSON-164481335 tempest-ServerActionsTestJSON-164481335-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 624.967940] env[62066]: DEBUG nova.virt.hardware [None req-7abf52e5-ee7c-4f92-a757-ffa8ca735a57 tempest-ServerActionsTestJSON-164481335 tempest-ServerActionsTestJSON-164481335-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 624.968259] env[62066]: DEBUG nova.virt.hardware [None req-7abf52e5-ee7c-4f92-a757-ffa8ca735a57 tempest-ServerActionsTestJSON-164481335 tempest-ServerActionsTestJSON-164481335-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 624.968441] env[62066]: DEBUG nova.virt.hardware [None req-7abf52e5-ee7c-4f92-a757-ffa8ca735a57 tempest-ServerActionsTestJSON-164481335 tempest-ServerActionsTestJSON-164481335-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 624.968618] env[62066]: DEBUG nova.virt.hardware [None req-7abf52e5-ee7c-4f92-a757-ffa8ca735a57 tempest-ServerActionsTestJSON-164481335 tempest-ServerActionsTestJSON-164481335-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 624.968781] env[62066]: DEBUG nova.virt.hardware [None req-7abf52e5-ee7c-4f92-a757-ffa8ca735a57 tempest-ServerActionsTestJSON-164481335 tempest-ServerActionsTestJSON-164481335-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 624.968958] env[62066]: DEBUG nova.virt.hardware [None req-7abf52e5-ee7c-4f92-a757-ffa8ca735a57 tempest-ServerActionsTestJSON-164481335 tempest-ServerActionsTestJSON-164481335-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 624.970654] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66125e83-86e2-41c0-a96f-42426d1e1fe1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.977818] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5648485-00ce-48de-bc96-26b0f956d89d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.018162] env[62066]: ERROR nova.compute.manager [None req-7abf52e5-ee7c-4f92-a757-ffa8ca735a57 tempest-ServerActionsTestJSON-164481335 tempest-ServerActionsTestJSON-164481335-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port c80661ff-38e1-4e3e-9647-43d6309380db, please check neutron logs for more information. [ 625.018162] env[62066]: ERROR nova.compute.manager Traceback (most recent call last): [ 625.018162] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 625.018162] env[62066]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 625.018162] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 625.018162] env[62066]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 625.018162] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 625.018162] env[62066]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 625.018162] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 625.018162] env[62066]: ERROR nova.compute.manager self.force_reraise() [ 625.018162] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 625.018162] env[62066]: ERROR nova.compute.manager raise self.value [ 625.018162] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 625.018162] env[62066]: ERROR nova.compute.manager updated_port = self._update_port( [ 625.018162] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 625.018162] env[62066]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 625.018639] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 625.018639] env[62066]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 625.018639] env[62066]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port c80661ff-38e1-4e3e-9647-43d6309380db, please check neutron logs for more information. [ 625.018639] env[62066]: ERROR nova.compute.manager [ 625.018639] env[62066]: Traceback (most recent call last): [ 625.018639] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 625.018639] env[62066]: listener.cb(fileno) [ 625.018639] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 625.018639] env[62066]: result = function(*args, **kwargs) [ 625.018639] env[62066]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 625.018639] env[62066]: return func(*args, **kwargs) [ 625.018639] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 625.018639] env[62066]: raise e [ 625.018639] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 625.018639] env[62066]: nwinfo = self.network_api.allocate_for_instance( [ 625.018639] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 625.018639] env[62066]: created_port_ids = self._update_ports_for_instance( [ 625.018639] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 625.018639] env[62066]: with excutils.save_and_reraise_exception(): [ 625.018639] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 625.018639] env[62066]: self.force_reraise() [ 625.018639] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 625.018639] env[62066]: raise self.value [ 625.018639] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 625.018639] env[62066]: updated_port = self._update_port( [ 625.018639] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 625.018639] env[62066]: _ensure_no_port_binding_failure(port) [ 625.018639] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 625.018639] env[62066]: raise exception.PortBindingFailed(port_id=port['id']) [ 625.019428] env[62066]: nova.exception.PortBindingFailed: Binding failed for port c80661ff-38e1-4e3e-9647-43d6309380db, please check neutron logs for more information. [ 625.019428] env[62066]: Removing descriptor: 15 [ 625.019428] env[62066]: ERROR nova.compute.manager [None req-7abf52e5-ee7c-4f92-a757-ffa8ca735a57 tempest-ServerActionsTestJSON-164481335 tempest-ServerActionsTestJSON-164481335-project-member] [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port c80661ff-38e1-4e3e-9647-43d6309380db, please check neutron logs for more information. [ 625.019428] env[62066]: ERROR nova.compute.manager [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] Traceback (most recent call last): [ 625.019428] env[62066]: ERROR nova.compute.manager [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 625.019428] env[62066]: ERROR nova.compute.manager [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] yield resources [ 625.019428] env[62066]: ERROR nova.compute.manager [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 625.019428] env[62066]: ERROR nova.compute.manager [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] self.driver.spawn(context, instance, image_meta, [ 625.019428] env[62066]: ERROR nova.compute.manager [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 625.019428] env[62066]: ERROR nova.compute.manager [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] self._vmops.spawn(context, instance, image_meta, injected_files, [ 625.019428] env[62066]: ERROR nova.compute.manager [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 625.019428] env[62066]: ERROR nova.compute.manager [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] vm_ref = self.build_virtual_machine(instance, [ 625.019801] env[62066]: ERROR nova.compute.manager [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 625.019801] env[62066]: ERROR nova.compute.manager [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] vif_infos = vmwarevif.get_vif_info(self._session, [ 625.019801] env[62066]: ERROR nova.compute.manager [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 625.019801] env[62066]: ERROR nova.compute.manager [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] for vif in network_info: [ 625.019801] env[62066]: ERROR nova.compute.manager [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 625.019801] env[62066]: ERROR nova.compute.manager [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] return self._sync_wrapper(fn, *args, **kwargs) [ 625.019801] env[62066]: ERROR nova.compute.manager [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 625.019801] env[62066]: ERROR nova.compute.manager [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] self.wait() [ 625.019801] env[62066]: ERROR nova.compute.manager [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 625.019801] env[62066]: ERROR nova.compute.manager [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] self[:] = self._gt.wait() [ 625.019801] env[62066]: ERROR nova.compute.manager [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 625.019801] env[62066]: ERROR nova.compute.manager [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] return self._exit_event.wait() [ 625.019801] env[62066]: ERROR nova.compute.manager [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 625.020170] env[62066]: ERROR nova.compute.manager [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] result = hub.switch() [ 625.020170] env[62066]: ERROR nova.compute.manager [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 625.020170] env[62066]: ERROR nova.compute.manager [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] return self.greenlet.switch() [ 625.020170] env[62066]: ERROR nova.compute.manager [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 625.020170] env[62066]: ERROR nova.compute.manager [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] result = function(*args, **kwargs) [ 625.020170] env[62066]: ERROR nova.compute.manager [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 625.020170] env[62066]: ERROR nova.compute.manager [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] return func(*args, **kwargs) [ 625.020170] env[62066]: ERROR nova.compute.manager [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 625.020170] env[62066]: ERROR nova.compute.manager [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] raise e [ 625.020170] env[62066]: ERROR nova.compute.manager [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 625.020170] env[62066]: ERROR nova.compute.manager [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] nwinfo = self.network_api.allocate_for_instance( [ 625.020170] env[62066]: ERROR nova.compute.manager [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 625.020170] env[62066]: ERROR nova.compute.manager [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] created_port_ids = self._update_ports_for_instance( [ 625.020559] env[62066]: ERROR nova.compute.manager [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 625.020559] env[62066]: ERROR nova.compute.manager [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] with excutils.save_and_reraise_exception(): [ 625.020559] env[62066]: ERROR nova.compute.manager [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 625.020559] env[62066]: ERROR nova.compute.manager [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] self.force_reraise() [ 625.020559] env[62066]: ERROR nova.compute.manager [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 625.020559] env[62066]: ERROR nova.compute.manager [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] raise self.value [ 625.020559] env[62066]: ERROR nova.compute.manager [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 625.020559] env[62066]: ERROR nova.compute.manager [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] updated_port = self._update_port( [ 625.020559] env[62066]: ERROR nova.compute.manager [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 625.020559] env[62066]: ERROR nova.compute.manager [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] _ensure_no_port_binding_failure(port) [ 625.020559] env[62066]: ERROR nova.compute.manager [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 625.020559] env[62066]: ERROR nova.compute.manager [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] raise exception.PortBindingFailed(port_id=port['id']) [ 625.020867] env[62066]: ERROR nova.compute.manager [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] nova.exception.PortBindingFailed: Binding failed for port c80661ff-38e1-4e3e-9647-43d6309380db, please check neutron logs for more information. [ 625.020867] env[62066]: ERROR nova.compute.manager [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] [ 625.020867] env[62066]: INFO nova.compute.manager [None req-7abf52e5-ee7c-4f92-a757-ffa8ca735a57 tempest-ServerActionsTestJSON-164481335 tempest-ServerActionsTestJSON-164481335-project-member] [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] Terminating instance [ 625.021290] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7abf52e5-ee7c-4f92-a757-ffa8ca735a57 tempest-ServerActionsTestJSON-164481335 tempest-ServerActionsTestJSON-164481335-project-member] Acquiring lock "refresh_cache-bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 625.021449] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7abf52e5-ee7c-4f92-a757-ffa8ca735a57 tempest-ServerActionsTestJSON-164481335 tempest-ServerActionsTestJSON-164481335-project-member] Acquired lock "refresh_cache-bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 625.021614] env[62066]: DEBUG nova.network.neutron [None req-7abf52e5-ee7c-4f92-a757-ffa8ca735a57 tempest-ServerActionsTestJSON-164481335 tempest-ServerActionsTestJSON-164481335-project-member] [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 625.112740] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.217s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 625.113376] env[62066]: ERROR nova.compute.manager [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port afd58cf2-370e-45b9-bc5a-7bea27a39c71, please check neutron logs for more information. [ 625.113376] env[62066]: ERROR nova.compute.manager [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] Traceback (most recent call last): [ 625.113376] env[62066]: ERROR nova.compute.manager [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 625.113376] env[62066]: ERROR nova.compute.manager [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] self.driver.spawn(context, instance, image_meta, [ 625.113376] env[62066]: ERROR nova.compute.manager [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 625.113376] env[62066]: ERROR nova.compute.manager [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] self._vmops.spawn(context, instance, image_meta, injected_files, [ 625.113376] env[62066]: ERROR nova.compute.manager [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 625.113376] env[62066]: ERROR nova.compute.manager [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] vm_ref = self.build_virtual_machine(instance, [ 625.113376] env[62066]: ERROR nova.compute.manager [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 625.113376] env[62066]: ERROR nova.compute.manager [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] vif_infos = vmwarevif.get_vif_info(self._session, [ 625.113376] env[62066]: ERROR nova.compute.manager [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 625.114738] env[62066]: ERROR nova.compute.manager [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] for vif in network_info: [ 625.114738] env[62066]: ERROR nova.compute.manager [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 625.114738] env[62066]: ERROR nova.compute.manager [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] return self._sync_wrapper(fn, *args, **kwargs) [ 625.114738] env[62066]: ERROR nova.compute.manager [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 625.114738] env[62066]: ERROR nova.compute.manager [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] self.wait() [ 625.114738] env[62066]: ERROR nova.compute.manager [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 625.114738] env[62066]: ERROR nova.compute.manager [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] self[:] = self._gt.wait() [ 625.114738] env[62066]: ERROR nova.compute.manager [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 625.114738] env[62066]: ERROR nova.compute.manager [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] return self._exit_event.wait() [ 625.114738] env[62066]: ERROR nova.compute.manager [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 625.114738] env[62066]: ERROR nova.compute.manager [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] current.throw(*self._exc) [ 625.114738] env[62066]: ERROR nova.compute.manager [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 625.114738] env[62066]: ERROR nova.compute.manager [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] result = function(*args, **kwargs) [ 625.115068] env[62066]: ERROR nova.compute.manager [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 625.115068] env[62066]: ERROR nova.compute.manager [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] return func(*args, **kwargs) [ 625.115068] env[62066]: ERROR nova.compute.manager [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 625.115068] env[62066]: ERROR nova.compute.manager [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] raise e [ 625.115068] env[62066]: ERROR nova.compute.manager [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 625.115068] env[62066]: ERROR nova.compute.manager [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] nwinfo = self.network_api.allocate_for_instance( [ 625.115068] env[62066]: ERROR nova.compute.manager [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 625.115068] env[62066]: ERROR nova.compute.manager [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] created_port_ids = self._update_ports_for_instance( [ 625.115068] env[62066]: ERROR nova.compute.manager [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 625.115068] env[62066]: ERROR nova.compute.manager [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] with excutils.save_and_reraise_exception(): [ 625.115068] env[62066]: ERROR nova.compute.manager [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 625.115068] env[62066]: ERROR nova.compute.manager [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] self.force_reraise() [ 625.115068] env[62066]: ERROR nova.compute.manager [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 625.115393] env[62066]: ERROR nova.compute.manager [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] raise self.value [ 625.115393] env[62066]: ERROR nova.compute.manager [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 625.115393] env[62066]: ERROR nova.compute.manager [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] updated_port = self._update_port( [ 625.115393] env[62066]: ERROR nova.compute.manager [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 625.115393] env[62066]: ERROR nova.compute.manager [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] _ensure_no_port_binding_failure(port) [ 625.115393] env[62066]: ERROR nova.compute.manager [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 625.115393] env[62066]: ERROR nova.compute.manager [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] raise exception.PortBindingFailed(port_id=port['id']) [ 625.115393] env[62066]: ERROR nova.compute.manager [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] nova.exception.PortBindingFailed: Binding failed for port afd58cf2-370e-45b9-bc5a-7bea27a39c71, please check neutron logs for more information. [ 625.115393] env[62066]: ERROR nova.compute.manager [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] [ 625.115393] env[62066]: DEBUG nova.compute.utils [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] Binding failed for port afd58cf2-370e-45b9-bc5a-7bea27a39c71, please check neutron logs for more information. {{(pid=62066) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 625.115672] env[62066]: DEBUG oslo_concurrency.lockutils [None req-29878964-568e-4c92-80e6-64ee3e8e11bb tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.299s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 625.117204] env[62066]: INFO nova.compute.claims [None req-29878964-568e-4c92-80e6-64ee3e8e11bb tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 625.120954] env[62066]: DEBUG nova.compute.manager [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] Build of instance 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962 was re-scheduled: Binding failed for port afd58cf2-370e-45b9-bc5a-7bea27a39c71, please check neutron logs for more information. {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 625.121101] env[62066]: DEBUG nova.compute.manager [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] Unplugging VIFs for instance {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 625.121327] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Acquiring lock "refresh_cache-5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 625.123208] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Acquired lock "refresh_cache-5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 625.123406] env[62066]: DEBUG nova.network.neutron [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 625.182160] env[62066]: DEBUG nova.compute.manager [req-1c6e0409-08a1-41bf-a04e-cd32fa5c5eff req-750c924c-5689-4bf4-a12c-c7a8dd654589 service nova] [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] Received event network-vif-deleted-3e3a7d95-33be-49e4-9f86-f4315cbe8f18 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 625.182160] env[62066]: DEBUG nova.compute.manager [req-1c6e0409-08a1-41bf-a04e-cd32fa5c5eff req-750c924c-5689-4bf4-a12c-c7a8dd654589 service nova] [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] Received event network-changed-c80661ff-38e1-4e3e-9647-43d6309380db {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 625.182160] env[62066]: DEBUG nova.compute.manager [req-1c6e0409-08a1-41bf-a04e-cd32fa5c5eff req-750c924c-5689-4bf4-a12c-c7a8dd654589 service nova] [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] Refreshing instance network info cache due to event network-changed-c80661ff-38e1-4e3e-9647-43d6309380db. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 625.182160] env[62066]: DEBUG oslo_concurrency.lockutils [req-1c6e0409-08a1-41bf-a04e-cd32fa5c5eff req-750c924c-5689-4bf4-a12c-c7a8dd654589 service nova] Acquiring lock "refresh_cache-bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 625.374031] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9371bde4-46e3-4272-a1fe-5125bccbd2c7 tempest-ServersTestJSON-680983284 tempest-ServersTestJSON-680983284-project-member] Releasing lock "refresh_cache-0040f46d-9a60-4fa1-8a00-c08022e17df0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 625.374167] env[62066]: DEBUG nova.compute.manager [None req-9371bde4-46e3-4272-a1fe-5125bccbd2c7 tempest-ServersTestJSON-680983284 tempest-ServersTestJSON-680983284-project-member] [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 625.374336] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-9371bde4-46e3-4272-a1fe-5125bccbd2c7 tempest-ServersTestJSON-680983284 tempest-ServersTestJSON-680983284-project-member] [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 625.374644] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-983f7090-98f7-4c35-9e0c-e820bdfbaa18 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.383301] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e68036b6-a71d-49bc-b7f4-4b936055acb4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.404206] env[62066]: WARNING nova.virt.vmwareapi.vmops [None req-9371bde4-46e3-4272-a1fe-5125bccbd2c7 tempest-ServersTestJSON-680983284 tempest-ServersTestJSON-680983284-project-member] [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 0040f46d-9a60-4fa1-8a00-c08022e17df0 could not be found. [ 625.405354] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-9371bde4-46e3-4272-a1fe-5125bccbd2c7 tempest-ServersTestJSON-680983284 tempest-ServersTestJSON-680983284-project-member] [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 625.405354] env[62066]: INFO nova.compute.manager [None req-9371bde4-46e3-4272-a1fe-5125bccbd2c7 tempest-ServersTestJSON-680983284 tempest-ServersTestJSON-680983284-project-member] [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] Took 0.03 seconds to destroy the instance on the hypervisor. [ 625.405354] env[62066]: DEBUG oslo.service.loopingcall [None req-9371bde4-46e3-4272-a1fe-5125bccbd2c7 tempest-ServersTestJSON-680983284 tempest-ServersTestJSON-680983284-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 625.405354] env[62066]: DEBUG nova.compute.manager [-] [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 625.405354] env[62066]: DEBUG nova.network.neutron [-] [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 625.546504] env[62066]: DEBUG nova.network.neutron [None req-7abf52e5-ee7c-4f92-a757-ffa8ca735a57 tempest-ServerActionsTestJSON-164481335 tempest-ServerActionsTestJSON-164481335-project-member] [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 625.559624] env[62066]: DEBUG nova.network.neutron [-] [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 625.646049] env[62066]: DEBUG nova.network.neutron [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 625.696262] env[62066]: DEBUG nova.network.neutron [None req-7abf52e5-ee7c-4f92-a757-ffa8ca735a57 tempest-ServerActionsTestJSON-164481335 tempest-ServerActionsTestJSON-164481335-project-member] [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 625.736045] env[62066]: DEBUG nova.network.neutron [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 626.063495] env[62066]: DEBUG nova.network.neutron [-] [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 626.200568] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7abf52e5-ee7c-4f92-a757-ffa8ca735a57 tempest-ServerActionsTestJSON-164481335 tempest-ServerActionsTestJSON-164481335-project-member] Releasing lock "refresh_cache-bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 626.200816] env[62066]: DEBUG nova.compute.manager [None req-7abf52e5-ee7c-4f92-a757-ffa8ca735a57 tempest-ServerActionsTestJSON-164481335 tempest-ServerActionsTestJSON-164481335-project-member] [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 626.200963] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-7abf52e5-ee7c-4f92-a757-ffa8ca735a57 tempest-ServerActionsTestJSON-164481335 tempest-ServerActionsTestJSON-164481335-project-member] [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 626.201281] env[62066]: DEBUG oslo_concurrency.lockutils [req-1c6e0409-08a1-41bf-a04e-cd32fa5c5eff req-750c924c-5689-4bf4-a12c-c7a8dd654589 service nova] Acquired lock "refresh_cache-bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 626.201442] env[62066]: DEBUG nova.network.neutron [req-1c6e0409-08a1-41bf-a04e-cd32fa5c5eff req-750c924c-5689-4bf4-a12c-c7a8dd654589 service nova] [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] Refreshing network info cache for port c80661ff-38e1-4e3e-9647-43d6309380db {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 626.202481] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-04fa70b6-a287-4905-9b87-befe20fb32f4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.213348] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad057589-132a-4c14-8ad9-ad178d3da376 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.236867] env[62066]: WARNING nova.virt.vmwareapi.vmops [None req-7abf52e5-ee7c-4f92-a757-ffa8ca735a57 tempest-ServerActionsTestJSON-164481335 tempest-ServerActionsTestJSON-164481335-project-member] [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94 could not be found. [ 626.237411] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-7abf52e5-ee7c-4f92-a757-ffa8ca735a57 tempest-ServerActionsTestJSON-164481335 tempest-ServerActionsTestJSON-164481335-project-member] [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 626.237411] env[62066]: INFO nova.compute.manager [None req-7abf52e5-ee7c-4f92-a757-ffa8ca735a57 tempest-ServerActionsTestJSON-164481335 tempest-ServerActionsTestJSON-164481335-project-member] [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] Took 0.04 seconds to destroy the instance on the hypervisor. [ 626.237595] env[62066]: DEBUG oslo.service.loopingcall [None req-7abf52e5-ee7c-4f92-a757-ffa8ca735a57 tempest-ServerActionsTestJSON-164481335 tempest-ServerActionsTestJSON-164481335-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 626.240097] env[62066]: DEBUG nova.compute.manager [-] [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 626.240210] env[62066]: DEBUG nova.network.neutron [-] [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 626.241782] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Releasing lock "refresh_cache-5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 626.241982] env[62066]: DEBUG nova.compute.manager [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 626.242165] env[62066]: DEBUG nova.compute.manager [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 626.242321] env[62066]: DEBUG nova.network.neutron [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 626.259356] env[62066]: DEBUG nova.network.neutron [-] [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 626.400949] env[62066]: DEBUG nova.network.neutron [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 626.560828] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8d1e903-660c-446b-b690-d48d4c72c2bb {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.566127] env[62066]: INFO nova.compute.manager [-] [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] Took 1.16 seconds to deallocate network for instance. [ 626.568886] env[62066]: DEBUG nova.compute.claims [None req-9371bde4-46e3-4272-a1fe-5125bccbd2c7 tempest-ServersTestJSON-680983284 tempest-ServersTestJSON-680983284-project-member] [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] Aborting claim: {{(pid=62066) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 626.569079] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9371bde4-46e3-4272-a1fe-5125bccbd2c7 tempest-ServersTestJSON-680983284 tempest-ServersTestJSON-680983284-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 626.570041] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c82f6c4-38a1-40f0-b6ec-c5faa9cdabb4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.602109] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bae2701-d4a5-44c2-afe8-2c85118e8741 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.609417] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3de04c4-93a5-46b3-b365-4d39471c65e9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.622823] env[62066]: DEBUG nova.compute.provider_tree [None req-29878964-568e-4c92-80e6-64ee3e8e11bb tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 626.726369] env[62066]: DEBUG nova.network.neutron [req-1c6e0409-08a1-41bf-a04e-cd32fa5c5eff req-750c924c-5689-4bf4-a12c-c7a8dd654589 service nova] [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 626.761955] env[62066]: DEBUG nova.network.neutron [-] [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 626.842920] env[62066]: DEBUG nova.network.neutron [req-1c6e0409-08a1-41bf-a04e-cd32fa5c5eff req-750c924c-5689-4bf4-a12c-c7a8dd654589 service nova] [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 626.904083] env[62066]: DEBUG nova.network.neutron [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 627.125983] env[62066]: DEBUG nova.scheduler.client.report [None req-29878964-568e-4c92-80e6-64ee3e8e11bb tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 627.204070] env[62066]: DEBUG nova.compute.manager [req-c989b5c9-b463-423c-90de-7c6be578fd3d req-7d1c09d0-7e3e-4eca-b0bc-7d9c04a251c9 service nova] [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] Received event network-vif-deleted-c80661ff-38e1-4e3e-9647-43d6309380db {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 627.266196] env[62066]: INFO nova.compute.manager [-] [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] Took 1.02 seconds to deallocate network for instance. [ 627.267113] env[62066]: DEBUG nova.compute.claims [None req-7abf52e5-ee7c-4f92-a757-ffa8ca735a57 tempest-ServerActionsTestJSON-164481335 tempest-ServerActionsTestJSON-164481335-project-member] [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] Aborting claim: {{(pid=62066) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 627.267113] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7abf52e5-ee7c-4f92-a757-ffa8ca735a57 tempest-ServerActionsTestJSON-164481335 tempest-ServerActionsTestJSON-164481335-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 627.346369] env[62066]: DEBUG oslo_concurrency.lockutils [req-1c6e0409-08a1-41bf-a04e-cd32fa5c5eff req-750c924c-5689-4bf4-a12c-c7a8dd654589 service nova] Releasing lock "refresh_cache-bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 627.410881] env[62066]: INFO nova.compute.manager [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962] Took 1.17 seconds to deallocate network for instance. [ 627.631176] env[62066]: DEBUG oslo_concurrency.lockutils [None req-29878964-568e-4c92-80e6-64ee3e8e11bb tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.516s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 627.631712] env[62066]: DEBUG nova.compute.manager [None req-29878964-568e-4c92-80e6-64ee3e8e11bb tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 627.634294] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.672s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 628.139402] env[62066]: DEBUG nova.compute.utils [None req-29878964-568e-4c92-80e6-64ee3e8e11bb tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 628.145110] env[62066]: DEBUG nova.compute.manager [None req-29878964-568e-4c92-80e6-64ee3e8e11bb tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 628.145110] env[62066]: DEBUG nova.network.neutron [None req-29878964-568e-4c92-80e6-64ee3e8e11bb tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 628.231116] env[62066]: DEBUG nova.policy [None req-29878964-568e-4c92-80e6-64ee3e8e11bb tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '52f785738b1e4ea287bce066aab4db36', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '45d7a77f838249699a6f6f303b3d6670', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 628.442780] env[62066]: INFO nova.scheduler.client.report [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Deleted allocations for instance 5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962 [ 628.647212] env[62066]: DEBUG nova.compute.manager [None req-29878964-568e-4c92-80e6-64ee3e8e11bb tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 628.680394] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f262f034-5820-4e55-af6c-d1cef4b18013 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.689297] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d22e26b3-8870-4f84-b967-889582aa2a63 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.725744] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90374467-94c4-45cd-8caa-425fa159d1df {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.740291] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-468e3671-7be9-43de-a92b-cbad14c92007 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.756879] env[62066]: DEBUG nova.compute.provider_tree [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 628.789338] env[62066]: DEBUG nova.network.neutron [None req-29878964-568e-4c92-80e6-64ee3e8e11bb tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] Successfully created port: d26b2533-b09b-4f8f-a9a3-31ed85d82aa4 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 628.953458] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Lock "5dcebdc4-5b1f-4dad-aca1-2bdcc68bc962" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 103.704s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 629.260459] env[62066]: DEBUG nova.scheduler.client.report [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 629.458217] env[62066]: DEBUG nova.compute.manager [None req-4d395966-f335-497d-b8a0-23e22d9b79f7 tempest-ServerActionsTestOtherA-1520864800 tempest-ServerActionsTestOtherA-1520864800-project-member] [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 629.660946] env[62066]: DEBUG nova.compute.manager [None req-29878964-568e-4c92-80e6-64ee3e8e11bb tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 629.688497] env[62066]: DEBUG nova.virt.hardware [None req-29878964-568e-4c92-80e6-64ee3e8e11bb tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-23T13:40:41Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-23T13:40:23Z,direct_url=,disk_format='vmdk',id=50ff584c-3b50-4395-af07-3e66769bc9f7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='eb52f7069a374c61ae946f052007c6d9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-23T13:40:24Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 629.692029] env[62066]: DEBUG nova.virt.hardware [None req-29878964-568e-4c92-80e6-64ee3e8e11bb tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 629.692029] env[62066]: DEBUG nova.virt.hardware [None req-29878964-568e-4c92-80e6-64ee3e8e11bb tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 629.692029] env[62066]: DEBUG nova.virt.hardware [None req-29878964-568e-4c92-80e6-64ee3e8e11bb tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 629.692029] env[62066]: DEBUG nova.virt.hardware [None req-29878964-568e-4c92-80e6-64ee3e8e11bb tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 629.692029] env[62066]: DEBUG nova.virt.hardware [None req-29878964-568e-4c92-80e6-64ee3e8e11bb tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 629.692278] env[62066]: DEBUG nova.virt.hardware [None req-29878964-568e-4c92-80e6-64ee3e8e11bb tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 629.692278] env[62066]: DEBUG nova.virt.hardware [None req-29878964-568e-4c92-80e6-64ee3e8e11bb tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 629.692278] env[62066]: DEBUG nova.virt.hardware [None req-29878964-568e-4c92-80e6-64ee3e8e11bb tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 629.692278] env[62066]: DEBUG nova.virt.hardware [None req-29878964-568e-4c92-80e6-64ee3e8e11bb tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 629.692278] env[62066]: DEBUG nova.virt.hardware [None req-29878964-568e-4c92-80e6-64ee3e8e11bb tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 629.693052] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-618bc32b-aada-4890-a738-e8d0d7095de5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.700889] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1226142-ba14-49d0-8555-74573fe390cb {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.766150] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.132s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 629.767371] env[62066]: ERROR nova.compute.manager [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port fd8616a9-d9b8-42e8-b0d3-4ca9121e33e7, please check neutron logs for more information. [ 629.767371] env[62066]: ERROR nova.compute.manager [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] Traceback (most recent call last): [ 629.767371] env[62066]: ERROR nova.compute.manager [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 629.767371] env[62066]: ERROR nova.compute.manager [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] self.driver.spawn(context, instance, image_meta, [ 629.767371] env[62066]: ERROR nova.compute.manager [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 629.767371] env[62066]: ERROR nova.compute.manager [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] self._vmops.spawn(context, instance, image_meta, injected_files, [ 629.767371] env[62066]: ERROR nova.compute.manager [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 629.767371] env[62066]: ERROR nova.compute.manager [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] vm_ref = self.build_virtual_machine(instance, [ 629.767371] env[62066]: ERROR nova.compute.manager [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 629.767371] env[62066]: ERROR nova.compute.manager [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] vif_infos = vmwarevif.get_vif_info(self._session, [ 629.767371] env[62066]: ERROR nova.compute.manager [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 629.767891] env[62066]: ERROR nova.compute.manager [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] for vif in network_info: [ 629.767891] env[62066]: ERROR nova.compute.manager [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 629.767891] env[62066]: ERROR nova.compute.manager [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] return self._sync_wrapper(fn, *args, **kwargs) [ 629.767891] env[62066]: ERROR nova.compute.manager [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 629.767891] env[62066]: ERROR nova.compute.manager [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] self.wait() [ 629.767891] env[62066]: ERROR nova.compute.manager [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 629.767891] env[62066]: ERROR nova.compute.manager [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] self[:] = self._gt.wait() [ 629.767891] env[62066]: ERROR nova.compute.manager [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 629.767891] env[62066]: ERROR nova.compute.manager [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] return self._exit_event.wait() [ 629.767891] env[62066]: ERROR nova.compute.manager [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 629.767891] env[62066]: ERROR nova.compute.manager [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] result = hub.switch() [ 629.767891] env[62066]: ERROR nova.compute.manager [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 629.767891] env[62066]: ERROR nova.compute.manager [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] return self.greenlet.switch() [ 629.768492] env[62066]: ERROR nova.compute.manager [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 629.768492] env[62066]: ERROR nova.compute.manager [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] result = function(*args, **kwargs) [ 629.768492] env[62066]: ERROR nova.compute.manager [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 629.768492] env[62066]: ERROR nova.compute.manager [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] return func(*args, **kwargs) [ 629.768492] env[62066]: ERROR nova.compute.manager [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 629.768492] env[62066]: ERROR nova.compute.manager [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] raise e [ 629.768492] env[62066]: ERROR nova.compute.manager [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 629.768492] env[62066]: ERROR nova.compute.manager [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] nwinfo = self.network_api.allocate_for_instance( [ 629.768492] env[62066]: ERROR nova.compute.manager [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 629.768492] env[62066]: ERROR nova.compute.manager [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] created_port_ids = self._update_ports_for_instance( [ 629.768492] env[62066]: ERROR nova.compute.manager [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 629.768492] env[62066]: ERROR nova.compute.manager [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] with excutils.save_and_reraise_exception(): [ 629.768492] env[62066]: ERROR nova.compute.manager [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 629.769241] env[62066]: ERROR nova.compute.manager [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] self.force_reraise() [ 629.769241] env[62066]: ERROR nova.compute.manager [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 629.769241] env[62066]: ERROR nova.compute.manager [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] raise self.value [ 629.769241] env[62066]: ERROR nova.compute.manager [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 629.769241] env[62066]: ERROR nova.compute.manager [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] updated_port = self._update_port( [ 629.769241] env[62066]: ERROR nova.compute.manager [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 629.769241] env[62066]: ERROR nova.compute.manager [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] _ensure_no_port_binding_failure(port) [ 629.769241] env[62066]: ERROR nova.compute.manager [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 629.769241] env[62066]: ERROR nova.compute.manager [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] raise exception.PortBindingFailed(port_id=port['id']) [ 629.769241] env[62066]: ERROR nova.compute.manager [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] nova.exception.PortBindingFailed: Binding failed for port fd8616a9-d9b8-42e8-b0d3-4ca9121e33e7, please check neutron logs for more information. [ 629.769241] env[62066]: ERROR nova.compute.manager [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] [ 629.769512] env[62066]: DEBUG nova.compute.utils [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] Binding failed for port fd8616a9-d9b8-42e8-b0d3-4ca9121e33e7, please check neutron logs for more information. {{(pid=62066) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 629.769996] env[62066]: DEBUG oslo_concurrency.lockutils [None req-14085d2f-7e58-404a-a2d0-4487ac658e45 tempest-FloatingIPsAssociationTestJSON-608653087 tempest-FloatingIPsAssociationTestJSON-608653087-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.583s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 629.776552] env[62066]: DEBUG nova.compute.manager [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] Build of instance 72c2dbe0-78de-4a7d-98df-cf405a5f20e2 was re-scheduled: Binding failed for port fd8616a9-d9b8-42e8-b0d3-4ca9121e33e7, please check neutron logs for more information. {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 629.776552] env[62066]: DEBUG nova.compute.manager [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] Unplugging VIFs for instance {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 629.776552] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Acquiring lock "refresh_cache-72c2dbe0-78de-4a7d-98df-cf405a5f20e2" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 629.776552] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Acquired lock "refresh_cache-72c2dbe0-78de-4a7d-98df-cf405a5f20e2" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 629.776702] env[62066]: DEBUG nova.network.neutron [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 629.980755] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4d395966-f335-497d-b8a0-23e22d9b79f7 tempest-ServerActionsTestOtherA-1520864800 tempest-ServerActionsTestOtherA-1520864800-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 630.015562] env[62066]: DEBUG nova.compute.manager [req-fff39c4c-2abd-4c7a-b756-b0cc765331e5 req-90a2a7e8-f401-4d7b-be71-8b9937946744 service nova] [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] Received event network-changed-d26b2533-b09b-4f8f-a9a3-31ed85d82aa4 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 630.015787] env[62066]: DEBUG nova.compute.manager [req-fff39c4c-2abd-4c7a-b756-b0cc765331e5 req-90a2a7e8-f401-4d7b-be71-8b9937946744 service nova] [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] Refreshing instance network info cache due to event network-changed-d26b2533-b09b-4f8f-a9a3-31ed85d82aa4. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 630.015972] env[62066]: DEBUG oslo_concurrency.lockutils [req-fff39c4c-2abd-4c7a-b756-b0cc765331e5 req-90a2a7e8-f401-4d7b-be71-8b9937946744 service nova] Acquiring lock "refresh_cache-0e53c28c-a440-4e0e-9a4d-ff1d52557752" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 630.016163] env[62066]: DEBUG oslo_concurrency.lockutils [req-fff39c4c-2abd-4c7a-b756-b0cc765331e5 req-90a2a7e8-f401-4d7b-be71-8b9937946744 service nova] Acquired lock "refresh_cache-0e53c28c-a440-4e0e-9a4d-ff1d52557752" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 630.016288] env[62066]: DEBUG nova.network.neutron [req-fff39c4c-2abd-4c7a-b756-b0cc765331e5 req-90a2a7e8-f401-4d7b-be71-8b9937946744 service nova] [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] Refreshing network info cache for port d26b2533-b09b-4f8f-a9a3-31ed85d82aa4 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 630.214735] env[62066]: ERROR nova.compute.manager [None req-29878964-568e-4c92-80e6-64ee3e8e11bb tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port d26b2533-b09b-4f8f-a9a3-31ed85d82aa4, please check neutron logs for more information. [ 630.214735] env[62066]: ERROR nova.compute.manager Traceback (most recent call last): [ 630.214735] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 630.214735] env[62066]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 630.214735] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 630.214735] env[62066]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 630.214735] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 630.214735] env[62066]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 630.214735] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 630.214735] env[62066]: ERROR nova.compute.manager self.force_reraise() [ 630.214735] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 630.214735] env[62066]: ERROR nova.compute.manager raise self.value [ 630.214735] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 630.214735] env[62066]: ERROR nova.compute.manager updated_port = self._update_port( [ 630.214735] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 630.214735] env[62066]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 630.215325] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 630.215325] env[62066]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 630.215325] env[62066]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port d26b2533-b09b-4f8f-a9a3-31ed85d82aa4, please check neutron logs for more information. [ 630.215325] env[62066]: ERROR nova.compute.manager [ 630.215325] env[62066]: Traceback (most recent call last): [ 630.215325] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 630.215325] env[62066]: listener.cb(fileno) [ 630.215325] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 630.215325] env[62066]: result = function(*args, **kwargs) [ 630.215325] env[62066]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 630.215325] env[62066]: return func(*args, **kwargs) [ 630.215325] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 630.215325] env[62066]: raise e [ 630.215325] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 630.215325] env[62066]: nwinfo = self.network_api.allocate_for_instance( [ 630.215325] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 630.215325] env[62066]: created_port_ids = self._update_ports_for_instance( [ 630.215325] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 630.215325] env[62066]: with excutils.save_and_reraise_exception(): [ 630.215325] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 630.215325] env[62066]: self.force_reraise() [ 630.215325] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 630.215325] env[62066]: raise self.value [ 630.215325] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 630.215325] env[62066]: updated_port = self._update_port( [ 630.215325] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 630.215325] env[62066]: _ensure_no_port_binding_failure(port) [ 630.215325] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 630.215325] env[62066]: raise exception.PortBindingFailed(port_id=port['id']) [ 630.216296] env[62066]: nova.exception.PortBindingFailed: Binding failed for port d26b2533-b09b-4f8f-a9a3-31ed85d82aa4, please check neutron logs for more information. [ 630.216296] env[62066]: Removing descriptor: 20 [ 630.216296] env[62066]: ERROR nova.compute.manager [None req-29878964-568e-4c92-80e6-64ee3e8e11bb tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port d26b2533-b09b-4f8f-a9a3-31ed85d82aa4, please check neutron logs for more information. [ 630.216296] env[62066]: ERROR nova.compute.manager [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] Traceback (most recent call last): [ 630.216296] env[62066]: ERROR nova.compute.manager [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 630.216296] env[62066]: ERROR nova.compute.manager [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] yield resources [ 630.216296] env[62066]: ERROR nova.compute.manager [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 630.216296] env[62066]: ERROR nova.compute.manager [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] self.driver.spawn(context, instance, image_meta, [ 630.216296] env[62066]: ERROR nova.compute.manager [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 630.216296] env[62066]: ERROR nova.compute.manager [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] self._vmops.spawn(context, instance, image_meta, injected_files, [ 630.216296] env[62066]: ERROR nova.compute.manager [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 630.216296] env[62066]: ERROR nova.compute.manager [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] vm_ref = self.build_virtual_machine(instance, [ 630.216664] env[62066]: ERROR nova.compute.manager [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 630.216664] env[62066]: ERROR nova.compute.manager [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] vif_infos = vmwarevif.get_vif_info(self._session, [ 630.216664] env[62066]: ERROR nova.compute.manager [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 630.216664] env[62066]: ERROR nova.compute.manager [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] for vif in network_info: [ 630.216664] env[62066]: ERROR nova.compute.manager [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 630.216664] env[62066]: ERROR nova.compute.manager [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] return self._sync_wrapper(fn, *args, **kwargs) [ 630.216664] env[62066]: ERROR nova.compute.manager [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 630.216664] env[62066]: ERROR nova.compute.manager [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] self.wait() [ 630.216664] env[62066]: ERROR nova.compute.manager [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 630.216664] env[62066]: ERROR nova.compute.manager [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] self[:] = self._gt.wait() [ 630.216664] env[62066]: ERROR nova.compute.manager [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 630.216664] env[62066]: ERROR nova.compute.manager [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] return self._exit_event.wait() [ 630.216664] env[62066]: ERROR nova.compute.manager [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 630.217064] env[62066]: ERROR nova.compute.manager [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] result = hub.switch() [ 630.217064] env[62066]: ERROR nova.compute.manager [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 630.217064] env[62066]: ERROR nova.compute.manager [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] return self.greenlet.switch() [ 630.217064] env[62066]: ERROR nova.compute.manager [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 630.217064] env[62066]: ERROR nova.compute.manager [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] result = function(*args, **kwargs) [ 630.217064] env[62066]: ERROR nova.compute.manager [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 630.217064] env[62066]: ERROR nova.compute.manager [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] return func(*args, **kwargs) [ 630.217064] env[62066]: ERROR nova.compute.manager [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 630.217064] env[62066]: ERROR nova.compute.manager [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] raise e [ 630.217064] env[62066]: ERROR nova.compute.manager [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 630.217064] env[62066]: ERROR nova.compute.manager [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] nwinfo = self.network_api.allocate_for_instance( [ 630.217064] env[62066]: ERROR nova.compute.manager [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 630.217064] env[62066]: ERROR nova.compute.manager [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] created_port_ids = self._update_ports_for_instance( [ 630.217456] env[62066]: ERROR nova.compute.manager [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 630.217456] env[62066]: ERROR nova.compute.manager [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] with excutils.save_and_reraise_exception(): [ 630.217456] env[62066]: ERROR nova.compute.manager [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 630.217456] env[62066]: ERROR nova.compute.manager [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] self.force_reraise() [ 630.217456] env[62066]: ERROR nova.compute.manager [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 630.217456] env[62066]: ERROR nova.compute.manager [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] raise self.value [ 630.217456] env[62066]: ERROR nova.compute.manager [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 630.217456] env[62066]: ERROR nova.compute.manager [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] updated_port = self._update_port( [ 630.217456] env[62066]: ERROR nova.compute.manager [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 630.217456] env[62066]: ERROR nova.compute.manager [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] _ensure_no_port_binding_failure(port) [ 630.217456] env[62066]: ERROR nova.compute.manager [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 630.217456] env[62066]: ERROR nova.compute.manager [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] raise exception.PortBindingFailed(port_id=port['id']) [ 630.218481] env[62066]: ERROR nova.compute.manager [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] nova.exception.PortBindingFailed: Binding failed for port d26b2533-b09b-4f8f-a9a3-31ed85d82aa4, please check neutron logs for more information. [ 630.218481] env[62066]: ERROR nova.compute.manager [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] [ 630.218481] env[62066]: INFO nova.compute.manager [None req-29878964-568e-4c92-80e6-64ee3e8e11bb tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] Terminating instance [ 630.218784] env[62066]: DEBUG oslo_concurrency.lockutils [None req-29878964-568e-4c92-80e6-64ee3e8e11bb tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] Acquiring lock "refresh_cache-0e53c28c-a440-4e0e-9a4d-ff1d52557752" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 630.302574] env[62066]: DEBUG nova.network.neutron [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 630.465253] env[62066]: DEBUG nova.network.neutron [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 630.537804] env[62066]: DEBUG nova.network.neutron [req-fff39c4c-2abd-4c7a-b756-b0cc765331e5 req-90a2a7e8-f401-4d7b-be71-8b9937946744 service nova] [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 630.594935] env[62066]: DEBUG nova.network.neutron [req-fff39c4c-2abd-4c7a-b756-b0cc765331e5 req-90a2a7e8-f401-4d7b-be71-8b9937946744 service nova] [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 630.672091] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9b1e489-2603-4ffc-8453-a1d8b5eb8876 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.680793] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-199b58f1-1c2a-4000-9808-13cbf91205e7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.713575] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3bf4f89-cbda-4a0b-b90f-bb29db6b0ebd {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.721715] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7533f4d2-808b-4aea-8453-02a414d07be5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.736902] env[62066]: DEBUG nova.compute.provider_tree [None req-14085d2f-7e58-404a-a2d0-4487ac658e45 tempest-FloatingIPsAssociationTestJSON-608653087 tempest-FloatingIPsAssociationTestJSON-608653087-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 630.968158] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Releasing lock "refresh_cache-72c2dbe0-78de-4a7d-98df-cf405a5f20e2" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 630.968416] env[62066]: DEBUG nova.compute.manager [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 630.968635] env[62066]: DEBUG nova.compute.manager [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 630.968767] env[62066]: DEBUG nova.network.neutron [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 630.987729] env[62066]: DEBUG nova.network.neutron [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 631.098157] env[62066]: DEBUG oslo_concurrency.lockutils [req-fff39c4c-2abd-4c7a-b756-b0cc765331e5 req-90a2a7e8-f401-4d7b-be71-8b9937946744 service nova] Releasing lock "refresh_cache-0e53c28c-a440-4e0e-9a4d-ff1d52557752" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 631.098603] env[62066]: DEBUG oslo_concurrency.lockutils [None req-29878964-568e-4c92-80e6-64ee3e8e11bb tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] Acquired lock "refresh_cache-0e53c28c-a440-4e0e-9a4d-ff1d52557752" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 631.098805] env[62066]: DEBUG nova.network.neutron [None req-29878964-568e-4c92-80e6-64ee3e8e11bb tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 631.240117] env[62066]: DEBUG nova.scheduler.client.report [None req-14085d2f-7e58-404a-a2d0-4487ac658e45 tempest-FloatingIPsAssociationTestJSON-608653087 tempest-FloatingIPsAssociationTestJSON-608653087-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 631.275903] env[62066]: DEBUG oslo_concurrency.lockutils [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Acquiring lock "0ebba3b2-f82d-4c1d-b01a-f75c3559c117" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 631.276170] env[62066]: DEBUG oslo_concurrency.lockutils [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Lock "0ebba3b2-f82d-4c1d-b01a-f75c3559c117" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 631.298880] env[62066]: DEBUG oslo_concurrency.lockutils [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Acquiring lock "98fb270c-f3f8-4375-8b5c-c2279305d476" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 631.299105] env[62066]: DEBUG oslo_concurrency.lockutils [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Lock "98fb270c-f3f8-4375-8b5c-c2279305d476" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 631.489927] env[62066]: DEBUG nova.network.neutron [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 631.614983] env[62066]: DEBUG nova.network.neutron [None req-29878964-568e-4c92-80e6-64ee3e8e11bb tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 631.678657] env[62066]: DEBUG nova.network.neutron [None req-29878964-568e-4c92-80e6-64ee3e8e11bb tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 631.744826] env[62066]: DEBUG oslo_concurrency.lockutils [None req-14085d2f-7e58-404a-a2d0-4487ac658e45 tempest-FloatingIPsAssociationTestJSON-608653087 tempest-FloatingIPsAssociationTestJSON-608653087-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.975s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 631.745581] env[62066]: ERROR nova.compute.manager [None req-14085d2f-7e58-404a-a2d0-4487ac658e45 tempest-FloatingIPsAssociationTestJSON-608653087 tempest-FloatingIPsAssociationTestJSON-608653087-project-member] [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port d641e931-fa53-4d0a-ad67-1b5304b8fbee, please check neutron logs for more information. [ 631.745581] env[62066]: ERROR nova.compute.manager [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] Traceback (most recent call last): [ 631.745581] env[62066]: ERROR nova.compute.manager [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 631.745581] env[62066]: ERROR nova.compute.manager [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] self.driver.spawn(context, instance, image_meta, [ 631.745581] env[62066]: ERROR nova.compute.manager [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 631.745581] env[62066]: ERROR nova.compute.manager [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] self._vmops.spawn(context, instance, image_meta, injected_files, [ 631.745581] env[62066]: ERROR nova.compute.manager [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 631.745581] env[62066]: ERROR nova.compute.manager [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] vm_ref = self.build_virtual_machine(instance, [ 631.745581] env[62066]: ERROR nova.compute.manager [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 631.745581] env[62066]: ERROR nova.compute.manager [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] vif_infos = vmwarevif.get_vif_info(self._session, [ 631.745581] env[62066]: ERROR nova.compute.manager [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 631.746074] env[62066]: ERROR nova.compute.manager [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] for vif in network_info: [ 631.746074] env[62066]: ERROR nova.compute.manager [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 631.746074] env[62066]: ERROR nova.compute.manager [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] return self._sync_wrapper(fn, *args, **kwargs) [ 631.746074] env[62066]: ERROR nova.compute.manager [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 631.746074] env[62066]: ERROR nova.compute.manager [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] self.wait() [ 631.746074] env[62066]: ERROR nova.compute.manager [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 631.746074] env[62066]: ERROR nova.compute.manager [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] self[:] = self._gt.wait() [ 631.746074] env[62066]: ERROR nova.compute.manager [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 631.746074] env[62066]: ERROR nova.compute.manager [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] return self._exit_event.wait() [ 631.746074] env[62066]: ERROR nova.compute.manager [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 631.746074] env[62066]: ERROR nova.compute.manager [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] result = hub.switch() [ 631.746074] env[62066]: ERROR nova.compute.manager [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 631.746074] env[62066]: ERROR nova.compute.manager [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] return self.greenlet.switch() [ 631.746434] env[62066]: ERROR nova.compute.manager [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 631.746434] env[62066]: ERROR nova.compute.manager [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] result = function(*args, **kwargs) [ 631.746434] env[62066]: ERROR nova.compute.manager [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 631.746434] env[62066]: ERROR nova.compute.manager [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] return func(*args, **kwargs) [ 631.746434] env[62066]: ERROR nova.compute.manager [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 631.746434] env[62066]: ERROR nova.compute.manager [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] raise e [ 631.746434] env[62066]: ERROR nova.compute.manager [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 631.746434] env[62066]: ERROR nova.compute.manager [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] nwinfo = self.network_api.allocate_for_instance( [ 631.746434] env[62066]: ERROR nova.compute.manager [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 631.746434] env[62066]: ERROR nova.compute.manager [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] created_port_ids = self._update_ports_for_instance( [ 631.746434] env[62066]: ERROR nova.compute.manager [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 631.746434] env[62066]: ERROR nova.compute.manager [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] with excutils.save_and_reraise_exception(): [ 631.746434] env[62066]: ERROR nova.compute.manager [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 631.746799] env[62066]: ERROR nova.compute.manager [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] self.force_reraise() [ 631.746799] env[62066]: ERROR nova.compute.manager [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 631.746799] env[62066]: ERROR nova.compute.manager [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] raise self.value [ 631.746799] env[62066]: ERROR nova.compute.manager [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 631.746799] env[62066]: ERROR nova.compute.manager [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] updated_port = self._update_port( [ 631.746799] env[62066]: ERROR nova.compute.manager [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 631.746799] env[62066]: ERROR nova.compute.manager [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] _ensure_no_port_binding_failure(port) [ 631.746799] env[62066]: ERROR nova.compute.manager [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 631.746799] env[62066]: ERROR nova.compute.manager [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] raise exception.PortBindingFailed(port_id=port['id']) [ 631.746799] env[62066]: ERROR nova.compute.manager [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] nova.exception.PortBindingFailed: Binding failed for port d641e931-fa53-4d0a-ad67-1b5304b8fbee, please check neutron logs for more information. [ 631.746799] env[62066]: ERROR nova.compute.manager [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] [ 631.747069] env[62066]: DEBUG nova.compute.utils [None req-14085d2f-7e58-404a-a2d0-4487ac658e45 tempest-FloatingIPsAssociationTestJSON-608653087 tempest-FloatingIPsAssociationTestJSON-608653087-project-member] [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] Binding failed for port d641e931-fa53-4d0a-ad67-1b5304b8fbee, please check neutron logs for more information. {{(pid=62066) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 631.748171] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5e085536-daab-46ff-8eb8-69c1da687135 tempest-ServersAaction247Test-1482016775 tempest-ServersAaction247Test-1482016775-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.266s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 631.749685] env[62066]: INFO nova.compute.claims [None req-5e085536-daab-46ff-8eb8-69c1da687135 tempest-ServersAaction247Test-1482016775 tempest-ServersAaction247Test-1482016775-project-member] [instance: 5d54c3fd-b239-4965-b7a3-5909e8de8bc0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 631.753221] env[62066]: DEBUG nova.compute.manager [None req-14085d2f-7e58-404a-a2d0-4487ac658e45 tempest-FloatingIPsAssociationTestJSON-608653087 tempest-FloatingIPsAssociationTestJSON-608653087-project-member] [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] Build of instance 08d41411-7928-4379-9f2e-c6ce00843d82 was re-scheduled: Binding failed for port d641e931-fa53-4d0a-ad67-1b5304b8fbee, please check neutron logs for more information. {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 631.753529] env[62066]: DEBUG nova.compute.manager [None req-14085d2f-7e58-404a-a2d0-4487ac658e45 tempest-FloatingIPsAssociationTestJSON-608653087 tempest-FloatingIPsAssociationTestJSON-608653087-project-member] [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] Unplugging VIFs for instance {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 631.753750] env[62066]: DEBUG oslo_concurrency.lockutils [None req-14085d2f-7e58-404a-a2d0-4487ac658e45 tempest-FloatingIPsAssociationTestJSON-608653087 tempest-FloatingIPsAssociationTestJSON-608653087-project-member] Acquiring lock "refresh_cache-08d41411-7928-4379-9f2e-c6ce00843d82" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 631.753898] env[62066]: DEBUG oslo_concurrency.lockutils [None req-14085d2f-7e58-404a-a2d0-4487ac658e45 tempest-FloatingIPsAssociationTestJSON-608653087 tempest-FloatingIPsAssociationTestJSON-608653087-project-member] Acquired lock "refresh_cache-08d41411-7928-4379-9f2e-c6ce00843d82" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 631.754067] env[62066]: DEBUG nova.network.neutron [None req-14085d2f-7e58-404a-a2d0-4487ac658e45 tempest-FloatingIPsAssociationTestJSON-608653087 tempest-FloatingIPsAssociationTestJSON-608653087-project-member] [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 631.994331] env[62066]: INFO nova.compute.manager [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 72c2dbe0-78de-4a7d-98df-cf405a5f20e2] Took 1.03 seconds to deallocate network for instance. [ 632.037730] env[62066]: DEBUG nova.compute.manager [req-9f781324-a7df-46ac-98f2-b90acf9d9a67 req-d1ca71cf-6d08-4d88-9b98-6cf3000e0d6a service nova] [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] Received event network-vif-deleted-d26b2533-b09b-4f8f-a9a3-31ed85d82aa4 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 632.181644] env[62066]: DEBUG oslo_concurrency.lockutils [None req-29878964-568e-4c92-80e6-64ee3e8e11bb tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] Releasing lock "refresh_cache-0e53c28c-a440-4e0e-9a4d-ff1d52557752" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 632.182947] env[62066]: DEBUG nova.compute.manager [None req-29878964-568e-4c92-80e6-64ee3e8e11bb tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 632.183203] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-29878964-568e-4c92-80e6-64ee3e8e11bb tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 632.183525] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-25f21012-e7bc-43ef-9a7a-4225939682f6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.193707] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49270117-95ca-41b0-8512-def0444fe418 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.215832] env[62066]: WARNING nova.virt.vmwareapi.vmops [None req-29878964-568e-4c92-80e6-64ee3e8e11bb tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 0e53c28c-a440-4e0e-9a4d-ff1d52557752 could not be found. [ 632.216083] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-29878964-568e-4c92-80e6-64ee3e8e11bb tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 632.216271] env[62066]: INFO nova.compute.manager [None req-29878964-568e-4c92-80e6-64ee3e8e11bb tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] Took 0.03 seconds to destroy the instance on the hypervisor. [ 632.216518] env[62066]: DEBUG oslo.service.loopingcall [None req-29878964-568e-4c92-80e6-64ee3e8e11bb tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 632.216732] env[62066]: DEBUG nova.compute.manager [-] [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 632.216824] env[62066]: DEBUG nova.network.neutron [-] [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 632.237228] env[62066]: DEBUG nova.network.neutron [-] [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 632.274574] env[62066]: DEBUG nova.network.neutron [None req-14085d2f-7e58-404a-a2d0-4487ac658e45 tempest-FloatingIPsAssociationTestJSON-608653087 tempest-FloatingIPsAssociationTestJSON-608653087-project-member] [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 632.411276] env[62066]: DEBUG nova.network.neutron [None req-14085d2f-7e58-404a-a2d0-4487ac658e45 tempest-FloatingIPsAssociationTestJSON-608653087 tempest-FloatingIPsAssociationTestJSON-608653087-project-member] [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 632.740806] env[62066]: DEBUG nova.network.neutron [-] [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 632.913294] env[62066]: DEBUG oslo_concurrency.lockutils [None req-14085d2f-7e58-404a-a2d0-4487ac658e45 tempest-FloatingIPsAssociationTestJSON-608653087 tempest-FloatingIPsAssociationTestJSON-608653087-project-member] Releasing lock "refresh_cache-08d41411-7928-4379-9f2e-c6ce00843d82" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 632.913534] env[62066]: DEBUG nova.compute.manager [None req-14085d2f-7e58-404a-a2d0-4487ac658e45 tempest-FloatingIPsAssociationTestJSON-608653087 tempest-FloatingIPsAssociationTestJSON-608653087-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 632.913716] env[62066]: DEBUG nova.compute.manager [None req-14085d2f-7e58-404a-a2d0-4487ac658e45 tempest-FloatingIPsAssociationTestJSON-608653087 tempest-FloatingIPsAssociationTestJSON-608653087-project-member] [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 632.913889] env[62066]: DEBUG nova.network.neutron [None req-14085d2f-7e58-404a-a2d0-4487ac658e45 tempest-FloatingIPsAssociationTestJSON-608653087 tempest-FloatingIPsAssociationTestJSON-608653087-project-member] [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 632.938646] env[62066]: DEBUG nova.network.neutron [None req-14085d2f-7e58-404a-a2d0-4487ac658e45 tempest-FloatingIPsAssociationTestJSON-608653087 tempest-FloatingIPsAssociationTestJSON-608653087-project-member] [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 633.165596] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad2af671-a424-46f1-b09d-263c1ea5b3f3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.172888] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3bcbedd-24b6-4ac8-b7eb-8072b3755da1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.203504] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84a5f624-0766-4ded-8951-c2b2e46ef8d7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.210965] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8535dafb-4bf0-418f-b984-6c5d1f2753e6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.223986] env[62066]: DEBUG nova.compute.provider_tree [None req-5e085536-daab-46ff-8eb8-69c1da687135 tempest-ServersAaction247Test-1482016775 tempest-ServersAaction247Test-1482016775-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 633.244525] env[62066]: INFO nova.compute.manager [-] [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] Took 1.03 seconds to deallocate network for instance. [ 633.246515] env[62066]: DEBUG nova.compute.claims [None req-29878964-568e-4c92-80e6-64ee3e8e11bb tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] Aborting claim: {{(pid=62066) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 633.246705] env[62066]: DEBUG oslo_concurrency.lockutils [None req-29878964-568e-4c92-80e6-64ee3e8e11bb tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 633.444316] env[62066]: DEBUG nova.network.neutron [None req-14085d2f-7e58-404a-a2d0-4487ac658e45 tempest-FloatingIPsAssociationTestJSON-608653087 tempest-FloatingIPsAssociationTestJSON-608653087-project-member] [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 633.727406] env[62066]: DEBUG nova.scheduler.client.report [None req-5e085536-daab-46ff-8eb8-69c1da687135 tempest-ServersAaction247Test-1482016775 tempest-ServersAaction247Test-1482016775-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 633.946562] env[62066]: INFO nova.compute.manager [None req-14085d2f-7e58-404a-a2d0-4487ac658e45 tempest-FloatingIPsAssociationTestJSON-608653087 tempest-FloatingIPsAssociationTestJSON-608653087-project-member] [instance: 08d41411-7928-4379-9f2e-c6ce00843d82] Took 1.03 seconds to deallocate network for instance. [ 634.028640] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b8913701-5021-4640-96ed-5162f5a7e9d3 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Lock "72c2dbe0-78de-4a7d-98df-cf405a5f20e2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 108.739s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 634.029300] env[62066]: Traceback (most recent call last): [ 634.029357] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 634.029357] env[62066]: self.driver.spawn(context, instance, image_meta, [ 634.029357] env[62066]: File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 634.029357] env[62066]: self._vmops.spawn(context, instance, image_meta, injected_files, [ 634.029357] env[62066]: File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 634.029357] env[62066]: vm_ref = self.build_virtual_machine(instance, [ 634.029357] env[62066]: File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 634.029357] env[62066]: vif_infos = vmwarevif.get_vif_info(self._session, [ 634.029357] env[62066]: File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 634.029357] env[62066]: for vif in network_info: [ 634.029357] env[62066]: File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 634.029357] env[62066]: return self._sync_wrapper(fn, *args, **kwargs) [ 634.029357] env[62066]: File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 634.029357] env[62066]: self.wait() [ 634.029357] env[62066]: File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 634.029357] env[62066]: self[:] = self._gt.wait() [ 634.029357] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 634.029357] env[62066]: return self._exit_event.wait() [ 634.029357] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 634.029357] env[62066]: result = hub.switch() [ 634.029856] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 634.029856] env[62066]: return self.greenlet.switch() [ 634.029856] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 634.029856] env[62066]: result = function(*args, **kwargs) [ 634.029856] env[62066]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 634.029856] env[62066]: return func(*args, **kwargs) [ 634.029856] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 634.029856] env[62066]: raise e [ 634.029856] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 634.029856] env[62066]: nwinfo = self.network_api.allocate_for_instance( [ 634.029856] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 634.029856] env[62066]: created_port_ids = self._update_ports_for_instance( [ 634.029856] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 634.029856] env[62066]: with excutils.save_and_reraise_exception(): [ 634.029856] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 634.029856] env[62066]: self.force_reraise() [ 634.029856] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 634.029856] env[62066]: raise self.value [ 634.029856] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 634.029856] env[62066]: updated_port = self._update_port( [ 634.029856] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 634.029856] env[62066]: _ensure_no_port_binding_failure(port) [ 634.029856] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 634.029856] env[62066]: raise exception.PortBindingFailed(port_id=port['id']) [ 634.029856] env[62066]: nova.exception.PortBindingFailed: Binding failed for port fd8616a9-d9b8-42e8-b0d3-4ca9121e33e7, please check neutron logs for more information. [ 634.029856] env[62066]: During handling of the above exception, another exception occurred: [ 634.029856] env[62066]: Traceback (most recent call last): [ 634.029856] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 2456, in _do_build_and_run_instance [ 634.029856] env[62066]: self._build_and_run_instance(context, instance, image, [ 634.029856] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 2748, in _build_and_run_instance [ 634.030971] env[62066]: raise exception.RescheduledException( [ 634.030971] env[62066]: nova.exception.RescheduledException: Build of instance 72c2dbe0-78de-4a7d-98df-cf405a5f20e2 was re-scheduled: Binding failed for port fd8616a9-d9b8-42e8-b0d3-4ca9121e33e7, please check neutron logs for more information. [ 634.030971] env[62066]: During handling of the above exception, another exception occurred: [ 634.030971] env[62066]: Traceback (most recent call last): [ 634.030971] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenpool.py", line 87, in _spawn_n_impl [ 634.030971] env[62066]: func(*args, **kwargs) [ 634.030971] env[62066]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 634.030971] env[62066]: return func(*args, **kwargs) [ 634.030971] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 634.030971] env[62066]: return f(*args, **kwargs) [ 634.030971] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 2347, in _locked_do_build_and_run_instance [ 634.030971] env[62066]: result = self._do_build_and_run_instance(*args, **kwargs) [ 634.030971] env[62066]: File "/opt/stack/nova/nova/exception_wrapper.py", line 65, in wrapped [ 634.030971] env[62066]: with excutils.save_and_reraise_exception(): [ 634.030971] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 634.030971] env[62066]: self.force_reraise() [ 634.030971] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 634.030971] env[62066]: raise self.value [ 634.030971] env[62066]: File "/opt/stack/nova/nova/exception_wrapper.py", line 63, in wrapped [ 634.030971] env[62066]: return f(self, context, *args, **kw) [ 634.030971] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 166, in decorated_function [ 634.030971] env[62066]: with excutils.save_and_reraise_exception(): [ 634.030971] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 634.030971] env[62066]: self.force_reraise() [ 634.030971] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 634.030971] env[62066]: raise self.value [ 634.030971] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 157, in decorated_function [ 634.030971] env[62066]: return function(self, context, *args, **kwargs) [ 634.030971] env[62066]: File "/opt/stack/nova/nova/compute/utils.py", line 1453, in decorated_function [ 634.030971] env[62066]: return function(self, context, *args, **kwargs) [ 634.031984] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 203, in decorated_function [ 634.031984] env[62066]: return function(self, context, *args, **kwargs) [ 634.031984] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 2491, in _do_build_and_run_instance [ 634.031984] env[62066]: instance.save() [ 634.031984] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_versionedobjects/base.py", line 209, in wrapper [ 634.031984] env[62066]: updates, result = self.indirection_api.object_action( [ 634.031984] env[62066]: File "/opt/stack/nova/nova/conductor/rpcapi.py", line 247, in object_action [ 634.031984] env[62066]: return cctxt.call(context, 'object_action', objinst=objinst, [ 634.031984] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/client.py", line 190, in call [ 634.031984] env[62066]: result = self.transport._send( [ 634.031984] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/transport.py", line 123, in _send [ 634.031984] env[62066]: return self._driver.send(target, ctxt, message, [ 634.031984] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 788, in send [ 634.031984] env[62066]: return self._send(target, ctxt, message, wait_for_reply, timeout, [ 634.031984] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 780, in _send [ 634.031984] env[62066]: raise result [ 634.031984] env[62066]: nova.exception_Remote.InstanceNotFound_Remote: Instance 72c2dbe0-78de-4a7d-98df-cf405a5f20e2 could not be found. [ 634.031984] env[62066]: Traceback (most recent call last): [ 634.031984] env[62066]: File "/opt/stack/nova/nova/conductor/manager.py", line 142, in _object_dispatch [ 634.031984] env[62066]: return getattr(target, method)(*args, **kwargs) [ 634.031984] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_versionedobjects/base.py", line 226, in wrapper [ 634.031984] env[62066]: return fn(self, *args, **kwargs) [ 634.031984] env[62066]: File "/opt/stack/nova/nova/objects/instance.py", line 878, in save [ 634.031984] env[62066]: old_ref, inst_ref = db.instance_update_and_get_original( [ 634.031984] env[62066]: File "/opt/stack/nova/nova/db/utils.py", line 35, in wrapper [ 634.031984] env[62066]: return f(*args, **kwargs) [ 634.031984] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/api.py", line 144, in wrapper [ 634.031984] env[62066]: with excutils.save_and_reraise_exception() as ectxt: [ 634.033193] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 634.033193] env[62066]: self.force_reraise() [ 634.033193] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 634.033193] env[62066]: raise self.value [ 634.033193] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/api.py", line 142, in wrapper [ 634.033193] env[62066]: return f(*args, **kwargs) [ 634.033193] env[62066]: File "/opt/stack/nova/nova/db/main/api.py", line 207, in wrapper [ 634.033193] env[62066]: return f(context, *args, **kwargs) [ 634.033193] env[62066]: File "/opt/stack/nova/nova/db/main/api.py", line 2283, in instance_update_and_get_original [ 634.033193] env[62066]: instance_ref = _instance_get_by_uuid(context, instance_uuid, [ 634.033193] env[62066]: File "/opt/stack/nova/nova/db/main/api.py", line 1405, in _instance_get_by_uuid [ 634.033193] env[62066]: raise exception.InstanceNotFound(instance_id=uuid) [ 634.033193] env[62066]: nova.exception.InstanceNotFound: Instance 72c2dbe0-78de-4a7d-98df-cf405a5f20e2 could not be found. [ 634.232848] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5e085536-daab-46ff-8eb8-69c1da687135 tempest-ServersAaction247Test-1482016775 tempest-ServersAaction247Test-1482016775-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.485s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 634.233453] env[62066]: DEBUG nova.compute.manager [None req-5e085536-daab-46ff-8eb8-69c1da687135 tempest-ServersAaction247Test-1482016775 tempest-ServersAaction247Test-1482016775-project-member] [instance: 5d54c3fd-b239-4965-b7a3-5909e8de8bc0] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 634.236102] env[62066]: DEBUG oslo_concurrency.lockutils [None req-2c107b45-3cbf-4be9-af5a-796fbdb9b5fe tempest-ImagesNegativeTestJSON-881899573 tempest-ImagesNegativeTestJSON-881899573-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.282s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 634.237469] env[62066]: INFO nova.compute.claims [None req-2c107b45-3cbf-4be9-af5a-796fbdb9b5fe tempest-ImagesNegativeTestJSON-881899573 tempest-ImagesNegativeTestJSON-881899573-project-member] [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 634.531608] env[62066]: DEBUG nova.compute.manager [None req-cf333fac-d104-4122-9317-eab87c4499f0 tempest-ServerDiagnosticsTest-1737887255 tempest-ServerDiagnosticsTest-1737887255-project-member] [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 634.743128] env[62066]: DEBUG nova.compute.utils [None req-5e085536-daab-46ff-8eb8-69c1da687135 tempest-ServersAaction247Test-1482016775 tempest-ServersAaction247Test-1482016775-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 634.746526] env[62066]: DEBUG nova.compute.manager [None req-5e085536-daab-46ff-8eb8-69c1da687135 tempest-ServersAaction247Test-1482016775 tempest-ServersAaction247Test-1482016775-project-member] [instance: 5d54c3fd-b239-4965-b7a3-5909e8de8bc0] Not allocating networking since 'none' was specified. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 634.978143] env[62066]: INFO nova.scheduler.client.report [None req-14085d2f-7e58-404a-a2d0-4487ac658e45 tempest-FloatingIPsAssociationTestJSON-608653087 tempest-FloatingIPsAssociationTestJSON-608653087-project-member] Deleted allocations for instance 08d41411-7928-4379-9f2e-c6ce00843d82 [ 635.050781] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cf333fac-d104-4122-9317-eab87c4499f0 tempest-ServerDiagnosticsTest-1737887255 tempest-ServerDiagnosticsTest-1737887255-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 635.247750] env[62066]: DEBUG nova.compute.manager [None req-5e085536-daab-46ff-8eb8-69c1da687135 tempest-ServersAaction247Test-1482016775 tempest-ServersAaction247Test-1482016775-project-member] [instance: 5d54c3fd-b239-4965-b7a3-5909e8de8bc0] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 635.487895] env[62066]: DEBUG oslo_concurrency.lockutils [None req-14085d2f-7e58-404a-a2d0-4487ac658e45 tempest-FloatingIPsAssociationTestJSON-608653087 tempest-FloatingIPsAssociationTestJSON-608653087-project-member] Lock "08d41411-7928-4379-9f2e-c6ce00843d82" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 108.009s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 635.648321] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18ebf127-cc4b-4e86-a786-4865dfbf4028 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.655882] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2d06c94-c795-437c-9781-929cef3758cf {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.684935] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abad00d7-27bc-4549-b4b5-c39e7dd019db {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.692502] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb56fb71-f55e-4bea-92d9-718db8f4e719 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.707260] env[62066]: DEBUG nova.compute.provider_tree [None req-2c107b45-3cbf-4be9-af5a-796fbdb9b5fe tempest-ImagesNegativeTestJSON-881899573 tempest-ImagesNegativeTestJSON-881899573-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 635.997438] env[62066]: DEBUG nova.compute.manager [None req-67a0ccea-4cde-4de9-9aa4-1bb67e83772e tempest-ServersNegativeTestJSON-789906455 tempest-ServersNegativeTestJSON-789906455-project-member] [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 636.214070] env[62066]: DEBUG nova.scheduler.client.report [None req-2c107b45-3cbf-4be9-af5a-796fbdb9b5fe tempest-ImagesNegativeTestJSON-881899573 tempest-ImagesNegativeTestJSON-881899573-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 636.263405] env[62066]: DEBUG nova.compute.manager [None req-5e085536-daab-46ff-8eb8-69c1da687135 tempest-ServersAaction247Test-1482016775 tempest-ServersAaction247Test-1482016775-project-member] [instance: 5d54c3fd-b239-4965-b7a3-5909e8de8bc0] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 636.294843] env[62066]: DEBUG nova.virt.hardware [None req-5e085536-daab-46ff-8eb8-69c1da687135 tempest-ServersAaction247Test-1482016775 tempest-ServersAaction247Test-1482016775-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-23T13:40:41Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-23T13:40:23Z,direct_url=,disk_format='vmdk',id=50ff584c-3b50-4395-af07-3e66769bc9f7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='eb52f7069a374c61ae946f052007c6d9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-23T13:40:24Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 636.295109] env[62066]: DEBUG nova.virt.hardware [None req-5e085536-daab-46ff-8eb8-69c1da687135 tempest-ServersAaction247Test-1482016775 tempest-ServersAaction247Test-1482016775-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 636.295269] env[62066]: DEBUG nova.virt.hardware [None req-5e085536-daab-46ff-8eb8-69c1da687135 tempest-ServersAaction247Test-1482016775 tempest-ServersAaction247Test-1482016775-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 636.295457] env[62066]: DEBUG nova.virt.hardware [None req-5e085536-daab-46ff-8eb8-69c1da687135 tempest-ServersAaction247Test-1482016775 tempest-ServersAaction247Test-1482016775-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 636.295717] env[62066]: DEBUG nova.virt.hardware [None req-5e085536-daab-46ff-8eb8-69c1da687135 tempest-ServersAaction247Test-1482016775 tempest-ServersAaction247Test-1482016775-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 636.295886] env[62066]: DEBUG nova.virt.hardware [None req-5e085536-daab-46ff-8eb8-69c1da687135 tempest-ServersAaction247Test-1482016775 tempest-ServersAaction247Test-1482016775-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 636.296115] env[62066]: DEBUG nova.virt.hardware [None req-5e085536-daab-46ff-8eb8-69c1da687135 tempest-ServersAaction247Test-1482016775 tempest-ServersAaction247Test-1482016775-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 636.296274] env[62066]: DEBUG nova.virt.hardware [None req-5e085536-daab-46ff-8eb8-69c1da687135 tempest-ServersAaction247Test-1482016775 tempest-ServersAaction247Test-1482016775-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 636.297206] env[62066]: DEBUG nova.virt.hardware [None req-5e085536-daab-46ff-8eb8-69c1da687135 tempest-ServersAaction247Test-1482016775 tempest-ServersAaction247Test-1482016775-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 636.297398] env[62066]: DEBUG nova.virt.hardware [None req-5e085536-daab-46ff-8eb8-69c1da687135 tempest-ServersAaction247Test-1482016775 tempest-ServersAaction247Test-1482016775-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 636.297573] env[62066]: DEBUG nova.virt.hardware [None req-5e085536-daab-46ff-8eb8-69c1da687135 tempest-ServersAaction247Test-1482016775 tempest-ServersAaction247Test-1482016775-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 636.298606] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb500729-b468-4ea8-a717-acdef17551b2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.306880] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57522287-ec0c-424b-9e71-b6f227a3d736 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.320467] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-5e085536-daab-46ff-8eb8-69c1da687135 tempest-ServersAaction247Test-1482016775 tempest-ServersAaction247Test-1482016775-project-member] [instance: 5d54c3fd-b239-4965-b7a3-5909e8de8bc0] Instance VIF info [] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 636.325953] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-5e085536-daab-46ff-8eb8-69c1da687135 tempest-ServersAaction247Test-1482016775 tempest-ServersAaction247Test-1482016775-project-member] Creating folder: Project (c482e32cd25f4a11a2ff1b39a3561d64). Parent ref: group-v285980. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 636.326267] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-00c446a4-95b6-4354-a3f7-be7816a4dd1a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.335524] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-5e085536-daab-46ff-8eb8-69c1da687135 tempest-ServersAaction247Test-1482016775 tempest-ServersAaction247Test-1482016775-project-member] Created folder: Project (c482e32cd25f4a11a2ff1b39a3561d64) in parent group-v285980. [ 636.335746] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-5e085536-daab-46ff-8eb8-69c1da687135 tempest-ServersAaction247Test-1482016775 tempest-ServersAaction247Test-1482016775-project-member] Creating folder: Instances. Parent ref: group-v285996. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 636.335976] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6816c51a-5c8a-40fe-bc48-c56507b21453 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.343310] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-5e085536-daab-46ff-8eb8-69c1da687135 tempest-ServersAaction247Test-1482016775 tempest-ServersAaction247Test-1482016775-project-member] Created folder: Instances in parent group-v285996. [ 636.343540] env[62066]: DEBUG oslo.service.loopingcall [None req-5e085536-daab-46ff-8eb8-69c1da687135 tempest-ServersAaction247Test-1482016775 tempest-ServersAaction247Test-1482016775-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 636.343724] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5d54c3fd-b239-4965-b7a3-5909e8de8bc0] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 636.343914] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a211bf9c-2938-4a3f-b257-000d7d39f2b0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.361023] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 636.361023] env[62066]: value = "task-1340661" [ 636.361023] env[62066]: _type = "Task" [ 636.361023] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 636.367670] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1340661, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 636.520852] env[62066]: DEBUG oslo_concurrency.lockutils [None req-67a0ccea-4cde-4de9-9aa4-1bb67e83772e tempest-ServersNegativeTestJSON-789906455 tempest-ServersNegativeTestJSON-789906455-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 636.720128] env[62066]: DEBUG oslo_concurrency.lockutils [None req-2c107b45-3cbf-4be9-af5a-796fbdb9b5fe tempest-ImagesNegativeTestJSON-881899573 tempest-ImagesNegativeTestJSON-881899573-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.481s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 636.720128] env[62066]: DEBUG nova.compute.manager [None req-2c107b45-3cbf-4be9-af5a-796fbdb9b5fe tempest-ImagesNegativeTestJSON-881899573 tempest-ImagesNegativeTestJSON-881899573-project-member] [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 636.721251] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 16.332s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 636.872768] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1340661, 'name': CreateVM_Task, 'duration_secs': 0.245994} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 636.872951] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5d54c3fd-b239-4965-b7a3-5909e8de8bc0] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 636.873536] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5e085536-daab-46ff-8eb8-69c1da687135 tempest-ServersAaction247Test-1482016775 tempest-ServersAaction247Test-1482016775-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 636.873536] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5e085536-daab-46ff-8eb8-69c1da687135 tempest-ServersAaction247Test-1482016775 tempest-ServersAaction247Test-1482016775-project-member] Acquired lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 636.873888] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5e085536-daab-46ff-8eb8-69c1da687135 tempest-ServersAaction247Test-1482016775 tempest-ServersAaction247Test-1482016775-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 636.874150] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0384708d-3f61-49d3-af2a-0bd657001df4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.878538] env[62066]: DEBUG oslo_vmware.api [None req-5e085536-daab-46ff-8eb8-69c1da687135 tempest-ServersAaction247Test-1482016775 tempest-ServersAaction247Test-1482016775-project-member] Waiting for the task: (returnval){ [ 636.878538] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]520d6b67-7cba-f445-631b-43fa741b282a" [ 636.878538] env[62066]: _type = "Task" [ 636.878538] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 636.886537] env[62066]: DEBUG oslo_vmware.api [None req-5e085536-daab-46ff-8eb8-69c1da687135 tempest-ServersAaction247Test-1482016775 tempest-ServersAaction247Test-1482016775-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]520d6b67-7cba-f445-631b-43fa741b282a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 637.231981] env[62066]: DEBUG nova.compute.utils [None req-2c107b45-3cbf-4be9-af5a-796fbdb9b5fe tempest-ImagesNegativeTestJSON-881899573 tempest-ImagesNegativeTestJSON-881899573-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 637.234848] env[62066]: DEBUG nova.compute.manager [None req-2c107b45-3cbf-4be9-af5a-796fbdb9b5fe tempest-ImagesNegativeTestJSON-881899573 tempest-ImagesNegativeTestJSON-881899573-project-member] [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 637.234848] env[62066]: DEBUG nova.network.neutron [None req-2c107b45-3cbf-4be9-af5a-796fbdb9b5fe tempest-ImagesNegativeTestJSON-881899573 tempest-ImagesNegativeTestJSON-881899573-project-member] [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 637.288156] env[62066]: DEBUG nova.policy [None req-2c107b45-3cbf-4be9-af5a-796fbdb9b5fe tempest-ImagesNegativeTestJSON-881899573 tempest-ImagesNegativeTestJSON-881899573-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '524a55542abb4f21887784f61826c9ea', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0b0531f280aa4addb2093f311bce05ae', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 637.392256] env[62066]: DEBUG oslo_vmware.api [None req-5e085536-daab-46ff-8eb8-69c1da687135 tempest-ServersAaction247Test-1482016775 tempest-ServersAaction247Test-1482016775-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]520d6b67-7cba-f445-631b-43fa741b282a, 'name': SearchDatastore_Task, 'duration_secs': 0.0091} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 637.392256] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5e085536-daab-46ff-8eb8-69c1da687135 tempest-ServersAaction247Test-1482016775 tempest-ServersAaction247Test-1482016775-project-member] Releasing lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 637.392256] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-5e085536-daab-46ff-8eb8-69c1da687135 tempest-ServersAaction247Test-1482016775 tempest-ServersAaction247Test-1482016775-project-member] [instance: 5d54c3fd-b239-4965-b7a3-5909e8de8bc0] Processing image 50ff584c-3b50-4395-af07-3e66769bc9f7 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 637.392256] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5e085536-daab-46ff-8eb8-69c1da687135 tempest-ServersAaction247Test-1482016775 tempest-ServersAaction247Test-1482016775-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 637.392486] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5e085536-daab-46ff-8eb8-69c1da687135 tempest-ServersAaction247Test-1482016775 tempest-ServersAaction247Test-1482016775-project-member] Acquired lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 637.392486] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-5e085536-daab-46ff-8eb8-69c1da687135 tempest-ServersAaction247Test-1482016775 tempest-ServersAaction247Test-1482016775-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 637.392486] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a4362294-33f8-470e-8c5f-41f10b18299c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.401097] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-5e085536-daab-46ff-8eb8-69c1da687135 tempest-ServersAaction247Test-1482016775 tempest-ServersAaction247Test-1482016775-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 637.401286] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-5e085536-daab-46ff-8eb8-69c1da687135 tempest-ServersAaction247Test-1482016775 tempest-ServersAaction247Test-1482016775-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 637.402047] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-06a08154-09cc-4b8d-9c30-92b42aec1d9f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.407522] env[62066]: DEBUG oslo_vmware.api [None req-5e085536-daab-46ff-8eb8-69c1da687135 tempest-ServersAaction247Test-1482016775 tempest-ServersAaction247Test-1482016775-project-member] Waiting for the task: (returnval){ [ 637.407522] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]52690669-2942-ba0f-f8da-d39e94a3a42f" [ 637.407522] env[62066]: _type = "Task" [ 637.407522] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 637.419460] env[62066]: DEBUG oslo_vmware.api [None req-5e085536-daab-46ff-8eb8-69c1da687135 tempest-ServersAaction247Test-1482016775 tempest-ServersAaction247Test-1482016775-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52690669-2942-ba0f-f8da-d39e94a3a42f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 637.740326] env[62066]: DEBUG nova.compute.manager [None req-2c107b45-3cbf-4be9-af5a-796fbdb9b5fe tempest-ImagesNegativeTestJSON-881899573 tempest-ImagesNegativeTestJSON-881899573-project-member] [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 637.761558] env[62066]: DEBUG nova.network.neutron [None req-2c107b45-3cbf-4be9-af5a-796fbdb9b5fe tempest-ImagesNegativeTestJSON-881899573 tempest-ImagesNegativeTestJSON-881899573-project-member] [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] Successfully created port: aa9ed432-13ef-49d1-94be-5f4666d101ef {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 637.771307] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Instance 623300f7-54d2-4b0a-b356-a1ae585682cb actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 637.772170] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Instance 0040f46d-9a60-4fa1-8a00-c08022e17df0 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 637.772170] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Instance bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 637.772170] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Instance 0e53c28c-a440-4e0e-9a4d-ff1d52557752 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 637.772170] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Instance 5d54c3fd-b239-4965-b7a3-5909e8de8bc0 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 637.772426] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Instance cb319720-29ab-4ff2-a71e-f77ae8c85735 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 637.918278] env[62066]: DEBUG oslo_vmware.api [None req-5e085536-daab-46ff-8eb8-69c1da687135 tempest-ServersAaction247Test-1482016775 tempest-ServersAaction247Test-1482016775-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52690669-2942-ba0f-f8da-d39e94a3a42f, 'name': SearchDatastore_Task, 'duration_secs': 0.010024} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 637.919136] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-40bdf584-a35c-4a57-bf75-50eb74f9429d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.924913] env[62066]: DEBUG oslo_vmware.api [None req-5e085536-daab-46ff-8eb8-69c1da687135 tempest-ServersAaction247Test-1482016775 tempest-ServersAaction247Test-1482016775-project-member] Waiting for the task: (returnval){ [ 637.924913] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]5219c0c3-a6dd-c2a7-89a7-628d6351bef7" [ 637.924913] env[62066]: _type = "Task" [ 637.924913] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 637.934037] env[62066]: DEBUG oslo_vmware.api [None req-5e085536-daab-46ff-8eb8-69c1da687135 tempest-ServersAaction247Test-1482016775 tempest-ServersAaction247Test-1482016775-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]5219c0c3-a6dd-c2a7-89a7-628d6351bef7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 638.275887] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Instance 97e51e6c-9a3a-4b68-b737-d48090a22b1b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 638.435852] env[62066]: DEBUG oslo_vmware.api [None req-5e085536-daab-46ff-8eb8-69c1da687135 tempest-ServersAaction247Test-1482016775 tempest-ServersAaction247Test-1482016775-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]5219c0c3-a6dd-c2a7-89a7-628d6351bef7, 'name': SearchDatastore_Task, 'duration_secs': 0.039805} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 638.437082] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5e085536-daab-46ff-8eb8-69c1da687135 tempest-ServersAaction247Test-1482016775 tempest-ServersAaction247Test-1482016775-project-member] Releasing lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 638.437474] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-5e085536-daab-46ff-8eb8-69c1da687135 tempest-ServersAaction247Test-1482016775 tempest-ServersAaction247Test-1482016775-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk to [datastore2] 5d54c3fd-b239-4965-b7a3-5909e8de8bc0/5d54c3fd-b239-4965-b7a3-5909e8de8bc0.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 638.437975] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4187d1fd-dd87-4e3a-aaf5-d3ec14faf126 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.446331] env[62066]: DEBUG oslo_vmware.api [None req-5e085536-daab-46ff-8eb8-69c1da687135 tempest-ServersAaction247Test-1482016775 tempest-ServersAaction247Test-1482016775-project-member] Waiting for the task: (returnval){ [ 638.446331] env[62066]: value = "task-1340663" [ 638.446331] env[62066]: _type = "Task" [ 638.446331] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 638.453835] env[62066]: DEBUG oslo_vmware.api [None req-5e085536-daab-46ff-8eb8-69c1da687135 tempest-ServersAaction247Test-1482016775 tempest-ServersAaction247Test-1482016775-project-member] Task: {'id': task-1340663, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 638.750690] env[62066]: DEBUG nova.compute.manager [None req-2c107b45-3cbf-4be9-af5a-796fbdb9b5fe tempest-ImagesNegativeTestJSON-881899573 tempest-ImagesNegativeTestJSON-881899573-project-member] [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 638.779214] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Instance 56984993-a8b1-464e-b20c-79ffe67d2491 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 638.784980] env[62066]: DEBUG nova.virt.hardware [None req-2c107b45-3cbf-4be9-af5a-796fbdb9b5fe tempest-ImagesNegativeTestJSON-881899573 tempest-ImagesNegativeTestJSON-881899573-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-23T13:40:41Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-23T13:40:23Z,direct_url=,disk_format='vmdk',id=50ff584c-3b50-4395-af07-3e66769bc9f7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='eb52f7069a374c61ae946f052007c6d9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-23T13:40:24Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 638.785281] env[62066]: DEBUG nova.virt.hardware [None req-2c107b45-3cbf-4be9-af5a-796fbdb9b5fe tempest-ImagesNegativeTestJSON-881899573 tempest-ImagesNegativeTestJSON-881899573-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 638.785442] env[62066]: DEBUG nova.virt.hardware [None req-2c107b45-3cbf-4be9-af5a-796fbdb9b5fe tempest-ImagesNegativeTestJSON-881899573 tempest-ImagesNegativeTestJSON-881899573-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 638.785892] env[62066]: DEBUG nova.virt.hardware [None req-2c107b45-3cbf-4be9-af5a-796fbdb9b5fe tempest-ImagesNegativeTestJSON-881899573 tempest-ImagesNegativeTestJSON-881899573-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 638.786163] env[62066]: DEBUG nova.virt.hardware [None req-2c107b45-3cbf-4be9-af5a-796fbdb9b5fe tempest-ImagesNegativeTestJSON-881899573 tempest-ImagesNegativeTestJSON-881899573-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 638.786383] env[62066]: DEBUG nova.virt.hardware [None req-2c107b45-3cbf-4be9-af5a-796fbdb9b5fe tempest-ImagesNegativeTestJSON-881899573 tempest-ImagesNegativeTestJSON-881899573-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 638.786648] env[62066]: DEBUG nova.virt.hardware [None req-2c107b45-3cbf-4be9-af5a-796fbdb9b5fe tempest-ImagesNegativeTestJSON-881899573 tempest-ImagesNegativeTestJSON-881899573-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 638.786890] env[62066]: DEBUG nova.virt.hardware [None req-2c107b45-3cbf-4be9-af5a-796fbdb9b5fe tempest-ImagesNegativeTestJSON-881899573 tempest-ImagesNegativeTestJSON-881899573-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 638.787152] env[62066]: DEBUG nova.virt.hardware [None req-2c107b45-3cbf-4be9-af5a-796fbdb9b5fe tempest-ImagesNegativeTestJSON-881899573 tempest-ImagesNegativeTestJSON-881899573-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 638.787403] env[62066]: DEBUG nova.virt.hardware [None req-2c107b45-3cbf-4be9-af5a-796fbdb9b5fe tempest-ImagesNegativeTestJSON-881899573 tempest-ImagesNegativeTestJSON-881899573-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 638.787599] env[62066]: DEBUG nova.virt.hardware [None req-2c107b45-3cbf-4be9-af5a-796fbdb9b5fe tempest-ImagesNegativeTestJSON-881899573 tempest-ImagesNegativeTestJSON-881899573-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 638.790951] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd397a94-c149-4934-9372-4d7428be3d79 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.798733] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Instance 2cf18a04-6bf9-4f47-a920-2c568207057d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 638.809511] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb8e256e-9cd4-4376-8292-8bf681fb79fb {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.923380] env[62066]: DEBUG nova.compute.manager [req-d14a01e0-55dd-4ea2-9281-396f2d37dab0 req-df52f844-7ef8-4179-9f50-fdfcedc3db33 service nova] [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] Received event network-changed-aa9ed432-13ef-49d1-94be-5f4666d101ef {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 638.923644] env[62066]: DEBUG nova.compute.manager [req-d14a01e0-55dd-4ea2-9281-396f2d37dab0 req-df52f844-7ef8-4179-9f50-fdfcedc3db33 service nova] [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] Refreshing instance network info cache due to event network-changed-aa9ed432-13ef-49d1-94be-5f4666d101ef. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 638.923908] env[62066]: DEBUG oslo_concurrency.lockutils [req-d14a01e0-55dd-4ea2-9281-396f2d37dab0 req-df52f844-7ef8-4179-9f50-fdfcedc3db33 service nova] Acquiring lock "refresh_cache-cb319720-29ab-4ff2-a71e-f77ae8c85735" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 638.924141] env[62066]: DEBUG oslo_concurrency.lockutils [req-d14a01e0-55dd-4ea2-9281-396f2d37dab0 req-df52f844-7ef8-4179-9f50-fdfcedc3db33 service nova] Acquired lock "refresh_cache-cb319720-29ab-4ff2-a71e-f77ae8c85735" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 638.924326] env[62066]: DEBUG nova.network.neutron [req-d14a01e0-55dd-4ea2-9281-396f2d37dab0 req-df52f844-7ef8-4179-9f50-fdfcedc3db33 service nova] [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] Refreshing network info cache for port aa9ed432-13ef-49d1-94be-5f4666d101ef {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 638.956462] env[62066]: DEBUG oslo_vmware.api [None req-5e085536-daab-46ff-8eb8-69c1da687135 tempest-ServersAaction247Test-1482016775 tempest-ServersAaction247Test-1482016775-project-member] Task: {'id': task-1340663, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 639.197019] env[62066]: ERROR nova.compute.manager [None req-2c107b45-3cbf-4be9-af5a-796fbdb9b5fe tempest-ImagesNegativeTestJSON-881899573 tempest-ImagesNegativeTestJSON-881899573-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port aa9ed432-13ef-49d1-94be-5f4666d101ef, please check neutron logs for more information. [ 639.197019] env[62066]: ERROR nova.compute.manager Traceback (most recent call last): [ 639.197019] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 639.197019] env[62066]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 639.197019] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 639.197019] env[62066]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 639.197019] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 639.197019] env[62066]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 639.197019] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 639.197019] env[62066]: ERROR nova.compute.manager self.force_reraise() [ 639.197019] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 639.197019] env[62066]: ERROR nova.compute.manager raise self.value [ 639.197019] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 639.197019] env[62066]: ERROR nova.compute.manager updated_port = self._update_port( [ 639.197019] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 639.197019] env[62066]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 639.198143] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 639.198143] env[62066]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 639.198143] env[62066]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port aa9ed432-13ef-49d1-94be-5f4666d101ef, please check neutron logs for more information. [ 639.198143] env[62066]: ERROR nova.compute.manager [ 639.198143] env[62066]: Traceback (most recent call last): [ 639.198143] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 639.198143] env[62066]: listener.cb(fileno) [ 639.198143] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 639.198143] env[62066]: result = function(*args, **kwargs) [ 639.198143] env[62066]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 639.198143] env[62066]: return func(*args, **kwargs) [ 639.198143] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 639.198143] env[62066]: raise e [ 639.198143] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 639.198143] env[62066]: nwinfo = self.network_api.allocate_for_instance( [ 639.198143] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 639.198143] env[62066]: created_port_ids = self._update_ports_for_instance( [ 639.198143] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 639.198143] env[62066]: with excutils.save_and_reraise_exception(): [ 639.198143] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 639.198143] env[62066]: self.force_reraise() [ 639.198143] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 639.198143] env[62066]: raise self.value [ 639.198143] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 639.198143] env[62066]: updated_port = self._update_port( [ 639.198143] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 639.198143] env[62066]: _ensure_no_port_binding_failure(port) [ 639.198143] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 639.198143] env[62066]: raise exception.PortBindingFailed(port_id=port['id']) [ 639.199298] env[62066]: nova.exception.PortBindingFailed: Binding failed for port aa9ed432-13ef-49d1-94be-5f4666d101ef, please check neutron logs for more information. [ 639.199298] env[62066]: Removing descriptor: 20 [ 639.199298] env[62066]: ERROR nova.compute.manager [None req-2c107b45-3cbf-4be9-af5a-796fbdb9b5fe tempest-ImagesNegativeTestJSON-881899573 tempest-ImagesNegativeTestJSON-881899573-project-member] [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port aa9ed432-13ef-49d1-94be-5f4666d101ef, please check neutron logs for more information. [ 639.199298] env[62066]: ERROR nova.compute.manager [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] Traceback (most recent call last): [ 639.199298] env[62066]: ERROR nova.compute.manager [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 639.199298] env[62066]: ERROR nova.compute.manager [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] yield resources [ 639.199298] env[62066]: ERROR nova.compute.manager [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 639.199298] env[62066]: ERROR nova.compute.manager [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] self.driver.spawn(context, instance, image_meta, [ 639.199298] env[62066]: ERROR nova.compute.manager [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 639.199298] env[62066]: ERROR nova.compute.manager [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] self._vmops.spawn(context, instance, image_meta, injected_files, [ 639.199298] env[62066]: ERROR nova.compute.manager [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 639.199298] env[62066]: ERROR nova.compute.manager [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] vm_ref = self.build_virtual_machine(instance, [ 639.199698] env[62066]: ERROR nova.compute.manager [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 639.199698] env[62066]: ERROR nova.compute.manager [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] vif_infos = vmwarevif.get_vif_info(self._session, [ 639.199698] env[62066]: ERROR nova.compute.manager [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 639.199698] env[62066]: ERROR nova.compute.manager [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] for vif in network_info: [ 639.199698] env[62066]: ERROR nova.compute.manager [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 639.199698] env[62066]: ERROR nova.compute.manager [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] return self._sync_wrapper(fn, *args, **kwargs) [ 639.199698] env[62066]: ERROR nova.compute.manager [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 639.199698] env[62066]: ERROR nova.compute.manager [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] self.wait() [ 639.199698] env[62066]: ERROR nova.compute.manager [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 639.199698] env[62066]: ERROR nova.compute.manager [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] self[:] = self._gt.wait() [ 639.199698] env[62066]: ERROR nova.compute.manager [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 639.199698] env[62066]: ERROR nova.compute.manager [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] return self._exit_event.wait() [ 639.199698] env[62066]: ERROR nova.compute.manager [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 639.200034] env[62066]: ERROR nova.compute.manager [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] result = hub.switch() [ 639.200034] env[62066]: ERROR nova.compute.manager [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 639.200034] env[62066]: ERROR nova.compute.manager [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] return self.greenlet.switch() [ 639.200034] env[62066]: ERROR nova.compute.manager [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 639.200034] env[62066]: ERROR nova.compute.manager [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] result = function(*args, **kwargs) [ 639.200034] env[62066]: ERROR nova.compute.manager [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 639.200034] env[62066]: ERROR nova.compute.manager [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] return func(*args, **kwargs) [ 639.200034] env[62066]: ERROR nova.compute.manager [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 639.200034] env[62066]: ERROR nova.compute.manager [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] raise e [ 639.200034] env[62066]: ERROR nova.compute.manager [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 639.200034] env[62066]: ERROR nova.compute.manager [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] nwinfo = self.network_api.allocate_for_instance( [ 639.200034] env[62066]: ERROR nova.compute.manager [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 639.200034] env[62066]: ERROR nova.compute.manager [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] created_port_ids = self._update_ports_for_instance( [ 639.200387] env[62066]: ERROR nova.compute.manager [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 639.200387] env[62066]: ERROR nova.compute.manager [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] with excutils.save_and_reraise_exception(): [ 639.200387] env[62066]: ERROR nova.compute.manager [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 639.200387] env[62066]: ERROR nova.compute.manager [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] self.force_reraise() [ 639.200387] env[62066]: ERROR nova.compute.manager [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 639.200387] env[62066]: ERROR nova.compute.manager [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] raise self.value [ 639.200387] env[62066]: ERROR nova.compute.manager [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 639.200387] env[62066]: ERROR nova.compute.manager [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] updated_port = self._update_port( [ 639.200387] env[62066]: ERROR nova.compute.manager [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 639.200387] env[62066]: ERROR nova.compute.manager [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] _ensure_no_port_binding_failure(port) [ 639.200387] env[62066]: ERROR nova.compute.manager [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 639.200387] env[62066]: ERROR nova.compute.manager [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] raise exception.PortBindingFailed(port_id=port['id']) [ 639.201426] env[62066]: ERROR nova.compute.manager [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] nova.exception.PortBindingFailed: Binding failed for port aa9ed432-13ef-49d1-94be-5f4666d101ef, please check neutron logs for more information. [ 639.201426] env[62066]: ERROR nova.compute.manager [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] [ 639.201426] env[62066]: INFO nova.compute.manager [None req-2c107b45-3cbf-4be9-af5a-796fbdb9b5fe tempest-ImagesNegativeTestJSON-881899573 tempest-ImagesNegativeTestJSON-881899573-project-member] [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] Terminating instance [ 639.201426] env[62066]: DEBUG oslo_concurrency.lockutils [None req-2c107b45-3cbf-4be9-af5a-796fbdb9b5fe tempest-ImagesNegativeTestJSON-881899573 tempest-ImagesNegativeTestJSON-881899573-project-member] Acquiring lock "refresh_cache-cb319720-29ab-4ff2-a71e-f77ae8c85735" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 639.302490] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Instance 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 639.443691] env[62066]: DEBUG nova.network.neutron [req-d14a01e0-55dd-4ea2-9281-396f2d37dab0 req-df52f844-7ef8-4179-9f50-fdfcedc3db33 service nova] [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 639.457794] env[62066]: DEBUG oslo_vmware.api [None req-5e085536-daab-46ff-8eb8-69c1da687135 tempest-ServersAaction247Test-1482016775 tempest-ServersAaction247Test-1482016775-project-member] Task: {'id': task-1340663, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.69049} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 639.458139] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-5e085536-daab-46ff-8eb8-69c1da687135 tempest-ServersAaction247Test-1482016775 tempest-ServersAaction247Test-1482016775-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk to [datastore2] 5d54c3fd-b239-4965-b7a3-5909e8de8bc0/5d54c3fd-b239-4965-b7a3-5909e8de8bc0.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 639.458386] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-5e085536-daab-46ff-8eb8-69c1da687135 tempest-ServersAaction247Test-1482016775 tempest-ServersAaction247Test-1482016775-project-member] [instance: 5d54c3fd-b239-4965-b7a3-5909e8de8bc0] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 639.458673] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e188418b-b0e9-45a4-bdba-be1d0dde1b72 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.465703] env[62066]: DEBUG oslo_vmware.api [None req-5e085536-daab-46ff-8eb8-69c1da687135 tempest-ServersAaction247Test-1482016775 tempest-ServersAaction247Test-1482016775-project-member] Waiting for the task: (returnval){ [ 639.465703] env[62066]: value = "task-1340665" [ 639.465703] env[62066]: _type = "Task" [ 639.465703] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 639.474138] env[62066]: DEBUG oslo_vmware.api [None req-5e085536-daab-46ff-8eb8-69c1da687135 tempest-ServersAaction247Test-1482016775 tempest-ServersAaction247Test-1482016775-project-member] Task: {'id': task-1340665, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 639.551046] env[62066]: DEBUG nova.network.neutron [req-d14a01e0-55dd-4ea2-9281-396f2d37dab0 req-df52f844-7ef8-4179-9f50-fdfcedc3db33 service nova] [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 639.805874] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Instance e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 639.977600] env[62066]: DEBUG oslo_vmware.api [None req-5e085536-daab-46ff-8eb8-69c1da687135 tempest-ServersAaction247Test-1482016775 tempest-ServersAaction247Test-1482016775-project-member] Task: {'id': task-1340665, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066156} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 639.977912] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-5e085536-daab-46ff-8eb8-69c1da687135 tempest-ServersAaction247Test-1482016775 tempest-ServersAaction247Test-1482016775-project-member] [instance: 5d54c3fd-b239-4965-b7a3-5909e8de8bc0] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 639.978746] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9768aeb8-6241-415e-8cb3-f1a173d9a27e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.998569] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-5e085536-daab-46ff-8eb8-69c1da687135 tempest-ServersAaction247Test-1482016775 tempest-ServersAaction247Test-1482016775-project-member] [instance: 5d54c3fd-b239-4965-b7a3-5909e8de8bc0] Reconfiguring VM instance instance-0000001f to attach disk [datastore2] 5d54c3fd-b239-4965-b7a3-5909e8de8bc0/5d54c3fd-b239-4965-b7a3-5909e8de8bc0.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 639.998812] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-22109a0c-546a-4adf-812a-fc0322647452 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.018338] env[62066]: DEBUG oslo_vmware.api [None req-5e085536-daab-46ff-8eb8-69c1da687135 tempest-ServersAaction247Test-1482016775 tempest-ServersAaction247Test-1482016775-project-member] Waiting for the task: (returnval){ [ 640.018338] env[62066]: value = "task-1340666" [ 640.018338] env[62066]: _type = "Task" [ 640.018338] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 640.027549] env[62066]: DEBUG oslo_vmware.api [None req-5e085536-daab-46ff-8eb8-69c1da687135 tempest-ServersAaction247Test-1482016775 tempest-ServersAaction247Test-1482016775-project-member] Task: {'id': task-1340666, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 640.057413] env[62066]: DEBUG oslo_concurrency.lockutils [req-d14a01e0-55dd-4ea2-9281-396f2d37dab0 req-df52f844-7ef8-4179-9f50-fdfcedc3db33 service nova] Releasing lock "refresh_cache-cb319720-29ab-4ff2-a71e-f77ae8c85735" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 640.057857] env[62066]: DEBUG oslo_concurrency.lockutils [None req-2c107b45-3cbf-4be9-af5a-796fbdb9b5fe tempest-ImagesNegativeTestJSON-881899573 tempest-ImagesNegativeTestJSON-881899573-project-member] Acquired lock "refresh_cache-cb319720-29ab-4ff2-a71e-f77ae8c85735" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 640.058130] env[62066]: DEBUG nova.network.neutron [None req-2c107b45-3cbf-4be9-af5a-796fbdb9b5fe tempest-ImagesNegativeTestJSON-881899573 tempest-ImagesNegativeTestJSON-881899573-project-member] [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 640.309903] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Instance e795e53e-af33-4d0c-8dbd-9c59f236acd7 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 640.528692] env[62066]: DEBUG oslo_vmware.api [None req-5e085536-daab-46ff-8eb8-69c1da687135 tempest-ServersAaction247Test-1482016775 tempest-ServersAaction247Test-1482016775-project-member] Task: {'id': task-1340666, 'name': ReconfigVM_Task, 'duration_secs': 0.271394} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 640.528990] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-5e085536-daab-46ff-8eb8-69c1da687135 tempest-ServersAaction247Test-1482016775 tempest-ServersAaction247Test-1482016775-project-member] [instance: 5d54c3fd-b239-4965-b7a3-5909e8de8bc0] Reconfigured VM instance instance-0000001f to attach disk [datastore2] 5d54c3fd-b239-4965-b7a3-5909e8de8bc0/5d54c3fd-b239-4965-b7a3-5909e8de8bc0.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 640.529615] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6c7e3ff8-0d49-4ead-bf89-4aba7c655488 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.535351] env[62066]: DEBUG oslo_vmware.api [None req-5e085536-daab-46ff-8eb8-69c1da687135 tempest-ServersAaction247Test-1482016775 tempest-ServersAaction247Test-1482016775-project-member] Waiting for the task: (returnval){ [ 640.535351] env[62066]: value = "task-1340667" [ 640.535351] env[62066]: _type = "Task" [ 640.535351] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 640.542483] env[62066]: DEBUG oslo_vmware.api [None req-5e085536-daab-46ff-8eb8-69c1da687135 tempest-ServersAaction247Test-1482016775 tempest-ServersAaction247Test-1482016775-project-member] Task: {'id': task-1340667, 'name': Rename_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 640.580741] env[62066]: DEBUG nova.network.neutron [None req-2c107b45-3cbf-4be9-af5a-796fbdb9b5fe tempest-ImagesNegativeTestJSON-881899573 tempest-ImagesNegativeTestJSON-881899573-project-member] [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 640.705259] env[62066]: DEBUG nova.network.neutron [None req-2c107b45-3cbf-4be9-af5a-796fbdb9b5fe tempest-ImagesNegativeTestJSON-881899573 tempest-ImagesNegativeTestJSON-881899573-project-member] [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 640.815459] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Instance a690ad15-2878-4fa7-a810-d155ce6b1a8c has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 640.955837] env[62066]: DEBUG nova.compute.manager [req-94762d9e-9745-48f2-90ab-58acccc0271b req-1648931f-bac0-47ea-9699-e5e946921ad3 service nova] [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] Received event network-vif-deleted-aa9ed432-13ef-49d1-94be-5f4666d101ef {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 641.044748] env[62066]: DEBUG oslo_vmware.api [None req-5e085536-daab-46ff-8eb8-69c1da687135 tempest-ServersAaction247Test-1482016775 tempest-ServersAaction247Test-1482016775-project-member] Task: {'id': task-1340667, 'name': Rename_Task, 'duration_secs': 0.133425} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 641.045392] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-5e085536-daab-46ff-8eb8-69c1da687135 tempest-ServersAaction247Test-1482016775 tempest-ServersAaction247Test-1482016775-project-member] [instance: 5d54c3fd-b239-4965-b7a3-5909e8de8bc0] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 641.045585] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d676c5f9-76b5-4b39-bcdb-0bdc921dd236 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.051031] env[62066]: DEBUG oslo_vmware.api [None req-5e085536-daab-46ff-8eb8-69c1da687135 tempest-ServersAaction247Test-1482016775 tempest-ServersAaction247Test-1482016775-project-member] Waiting for the task: (returnval){ [ 641.051031] env[62066]: value = "task-1340669" [ 641.051031] env[62066]: _type = "Task" [ 641.051031] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 641.058079] env[62066]: DEBUG oslo_vmware.api [None req-5e085536-daab-46ff-8eb8-69c1da687135 tempest-ServersAaction247Test-1482016775 tempest-ServersAaction247Test-1482016775-project-member] Task: {'id': task-1340669, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 641.211949] env[62066]: DEBUG oslo_concurrency.lockutils [None req-2c107b45-3cbf-4be9-af5a-796fbdb9b5fe tempest-ImagesNegativeTestJSON-881899573 tempest-ImagesNegativeTestJSON-881899573-project-member] Releasing lock "refresh_cache-cb319720-29ab-4ff2-a71e-f77ae8c85735" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 641.212389] env[62066]: DEBUG nova.compute.manager [None req-2c107b45-3cbf-4be9-af5a-796fbdb9b5fe tempest-ImagesNegativeTestJSON-881899573 tempest-ImagesNegativeTestJSON-881899573-project-member] [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 641.212523] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-2c107b45-3cbf-4be9-af5a-796fbdb9b5fe tempest-ImagesNegativeTestJSON-881899573 tempest-ImagesNegativeTestJSON-881899573-project-member] [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 641.212836] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3815a833-de41-46d2-92c3-4c8a1efa9d00 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.222276] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b80b856-07bc-4645-b3a7-9947f0b384cf {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.245035] env[62066]: WARNING nova.virt.vmwareapi.vmops [None req-2c107b45-3cbf-4be9-af5a-796fbdb9b5fe tempest-ImagesNegativeTestJSON-881899573 tempest-ImagesNegativeTestJSON-881899573-project-member] [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance cb319720-29ab-4ff2-a71e-f77ae8c85735 could not be found. [ 641.245322] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-2c107b45-3cbf-4be9-af5a-796fbdb9b5fe tempest-ImagesNegativeTestJSON-881899573 tempest-ImagesNegativeTestJSON-881899573-project-member] [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 641.245545] env[62066]: INFO nova.compute.manager [None req-2c107b45-3cbf-4be9-af5a-796fbdb9b5fe tempest-ImagesNegativeTestJSON-881899573 tempest-ImagesNegativeTestJSON-881899573-project-member] [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] Took 0.03 seconds to destroy the instance on the hypervisor. [ 641.245851] env[62066]: DEBUG oslo.service.loopingcall [None req-2c107b45-3cbf-4be9-af5a-796fbdb9b5fe tempest-ImagesNegativeTestJSON-881899573 tempest-ImagesNegativeTestJSON-881899573-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 641.246134] env[62066]: DEBUG nova.compute.manager [-] [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 641.246249] env[62066]: DEBUG nova.network.neutron [-] [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 641.261118] env[62066]: DEBUG nova.network.neutron [-] [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 641.318703] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Instance 5fad2258-f694-4993-bf64-bfa16abc09ba has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 641.564533] env[62066]: DEBUG oslo_vmware.api [None req-5e085536-daab-46ff-8eb8-69c1da687135 tempest-ServersAaction247Test-1482016775 tempest-ServersAaction247Test-1482016775-project-member] Task: {'id': task-1340669, 'name': PowerOnVM_Task, 'duration_secs': 0.417902} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 641.564533] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-5e085536-daab-46ff-8eb8-69c1da687135 tempest-ServersAaction247Test-1482016775 tempest-ServersAaction247Test-1482016775-project-member] [instance: 5d54c3fd-b239-4965-b7a3-5909e8de8bc0] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 641.564810] env[62066]: INFO nova.compute.manager [None req-5e085536-daab-46ff-8eb8-69c1da687135 tempest-ServersAaction247Test-1482016775 tempest-ServersAaction247Test-1482016775-project-member] [instance: 5d54c3fd-b239-4965-b7a3-5909e8de8bc0] Took 5.30 seconds to spawn the instance on the hypervisor. [ 641.565085] env[62066]: DEBUG nova.compute.manager [None req-5e085536-daab-46ff-8eb8-69c1da687135 tempest-ServersAaction247Test-1482016775 tempest-ServersAaction247Test-1482016775-project-member] [instance: 5d54c3fd-b239-4965-b7a3-5909e8de8bc0] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 641.566219] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0c66b38-8188-4800-97f1-7c76399c9f12 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.763913] env[62066]: DEBUG nova.network.neutron [-] [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 641.822079] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Instance 45c32181-01a8-4e1e-9e0e-37035a298d55 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 642.087163] env[62066]: INFO nova.compute.manager [None req-5e085536-daab-46ff-8eb8-69c1da687135 tempest-ServersAaction247Test-1482016775 tempest-ServersAaction247Test-1482016775-project-member] [instance: 5d54c3fd-b239-4965-b7a3-5909e8de8bc0] Took 24.63 seconds to build instance. [ 642.267728] env[62066]: INFO nova.compute.manager [-] [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] Took 1.02 seconds to deallocate network for instance. [ 642.270062] env[62066]: DEBUG nova.compute.claims [None req-2c107b45-3cbf-4be9-af5a-796fbdb9b5fe tempest-ImagesNegativeTestJSON-881899573 tempest-ImagesNegativeTestJSON-881899573-project-member] [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] Aborting claim: {{(pid=62066) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 642.270249] env[62066]: DEBUG oslo_concurrency.lockutils [None req-2c107b45-3cbf-4be9-af5a-796fbdb9b5fe tempest-ImagesNegativeTestJSON-881899573 tempest-ImagesNegativeTestJSON-881899573-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 642.324223] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Instance f9a8e863-6e7f-4f00-b54a-78802659cd59 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 642.588610] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5e085536-daab-46ff-8eb8-69c1da687135 tempest-ServersAaction247Test-1482016775 tempest-ServersAaction247Test-1482016775-project-member] Lock "5d54c3fd-b239-4965-b7a3-5909e8de8bc0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 107.272s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 642.828294] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Instance 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 642.845125] env[62066]: DEBUG nova.compute.manager [None req-6d4ff083-a88e-4da1-b01f-617fef635dac tempest-ServersAaction247Test-1482016775 tempest-ServersAaction247Test-1482016775-project-member] [instance: 5d54c3fd-b239-4965-b7a3-5909e8de8bc0] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 642.846419] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d08db466-3815-4a87-a2e7-0bf6e28c2c4a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.949246] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b5125342-7692-4177-9e88-9f6d8ef9f1dc tempest-ServersAaction247Test-1482016775 tempest-ServersAaction247Test-1482016775-project-member] Acquiring lock "5d54c3fd-b239-4965-b7a3-5909e8de8bc0" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 642.949521] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b5125342-7692-4177-9e88-9f6d8ef9f1dc tempest-ServersAaction247Test-1482016775 tempest-ServersAaction247Test-1482016775-project-member] Lock "5d54c3fd-b239-4965-b7a3-5909e8de8bc0" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 642.949727] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b5125342-7692-4177-9e88-9f6d8ef9f1dc tempest-ServersAaction247Test-1482016775 tempest-ServersAaction247Test-1482016775-project-member] Acquiring lock "5d54c3fd-b239-4965-b7a3-5909e8de8bc0-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 642.949912] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b5125342-7692-4177-9e88-9f6d8ef9f1dc tempest-ServersAaction247Test-1482016775 tempest-ServersAaction247Test-1482016775-project-member] Lock "5d54c3fd-b239-4965-b7a3-5909e8de8bc0-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 642.950099] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b5125342-7692-4177-9e88-9f6d8ef9f1dc tempest-ServersAaction247Test-1482016775 tempest-ServersAaction247Test-1482016775-project-member] Lock "5d54c3fd-b239-4965-b7a3-5909e8de8bc0-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 642.952297] env[62066]: INFO nova.compute.manager [None req-b5125342-7692-4177-9e88-9f6d8ef9f1dc tempest-ServersAaction247Test-1482016775 tempest-ServersAaction247Test-1482016775-project-member] [instance: 5d54c3fd-b239-4965-b7a3-5909e8de8bc0] Terminating instance [ 642.953915] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b5125342-7692-4177-9e88-9f6d8ef9f1dc tempest-ServersAaction247Test-1482016775 tempest-ServersAaction247Test-1482016775-project-member] Acquiring lock "refresh_cache-5d54c3fd-b239-4965-b7a3-5909e8de8bc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 642.954092] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b5125342-7692-4177-9e88-9f6d8ef9f1dc tempest-ServersAaction247Test-1482016775 tempest-ServersAaction247Test-1482016775-project-member] Acquired lock "refresh_cache-5d54c3fd-b239-4965-b7a3-5909e8de8bc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 642.954259] env[62066]: DEBUG nova.network.neutron [None req-b5125342-7692-4177-9e88-9f6d8ef9f1dc tempest-ServersAaction247Test-1482016775 tempest-ServersAaction247Test-1482016775-project-member] [instance: 5d54c3fd-b239-4965-b7a3-5909e8de8bc0] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 643.095430] env[62066]: DEBUG nova.compute.manager [None req-b96d7111-6e60-4fd2-a38a-3863242412fb tempest-ServerActionsTestOtherB-426172427 tempest-ServerActionsTestOtherB-426172427-project-member] [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 643.332066] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Instance f91a90f7-be73-424c-966c-1be6f37a0864 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 643.360718] env[62066]: INFO nova.compute.manager [None req-6d4ff083-a88e-4da1-b01f-617fef635dac tempest-ServersAaction247Test-1482016775 tempest-ServersAaction247Test-1482016775-project-member] [instance: 5d54c3fd-b239-4965-b7a3-5909e8de8bc0] instance snapshotting [ 643.361336] env[62066]: DEBUG nova.objects.instance [None req-6d4ff083-a88e-4da1-b01f-617fef635dac tempest-ServersAaction247Test-1482016775 tempest-ServersAaction247Test-1482016775-project-member] Lazy-loading 'flavor' on Instance uuid 5d54c3fd-b239-4965-b7a3-5909e8de8bc0 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 643.472259] env[62066]: DEBUG nova.network.neutron [None req-b5125342-7692-4177-9e88-9f6d8ef9f1dc tempest-ServersAaction247Test-1482016775 tempest-ServersAaction247Test-1482016775-project-member] [instance: 5d54c3fd-b239-4965-b7a3-5909e8de8bc0] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 643.531051] env[62066]: DEBUG nova.network.neutron [None req-b5125342-7692-4177-9e88-9f6d8ef9f1dc tempest-ServersAaction247Test-1482016775 tempest-ServersAaction247Test-1482016775-project-member] [instance: 5d54c3fd-b239-4965-b7a3-5909e8de8bc0] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 643.620801] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b96d7111-6e60-4fd2-a38a-3863242412fb tempest-ServerActionsTestOtherB-426172427 tempest-ServerActionsTestOtherB-426172427-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 643.834441] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Instance 0ff2f9f4-6d85-4465-8e0a-fb39f2a25c9b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 643.870426] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1cb8709d-13f4-4afe-a268-85cd2f218666 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.888597] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70895d33-d964-4611-aa17-b4bf0b21659d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.034305] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b5125342-7692-4177-9e88-9f6d8ef9f1dc tempest-ServersAaction247Test-1482016775 tempest-ServersAaction247Test-1482016775-project-member] Releasing lock "refresh_cache-5d54c3fd-b239-4965-b7a3-5909e8de8bc0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 644.034777] env[62066]: DEBUG nova.compute.manager [None req-b5125342-7692-4177-9e88-9f6d8ef9f1dc tempest-ServersAaction247Test-1482016775 tempest-ServersAaction247Test-1482016775-project-member] [instance: 5d54c3fd-b239-4965-b7a3-5909e8de8bc0] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 644.035388] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-b5125342-7692-4177-9e88-9f6d8ef9f1dc tempest-ServersAaction247Test-1482016775 tempest-ServersAaction247Test-1482016775-project-member] [instance: 5d54c3fd-b239-4965-b7a3-5909e8de8bc0] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 644.036030] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31495075-27f7-4295-b886-4cd5ae400726 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.044021] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-b5125342-7692-4177-9e88-9f6d8ef9f1dc tempest-ServersAaction247Test-1482016775 tempest-ServersAaction247Test-1482016775-project-member] [instance: 5d54c3fd-b239-4965-b7a3-5909e8de8bc0] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 644.044021] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d84bb8eb-89db-4783-b6bf-fa58287ae8d2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.050472] env[62066]: DEBUG oslo_vmware.api [None req-b5125342-7692-4177-9e88-9f6d8ef9f1dc tempest-ServersAaction247Test-1482016775 tempest-ServersAaction247Test-1482016775-project-member] Waiting for the task: (returnval){ [ 644.050472] env[62066]: value = "task-1340671" [ 644.050472] env[62066]: _type = "Task" [ 644.050472] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 644.058605] env[62066]: DEBUG oslo_vmware.api [None req-b5125342-7692-4177-9e88-9f6d8ef9f1dc tempest-ServersAaction247Test-1482016775 tempest-ServersAaction247Test-1482016775-project-member] Task: {'id': task-1340671, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 644.338064] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Instance d7de90c0-95e9-4899-92a0-81658831713a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 644.398698] env[62066]: DEBUG nova.compute.manager [None req-6d4ff083-a88e-4da1-b01f-617fef635dac tempest-ServersAaction247Test-1482016775 tempest-ServersAaction247Test-1482016775-project-member] [instance: 5d54c3fd-b239-4965-b7a3-5909e8de8bc0] Instance disappeared during snapshot {{(pid=62066) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4494}} [ 644.535463] env[62066]: DEBUG nova.compute.manager [None req-6d4ff083-a88e-4da1-b01f-617fef635dac tempest-ServersAaction247Test-1482016775 tempest-ServersAaction247Test-1482016775-project-member] [instance: 5d54c3fd-b239-4965-b7a3-5909e8de8bc0] Found 0 images (rotation: 2) {{(pid=62066) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4554}} [ 644.545846] env[62066]: DEBUG oslo_concurrency.lockutils [None req-064a3fac-0a45-4f27-8eed-cb07d9679634 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Acquiring lock "1a4b0637-1a56-41ef-b89b-6b56d24ed206" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 644.546113] env[62066]: DEBUG oslo_concurrency.lockutils [None req-064a3fac-0a45-4f27-8eed-cb07d9679634 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Lock "1a4b0637-1a56-41ef-b89b-6b56d24ed206" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 644.560649] env[62066]: DEBUG oslo_vmware.api [None req-b5125342-7692-4177-9e88-9f6d8ef9f1dc tempest-ServersAaction247Test-1482016775 tempest-ServersAaction247Test-1482016775-project-member] Task: {'id': task-1340671, 'name': PowerOffVM_Task, 'duration_secs': 0.187218} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 644.560914] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-b5125342-7692-4177-9e88-9f6d8ef9f1dc tempest-ServersAaction247Test-1482016775 tempest-ServersAaction247Test-1482016775-project-member] [instance: 5d54c3fd-b239-4965-b7a3-5909e8de8bc0] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 644.561101] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-b5125342-7692-4177-9e88-9f6d8ef9f1dc tempest-ServersAaction247Test-1482016775 tempest-ServersAaction247Test-1482016775-project-member] [instance: 5d54c3fd-b239-4965-b7a3-5909e8de8bc0] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 644.561400] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f2b7b18a-ed8f-43da-8065-f974e3ceaa0b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.587708] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-b5125342-7692-4177-9e88-9f6d8ef9f1dc tempest-ServersAaction247Test-1482016775 tempest-ServersAaction247Test-1482016775-project-member] [instance: 5d54c3fd-b239-4965-b7a3-5909e8de8bc0] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 644.587963] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-b5125342-7692-4177-9e88-9f6d8ef9f1dc tempest-ServersAaction247Test-1482016775 tempest-ServersAaction247Test-1482016775-project-member] [instance: 5d54c3fd-b239-4965-b7a3-5909e8de8bc0] Deleting contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 644.588173] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-b5125342-7692-4177-9e88-9f6d8ef9f1dc tempest-ServersAaction247Test-1482016775 tempest-ServersAaction247Test-1482016775-project-member] Deleting the datastore file [datastore2] 5d54c3fd-b239-4965-b7a3-5909e8de8bc0 {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 644.588432] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7fe6c2c1-5a73-496c-9861-d8a18866594e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.595973] env[62066]: DEBUG oslo_vmware.api [None req-b5125342-7692-4177-9e88-9f6d8ef9f1dc tempest-ServersAaction247Test-1482016775 tempest-ServersAaction247Test-1482016775-project-member] Waiting for the task: (returnval){ [ 644.595973] env[62066]: value = "task-1340673" [ 644.595973] env[62066]: _type = "Task" [ 644.595973] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 644.603696] env[62066]: DEBUG oslo_vmware.api [None req-b5125342-7692-4177-9e88-9f6d8ef9f1dc tempest-ServersAaction247Test-1482016775 tempest-ServersAaction247Test-1482016775-project-member] Task: {'id': task-1340673, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 644.763127] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3baf3239-22c0-4858-82f6-d3b75bd59a7c tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Acquiring lock "1e1a2ede-4ad8-4600-851e-6f2046b3f919" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 644.763369] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3baf3239-22c0-4858-82f6-d3b75bd59a7c tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Lock "1e1a2ede-4ad8-4600-851e-6f2046b3f919" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 644.840899] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Instance 3624ca43-9ee3-418a-9205-debb699295df has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 645.106201] env[62066]: DEBUG oslo_vmware.api [None req-b5125342-7692-4177-9e88-9f6d8ef9f1dc tempest-ServersAaction247Test-1482016775 tempest-ServersAaction247Test-1482016775-project-member] Task: {'id': task-1340673, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.090921} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 645.107212] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-b5125342-7692-4177-9e88-9f6d8ef9f1dc tempest-ServersAaction247Test-1482016775 tempest-ServersAaction247Test-1482016775-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 645.107212] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-b5125342-7692-4177-9e88-9f6d8ef9f1dc tempest-ServersAaction247Test-1482016775 tempest-ServersAaction247Test-1482016775-project-member] [instance: 5d54c3fd-b239-4965-b7a3-5909e8de8bc0] Deleted contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 645.107212] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-b5125342-7692-4177-9e88-9f6d8ef9f1dc tempest-ServersAaction247Test-1482016775 tempest-ServersAaction247Test-1482016775-project-member] [instance: 5d54c3fd-b239-4965-b7a3-5909e8de8bc0] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 645.107845] env[62066]: INFO nova.compute.manager [None req-b5125342-7692-4177-9e88-9f6d8ef9f1dc tempest-ServersAaction247Test-1482016775 tempest-ServersAaction247Test-1482016775-project-member] [instance: 5d54c3fd-b239-4965-b7a3-5909e8de8bc0] Took 1.07 seconds to destroy the instance on the hypervisor. [ 645.107845] env[62066]: DEBUG oslo.service.loopingcall [None req-b5125342-7692-4177-9e88-9f6d8ef9f1dc tempest-ServersAaction247Test-1482016775 tempest-ServersAaction247Test-1482016775-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 645.108130] env[62066]: DEBUG nova.compute.manager [-] [instance: 5d54c3fd-b239-4965-b7a3-5909e8de8bc0] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 645.108130] env[62066]: DEBUG nova.network.neutron [-] [instance: 5d54c3fd-b239-4965-b7a3-5909e8de8bc0] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 645.123816] env[62066]: DEBUG nova.network.neutron [-] [instance: 5d54c3fd-b239-4965-b7a3-5909e8de8bc0] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 645.345049] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Instance c908a3ed-88a6-4e79-8c27-2e16bf2bdca4 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 645.626533] env[62066]: DEBUG nova.network.neutron [-] [instance: 5d54c3fd-b239-4965-b7a3-5909e8de8bc0] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 645.848159] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Instance 0ed07533-9e79-416d-a495-a8e20b97e108 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 646.132034] env[62066]: INFO nova.compute.manager [-] [instance: 5d54c3fd-b239-4965-b7a3-5909e8de8bc0] Took 1.02 seconds to deallocate network for instance. [ 646.351632] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Instance 2b81d060-ecb7-4ae5-b51d-33197937a76f has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 646.637966] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b5125342-7692-4177-9e88-9f6d8ef9f1dc tempest-ServersAaction247Test-1482016775 tempest-ServersAaction247Test-1482016775-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 646.854124] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Instance 04d25af4-0e73-4650-9c3d-85817754bac9 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 647.356855] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Instance d973b09f-7727-4055-b05e-123079df9ce4 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 647.860354] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Instance 71c44a07-fdf0-4847-868a-47f16c033c3d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 648.036683] env[62066]: DEBUG oslo_concurrency.lockutils [None req-86820846-d757-4de7-918c-ad5501b560db tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] Acquiring lock "daffaf51-4c45-44aa-8fc2-4db066a09971" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 648.036938] env[62066]: DEBUG oslo_concurrency.lockutils [None req-86820846-d757-4de7-918c-ad5501b560db tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] Lock "daffaf51-4c45-44aa-8fc2-4db066a09971" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 648.362994] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Instance f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 648.866469] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Instance 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 649.369773] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Instance 0ebba3b2-f82d-4c1d-b01a-f75c3559c117 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 649.872269] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Instance 98fb270c-f3f8-4375-8b5c-c2279305d476 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 649.872538] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Total usable vcpus: 48, total allocated vcpus: 6 {{(pid=62066) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 649.872684] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1664MB phys_disk=200GB used_disk=6GB total_vcpus=48 used_vcpus=6 pci_stats=[] {{(pid=62066) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 650.268723] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27b38d85-6cc3-489b-a20c-11b06d1193b3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.276409] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad663131-347a-400f-8532-3246a34fbe79 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.305240] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49d40ea5-835e-40b8-9605-7dcf509468f8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.312480] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fed5272-696d-4187-b555-2c57492e6b2f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.325222] env[62066]: DEBUG nova.compute.provider_tree [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 650.828823] env[62066]: DEBUG nova.scheduler.client.report [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 651.333841] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62066) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 651.334166] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 14.613s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 651.334418] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a9ebb4d2-ff8d-4333-a9df-86618ea2d54c tempest-ImagesOneServerTestJSON-1591819949 tempest-ImagesOneServerTestJSON-1591819949-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 29.641s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 652.227713] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d8c5ce6-6b4a-47a3-a061-21167f87bebf {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.235350] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe1dd3d3-acfd-4868-a731-26dbc6fbc07f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.264633] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3012b287-8695-4349-b9b5-9fe49bfa22d6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.271738] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbc63025-9aac-4ada-be74-054614b63efa {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.284468] env[62066]: DEBUG nova.compute.provider_tree [None req-a9ebb4d2-ff8d-4333-a9df-86618ea2d54c tempest-ImagesOneServerTestJSON-1591819949 tempest-ImagesOneServerTestJSON-1591819949-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 652.787740] env[62066]: DEBUG nova.scheduler.client.report [None req-a9ebb4d2-ff8d-4333-a9df-86618ea2d54c tempest-ImagesOneServerTestJSON-1591819949 tempest-ImagesOneServerTestJSON-1591819949-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 653.292954] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a9ebb4d2-ff8d-4333-a9df-86618ea2d54c tempest-ImagesOneServerTestJSON-1591819949 tempest-ImagesOneServerTestJSON-1591819949-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.958s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 653.293644] env[62066]: ERROR nova.compute.manager [None req-a9ebb4d2-ff8d-4333-a9df-86618ea2d54c tempest-ImagesOneServerTestJSON-1591819949 tempest-ImagesOneServerTestJSON-1591819949-project-member] [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port e09cd9e1-44fe-4858-bf9a-0afd7b681ca4, please check neutron logs for more information. [ 653.293644] env[62066]: ERROR nova.compute.manager [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] Traceback (most recent call last): [ 653.293644] env[62066]: ERROR nova.compute.manager [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 653.293644] env[62066]: ERROR nova.compute.manager [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] self.driver.spawn(context, instance, image_meta, [ 653.293644] env[62066]: ERROR nova.compute.manager [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 653.293644] env[62066]: ERROR nova.compute.manager [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] self._vmops.spawn(context, instance, image_meta, injected_files, [ 653.293644] env[62066]: ERROR nova.compute.manager [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 653.293644] env[62066]: ERROR nova.compute.manager [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] vm_ref = self.build_virtual_machine(instance, [ 653.293644] env[62066]: ERROR nova.compute.manager [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 653.293644] env[62066]: ERROR nova.compute.manager [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] vif_infos = vmwarevif.get_vif_info(self._session, [ 653.293644] env[62066]: ERROR nova.compute.manager [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 653.293996] env[62066]: ERROR nova.compute.manager [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] for vif in network_info: [ 653.293996] env[62066]: ERROR nova.compute.manager [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 653.293996] env[62066]: ERROR nova.compute.manager [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] return self._sync_wrapper(fn, *args, **kwargs) [ 653.293996] env[62066]: ERROR nova.compute.manager [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 653.293996] env[62066]: ERROR nova.compute.manager [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] self.wait() [ 653.293996] env[62066]: ERROR nova.compute.manager [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 653.293996] env[62066]: ERROR nova.compute.manager [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] self[:] = self._gt.wait() [ 653.293996] env[62066]: ERROR nova.compute.manager [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 653.293996] env[62066]: ERROR nova.compute.manager [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] return self._exit_event.wait() [ 653.293996] env[62066]: ERROR nova.compute.manager [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 653.293996] env[62066]: ERROR nova.compute.manager [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] result = hub.switch() [ 653.293996] env[62066]: ERROR nova.compute.manager [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 653.293996] env[62066]: ERROR nova.compute.manager [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] return self.greenlet.switch() [ 653.294526] env[62066]: ERROR nova.compute.manager [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 653.294526] env[62066]: ERROR nova.compute.manager [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] result = function(*args, **kwargs) [ 653.294526] env[62066]: ERROR nova.compute.manager [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 653.294526] env[62066]: ERROR nova.compute.manager [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] return func(*args, **kwargs) [ 653.294526] env[62066]: ERROR nova.compute.manager [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 653.294526] env[62066]: ERROR nova.compute.manager [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] raise e [ 653.294526] env[62066]: ERROR nova.compute.manager [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 653.294526] env[62066]: ERROR nova.compute.manager [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] nwinfo = self.network_api.allocate_for_instance( [ 653.294526] env[62066]: ERROR nova.compute.manager [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 653.294526] env[62066]: ERROR nova.compute.manager [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] created_port_ids = self._update_ports_for_instance( [ 653.294526] env[62066]: ERROR nova.compute.manager [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 653.294526] env[62066]: ERROR nova.compute.manager [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] with excutils.save_and_reraise_exception(): [ 653.294526] env[62066]: ERROR nova.compute.manager [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 653.294907] env[62066]: ERROR nova.compute.manager [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] self.force_reraise() [ 653.294907] env[62066]: ERROR nova.compute.manager [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 653.294907] env[62066]: ERROR nova.compute.manager [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] raise self.value [ 653.294907] env[62066]: ERROR nova.compute.manager [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 653.294907] env[62066]: ERROR nova.compute.manager [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] updated_port = self._update_port( [ 653.294907] env[62066]: ERROR nova.compute.manager [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 653.294907] env[62066]: ERROR nova.compute.manager [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] _ensure_no_port_binding_failure(port) [ 653.294907] env[62066]: ERROR nova.compute.manager [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 653.294907] env[62066]: ERROR nova.compute.manager [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] raise exception.PortBindingFailed(port_id=port['id']) [ 653.294907] env[62066]: ERROR nova.compute.manager [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] nova.exception.PortBindingFailed: Binding failed for port e09cd9e1-44fe-4858-bf9a-0afd7b681ca4, please check neutron logs for more information. [ 653.294907] env[62066]: ERROR nova.compute.manager [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] [ 653.295273] env[62066]: DEBUG nova.compute.utils [None req-a9ebb4d2-ff8d-4333-a9df-86618ea2d54c tempest-ImagesOneServerTestJSON-1591819949 tempest-ImagesOneServerTestJSON-1591819949-project-member] [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] Binding failed for port e09cd9e1-44fe-4858-bf9a-0afd7b681ca4, please check neutron logs for more information. {{(pid=62066) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 653.295626] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ef1737fc-6565-4a6f-bd10-02fc3b07f848 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.817s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 653.297081] env[62066]: INFO nova.compute.claims [None req-ef1737fc-6565-4a6f-bd10-02fc3b07f848 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 653.299717] env[62066]: DEBUG nova.compute.manager [None req-a9ebb4d2-ff8d-4333-a9df-86618ea2d54c tempest-ImagesOneServerTestJSON-1591819949 tempest-ImagesOneServerTestJSON-1591819949-project-member] [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] Build of instance 623300f7-54d2-4b0a-b356-a1ae585682cb was re-scheduled: Binding failed for port e09cd9e1-44fe-4858-bf9a-0afd7b681ca4, please check neutron logs for more information. {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 653.300177] env[62066]: DEBUG nova.compute.manager [None req-a9ebb4d2-ff8d-4333-a9df-86618ea2d54c tempest-ImagesOneServerTestJSON-1591819949 tempest-ImagesOneServerTestJSON-1591819949-project-member] [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] Unplugging VIFs for instance {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 653.300409] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a9ebb4d2-ff8d-4333-a9df-86618ea2d54c tempest-ImagesOneServerTestJSON-1591819949 tempest-ImagesOneServerTestJSON-1591819949-project-member] Acquiring lock "refresh_cache-623300f7-54d2-4b0a-b356-a1ae585682cb" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 653.300557] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a9ebb4d2-ff8d-4333-a9df-86618ea2d54c tempest-ImagesOneServerTestJSON-1591819949 tempest-ImagesOneServerTestJSON-1591819949-project-member] Acquired lock "refresh_cache-623300f7-54d2-4b0a-b356-a1ae585682cb" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 653.300718] env[62066]: DEBUG nova.network.neutron [None req-a9ebb4d2-ff8d-4333-a9df-86618ea2d54c tempest-ImagesOneServerTestJSON-1591819949 tempest-ImagesOneServerTestJSON-1591819949-project-member] [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 653.820187] env[62066]: DEBUG nova.network.neutron [None req-a9ebb4d2-ff8d-4333-a9df-86618ea2d54c tempest-ImagesOneServerTestJSON-1591819949 tempest-ImagesOneServerTestJSON-1591819949-project-member] [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 653.940639] env[62066]: DEBUG nova.network.neutron [None req-a9ebb4d2-ff8d-4333-a9df-86618ea2d54c tempest-ImagesOneServerTestJSON-1591819949 tempest-ImagesOneServerTestJSON-1591819949-project-member] [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 654.443488] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a9ebb4d2-ff8d-4333-a9df-86618ea2d54c tempest-ImagesOneServerTestJSON-1591819949 tempest-ImagesOneServerTestJSON-1591819949-project-member] Releasing lock "refresh_cache-623300f7-54d2-4b0a-b356-a1ae585682cb" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 654.443765] env[62066]: DEBUG nova.compute.manager [None req-a9ebb4d2-ff8d-4333-a9df-86618ea2d54c tempest-ImagesOneServerTestJSON-1591819949 tempest-ImagesOneServerTestJSON-1591819949-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 654.443765] env[62066]: DEBUG nova.compute.manager [None req-a9ebb4d2-ff8d-4333-a9df-86618ea2d54c tempest-ImagesOneServerTestJSON-1591819949 tempest-ImagesOneServerTestJSON-1591819949-project-member] [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 654.443904] env[62066]: DEBUG nova.network.neutron [None req-a9ebb4d2-ff8d-4333-a9df-86618ea2d54c tempest-ImagesOneServerTestJSON-1591819949 tempest-ImagesOneServerTestJSON-1591819949-project-member] [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 654.458736] env[62066]: DEBUG nova.network.neutron [None req-a9ebb4d2-ff8d-4333-a9df-86618ea2d54c tempest-ImagesOneServerTestJSON-1591819949 tempest-ImagesOneServerTestJSON-1591819949-project-member] [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 654.672345] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-764035c2-82f4-4ed2-bea1-0f37c48f1a4a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.679993] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86439d80-8687-4686-9e81-71a89de3775e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.709365] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aba70631-6518-4695-9e8c-b4bf18ec276f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.716987] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22f92f20-a9ad-44fd-8618-30163976c443 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.730859] env[62066]: DEBUG nova.compute.provider_tree [None req-ef1737fc-6565-4a6f-bd10-02fc3b07f848 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 654.962664] env[62066]: DEBUG nova.network.neutron [None req-a9ebb4d2-ff8d-4333-a9df-86618ea2d54c tempest-ImagesOneServerTestJSON-1591819949 tempest-ImagesOneServerTestJSON-1591819949-project-member] [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 655.234329] env[62066]: DEBUG nova.scheduler.client.report [None req-ef1737fc-6565-4a6f-bd10-02fc3b07f848 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 655.465889] env[62066]: INFO nova.compute.manager [None req-a9ebb4d2-ff8d-4333-a9df-86618ea2d54c tempest-ImagesOneServerTestJSON-1591819949 tempest-ImagesOneServerTestJSON-1591819949-project-member] [instance: 623300f7-54d2-4b0a-b356-a1ae585682cb] Took 1.02 seconds to deallocate network for instance. [ 655.739697] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ef1737fc-6565-4a6f-bd10-02fc3b07f848 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.444s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 655.740592] env[62066]: DEBUG nova.compute.manager [None req-ef1737fc-6565-4a6f-bd10-02fc3b07f848 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 655.744580] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9371bde4-46e3-4272-a1fe-5125bccbd2c7 tempest-ServersTestJSON-680983284 tempest-ServersTestJSON-680983284-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 29.175s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 656.249350] env[62066]: DEBUG nova.compute.utils [None req-ef1737fc-6565-4a6f-bd10-02fc3b07f848 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 656.251210] env[62066]: DEBUG nova.compute.manager [None req-ef1737fc-6565-4a6f-bd10-02fc3b07f848 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 656.251453] env[62066]: DEBUG nova.network.neutron [None req-ef1737fc-6565-4a6f-bd10-02fc3b07f848 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 656.310696] env[62066]: DEBUG nova.policy [None req-ef1737fc-6565-4a6f-bd10-02fc3b07f848 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '60e2bc61839b4299912c53137bc23fbc', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '75f0352c852947369474c9e05766c584', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 656.494932] env[62066]: INFO nova.scheduler.client.report [None req-a9ebb4d2-ff8d-4333-a9df-86618ea2d54c tempest-ImagesOneServerTestJSON-1591819949 tempest-ImagesOneServerTestJSON-1591819949-project-member] Deleted allocations for instance 623300f7-54d2-4b0a-b356-a1ae585682cb [ 656.674104] env[62066]: DEBUG nova.network.neutron [None req-ef1737fc-6565-4a6f-bd10-02fc3b07f848 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] Successfully created port: d8e26205-c295-47d1-b068-e154150e67e5 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 656.677956] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41e8925c-387d-4e20-8762-f5b946a9dbd9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.685627] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cba32b10-6c67-43b1-bf27-556237e6930c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.714939] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8d801a4-339f-455c-a305-c3b780444011 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.722108] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed834ea0-34b9-4f32-be0c-a9ef83ddd1cd {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.735973] env[62066]: DEBUG nova.compute.provider_tree [None req-9371bde4-46e3-4272-a1fe-5125bccbd2c7 tempest-ServersTestJSON-680983284 tempest-ServersTestJSON-680983284-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 656.757257] env[62066]: DEBUG nova.compute.manager [None req-ef1737fc-6565-4a6f-bd10-02fc3b07f848 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 657.008468] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a9ebb4d2-ff8d-4333-a9df-86618ea2d54c tempest-ImagesOneServerTestJSON-1591819949 tempest-ImagesOneServerTestJSON-1591819949-project-member] Lock "623300f7-54d2-4b0a-b356-a1ae585682cb" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 127.183s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 657.241840] env[62066]: DEBUG nova.scheduler.client.report [None req-9371bde4-46e3-4272-a1fe-5125bccbd2c7 tempest-ServersTestJSON-680983284 tempest-ServersTestJSON-680983284-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 657.510988] env[62066]: DEBUG nova.compute.manager [None req-3515d98a-2869-41bb-bf8e-e38d995a4d22 tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 657.746079] env[62066]: DEBUG nova.compute.manager [req-f24d14dc-6793-4a92-a328-78b7d9e779a4 req-9620fccf-17e7-4c7e-9816-05d54253fd1d service nova] [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] Received event network-changed-d8e26205-c295-47d1-b068-e154150e67e5 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 657.746079] env[62066]: DEBUG nova.compute.manager [req-f24d14dc-6793-4a92-a328-78b7d9e779a4 req-9620fccf-17e7-4c7e-9816-05d54253fd1d service nova] [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] Refreshing instance network info cache due to event network-changed-d8e26205-c295-47d1-b068-e154150e67e5. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 657.746079] env[62066]: DEBUG oslo_concurrency.lockutils [req-f24d14dc-6793-4a92-a328-78b7d9e779a4 req-9620fccf-17e7-4c7e-9816-05d54253fd1d service nova] Acquiring lock "refresh_cache-97e51e6c-9a3a-4b68-b737-d48090a22b1b" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 657.746079] env[62066]: DEBUG oslo_concurrency.lockutils [req-f24d14dc-6793-4a92-a328-78b7d9e779a4 req-9620fccf-17e7-4c7e-9816-05d54253fd1d service nova] Acquired lock "refresh_cache-97e51e6c-9a3a-4b68-b737-d48090a22b1b" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 657.746079] env[62066]: DEBUG nova.network.neutron [req-f24d14dc-6793-4a92-a328-78b7d9e779a4 req-9620fccf-17e7-4c7e-9816-05d54253fd1d service nova] [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] Refreshing network info cache for port d8e26205-c295-47d1-b068-e154150e67e5 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 657.747559] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9371bde4-46e3-4272-a1fe-5125bccbd2c7 tempest-ServersTestJSON-680983284 tempest-ServersTestJSON-680983284-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.003s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 657.748222] env[62066]: ERROR nova.compute.manager [None req-9371bde4-46e3-4272-a1fe-5125bccbd2c7 tempest-ServersTestJSON-680983284 tempest-ServersTestJSON-680983284-project-member] [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 3e3a7d95-33be-49e4-9f86-f4315cbe8f18, please check neutron logs for more information. [ 657.748222] env[62066]: ERROR nova.compute.manager [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] Traceback (most recent call last): [ 657.748222] env[62066]: ERROR nova.compute.manager [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 657.748222] env[62066]: ERROR nova.compute.manager [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] self.driver.spawn(context, instance, image_meta, [ 657.748222] env[62066]: ERROR nova.compute.manager [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 657.748222] env[62066]: ERROR nova.compute.manager [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] self._vmops.spawn(context, instance, image_meta, injected_files, [ 657.748222] env[62066]: ERROR nova.compute.manager [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 657.748222] env[62066]: ERROR nova.compute.manager [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] vm_ref = self.build_virtual_machine(instance, [ 657.748222] env[62066]: ERROR nova.compute.manager [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 657.748222] env[62066]: ERROR nova.compute.manager [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] vif_infos = vmwarevif.get_vif_info(self._session, [ 657.748222] env[62066]: ERROR nova.compute.manager [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 657.748670] env[62066]: ERROR nova.compute.manager [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] for vif in network_info: [ 657.748670] env[62066]: ERROR nova.compute.manager [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 657.748670] env[62066]: ERROR nova.compute.manager [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] return self._sync_wrapper(fn, *args, **kwargs) [ 657.748670] env[62066]: ERROR nova.compute.manager [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 657.748670] env[62066]: ERROR nova.compute.manager [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] self.wait() [ 657.748670] env[62066]: ERROR nova.compute.manager [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 657.748670] env[62066]: ERROR nova.compute.manager [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] self[:] = self._gt.wait() [ 657.748670] env[62066]: ERROR nova.compute.manager [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 657.748670] env[62066]: ERROR nova.compute.manager [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] return self._exit_event.wait() [ 657.748670] env[62066]: ERROR nova.compute.manager [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 657.748670] env[62066]: ERROR nova.compute.manager [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] result = hub.switch() [ 657.748670] env[62066]: ERROR nova.compute.manager [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 657.748670] env[62066]: ERROR nova.compute.manager [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] return self.greenlet.switch() [ 657.749212] env[62066]: ERROR nova.compute.manager [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 657.749212] env[62066]: ERROR nova.compute.manager [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] result = function(*args, **kwargs) [ 657.749212] env[62066]: ERROR nova.compute.manager [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 657.749212] env[62066]: ERROR nova.compute.manager [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] return func(*args, **kwargs) [ 657.749212] env[62066]: ERROR nova.compute.manager [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 657.749212] env[62066]: ERROR nova.compute.manager [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] raise e [ 657.749212] env[62066]: ERROR nova.compute.manager [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 657.749212] env[62066]: ERROR nova.compute.manager [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] nwinfo = self.network_api.allocate_for_instance( [ 657.749212] env[62066]: ERROR nova.compute.manager [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 657.749212] env[62066]: ERROR nova.compute.manager [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] created_port_ids = self._update_ports_for_instance( [ 657.749212] env[62066]: ERROR nova.compute.manager [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 657.749212] env[62066]: ERROR nova.compute.manager [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] with excutils.save_and_reraise_exception(): [ 657.749212] env[62066]: ERROR nova.compute.manager [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 657.749554] env[62066]: ERROR nova.compute.manager [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] self.force_reraise() [ 657.749554] env[62066]: ERROR nova.compute.manager [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 657.749554] env[62066]: ERROR nova.compute.manager [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] raise self.value [ 657.749554] env[62066]: ERROR nova.compute.manager [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 657.749554] env[62066]: ERROR nova.compute.manager [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] updated_port = self._update_port( [ 657.749554] env[62066]: ERROR nova.compute.manager [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 657.749554] env[62066]: ERROR nova.compute.manager [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] _ensure_no_port_binding_failure(port) [ 657.749554] env[62066]: ERROR nova.compute.manager [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 657.749554] env[62066]: ERROR nova.compute.manager [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] raise exception.PortBindingFailed(port_id=port['id']) [ 657.749554] env[62066]: ERROR nova.compute.manager [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] nova.exception.PortBindingFailed: Binding failed for port 3e3a7d95-33be-49e4-9f86-f4315cbe8f18, please check neutron logs for more information. [ 657.749554] env[62066]: ERROR nova.compute.manager [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] [ 657.749835] env[62066]: DEBUG nova.compute.utils [None req-9371bde4-46e3-4272-a1fe-5125bccbd2c7 tempest-ServersTestJSON-680983284 tempest-ServersTestJSON-680983284-project-member] [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] Binding failed for port 3e3a7d95-33be-49e4-9f86-f4315cbe8f18, please check neutron logs for more information. {{(pid=62066) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 657.750956] env[62066]: DEBUG nova.compute.manager [None req-9371bde4-46e3-4272-a1fe-5125bccbd2c7 tempest-ServersTestJSON-680983284 tempest-ServersTestJSON-680983284-project-member] [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] Build of instance 0040f46d-9a60-4fa1-8a00-c08022e17df0 was re-scheduled: Binding failed for port 3e3a7d95-33be-49e4-9f86-f4315cbe8f18, please check neutron logs for more information. {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 657.752139] env[62066]: DEBUG nova.compute.manager [None req-9371bde4-46e3-4272-a1fe-5125bccbd2c7 tempest-ServersTestJSON-680983284 tempest-ServersTestJSON-680983284-project-member] [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] Unplugging VIFs for instance {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 657.752626] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9371bde4-46e3-4272-a1fe-5125bccbd2c7 tempest-ServersTestJSON-680983284 tempest-ServersTestJSON-680983284-project-member] Acquiring lock "refresh_cache-0040f46d-9a60-4fa1-8a00-c08022e17df0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 657.752790] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9371bde4-46e3-4272-a1fe-5125bccbd2c7 tempest-ServersTestJSON-680983284 tempest-ServersTestJSON-680983284-project-member] Acquired lock "refresh_cache-0040f46d-9a60-4fa1-8a00-c08022e17df0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 657.752949] env[62066]: DEBUG nova.network.neutron [None req-9371bde4-46e3-4272-a1fe-5125bccbd2c7 tempest-ServersTestJSON-680983284 tempest-ServersTestJSON-680983284-project-member] [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 657.755642] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7abf52e5-ee7c-4f92-a757-ffa8ca735a57 tempest-ServerActionsTestJSON-164481335 tempest-ServerActionsTestJSON-164481335-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 30.489s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 657.767414] env[62066]: DEBUG nova.compute.manager [None req-ef1737fc-6565-4a6f-bd10-02fc3b07f848 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 657.800712] env[62066]: DEBUG nova.virt.hardware [None req-ef1737fc-6565-4a6f-bd10-02fc3b07f848 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-23T13:40:41Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-23T13:40:23Z,direct_url=,disk_format='vmdk',id=50ff584c-3b50-4395-af07-3e66769bc9f7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='eb52f7069a374c61ae946f052007c6d9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-23T13:40:24Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 657.800903] env[62066]: DEBUG nova.virt.hardware [None req-ef1737fc-6565-4a6f-bd10-02fc3b07f848 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 657.801222] env[62066]: DEBUG nova.virt.hardware [None req-ef1737fc-6565-4a6f-bd10-02fc3b07f848 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 657.801896] env[62066]: DEBUG nova.virt.hardware [None req-ef1737fc-6565-4a6f-bd10-02fc3b07f848 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 657.802204] env[62066]: DEBUG nova.virt.hardware [None req-ef1737fc-6565-4a6f-bd10-02fc3b07f848 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 657.802494] env[62066]: DEBUG nova.virt.hardware [None req-ef1737fc-6565-4a6f-bd10-02fc3b07f848 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 657.802769] env[62066]: DEBUG nova.virt.hardware [None req-ef1737fc-6565-4a6f-bd10-02fc3b07f848 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 657.803030] env[62066]: DEBUG nova.virt.hardware [None req-ef1737fc-6565-4a6f-bd10-02fc3b07f848 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 657.803251] env[62066]: DEBUG nova.virt.hardware [None req-ef1737fc-6565-4a6f-bd10-02fc3b07f848 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 657.803930] env[62066]: DEBUG nova.virt.hardware [None req-ef1737fc-6565-4a6f-bd10-02fc3b07f848 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 657.803930] env[62066]: DEBUG nova.virt.hardware [None req-ef1737fc-6565-4a6f-bd10-02fc3b07f848 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 657.805465] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a5a4620-6ef7-4469-ad02-22088dae57f2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.815494] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4cf00b5-08f7-40a9-bdc0-08a663b7a965 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.893967] env[62066]: ERROR nova.compute.manager [None req-ef1737fc-6565-4a6f-bd10-02fc3b07f848 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port d8e26205-c295-47d1-b068-e154150e67e5, please check neutron logs for more information. [ 657.893967] env[62066]: ERROR nova.compute.manager Traceback (most recent call last): [ 657.893967] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 657.893967] env[62066]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 657.893967] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 657.893967] env[62066]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 657.893967] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 657.893967] env[62066]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 657.893967] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 657.893967] env[62066]: ERROR nova.compute.manager self.force_reraise() [ 657.893967] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 657.893967] env[62066]: ERROR nova.compute.manager raise self.value [ 657.893967] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 657.893967] env[62066]: ERROR nova.compute.manager updated_port = self._update_port( [ 657.893967] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 657.893967] env[62066]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 657.894675] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 657.894675] env[62066]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 657.894675] env[62066]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port d8e26205-c295-47d1-b068-e154150e67e5, please check neutron logs for more information. [ 657.894675] env[62066]: ERROR nova.compute.manager [ 657.894675] env[62066]: Traceback (most recent call last): [ 657.894675] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 657.894675] env[62066]: listener.cb(fileno) [ 657.894675] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 657.894675] env[62066]: result = function(*args, **kwargs) [ 657.894675] env[62066]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 657.894675] env[62066]: return func(*args, **kwargs) [ 657.894675] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 657.894675] env[62066]: raise e [ 657.894675] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 657.894675] env[62066]: nwinfo = self.network_api.allocate_for_instance( [ 657.894675] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 657.894675] env[62066]: created_port_ids = self._update_ports_for_instance( [ 657.894675] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 657.894675] env[62066]: with excutils.save_and_reraise_exception(): [ 657.894675] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 657.894675] env[62066]: self.force_reraise() [ 657.894675] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 657.894675] env[62066]: raise self.value [ 657.894675] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 657.894675] env[62066]: updated_port = self._update_port( [ 657.894675] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 657.894675] env[62066]: _ensure_no_port_binding_failure(port) [ 657.894675] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 657.894675] env[62066]: raise exception.PortBindingFailed(port_id=port['id']) [ 657.895823] env[62066]: nova.exception.PortBindingFailed: Binding failed for port d8e26205-c295-47d1-b068-e154150e67e5, please check neutron logs for more information. [ 657.895823] env[62066]: Removing descriptor: 20 [ 657.895823] env[62066]: ERROR nova.compute.manager [None req-ef1737fc-6565-4a6f-bd10-02fc3b07f848 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port d8e26205-c295-47d1-b068-e154150e67e5, please check neutron logs for more information. [ 657.895823] env[62066]: ERROR nova.compute.manager [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] Traceback (most recent call last): [ 657.895823] env[62066]: ERROR nova.compute.manager [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 657.895823] env[62066]: ERROR nova.compute.manager [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] yield resources [ 657.895823] env[62066]: ERROR nova.compute.manager [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 657.895823] env[62066]: ERROR nova.compute.manager [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] self.driver.spawn(context, instance, image_meta, [ 657.895823] env[62066]: ERROR nova.compute.manager [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 657.895823] env[62066]: ERROR nova.compute.manager [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 657.895823] env[62066]: ERROR nova.compute.manager [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 657.895823] env[62066]: ERROR nova.compute.manager [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] vm_ref = self.build_virtual_machine(instance, [ 657.896212] env[62066]: ERROR nova.compute.manager [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 657.896212] env[62066]: ERROR nova.compute.manager [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] vif_infos = vmwarevif.get_vif_info(self._session, [ 657.896212] env[62066]: ERROR nova.compute.manager [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 657.896212] env[62066]: ERROR nova.compute.manager [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] for vif in network_info: [ 657.896212] env[62066]: ERROR nova.compute.manager [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 657.896212] env[62066]: ERROR nova.compute.manager [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] return self._sync_wrapper(fn, *args, **kwargs) [ 657.896212] env[62066]: ERROR nova.compute.manager [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 657.896212] env[62066]: ERROR nova.compute.manager [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] self.wait() [ 657.896212] env[62066]: ERROR nova.compute.manager [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 657.896212] env[62066]: ERROR nova.compute.manager [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] self[:] = self._gt.wait() [ 657.896212] env[62066]: ERROR nova.compute.manager [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 657.896212] env[62066]: ERROR nova.compute.manager [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] return self._exit_event.wait() [ 657.896212] env[62066]: ERROR nova.compute.manager [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 657.896543] env[62066]: ERROR nova.compute.manager [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] result = hub.switch() [ 657.896543] env[62066]: ERROR nova.compute.manager [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 657.896543] env[62066]: ERROR nova.compute.manager [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] return self.greenlet.switch() [ 657.896543] env[62066]: ERROR nova.compute.manager [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 657.896543] env[62066]: ERROR nova.compute.manager [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] result = function(*args, **kwargs) [ 657.896543] env[62066]: ERROR nova.compute.manager [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 657.896543] env[62066]: ERROR nova.compute.manager [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] return func(*args, **kwargs) [ 657.896543] env[62066]: ERROR nova.compute.manager [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 657.896543] env[62066]: ERROR nova.compute.manager [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] raise e [ 657.896543] env[62066]: ERROR nova.compute.manager [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 657.896543] env[62066]: ERROR nova.compute.manager [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] nwinfo = self.network_api.allocate_for_instance( [ 657.896543] env[62066]: ERROR nova.compute.manager [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 657.896543] env[62066]: ERROR nova.compute.manager [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] created_port_ids = self._update_ports_for_instance( [ 657.897486] env[62066]: ERROR nova.compute.manager [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 657.897486] env[62066]: ERROR nova.compute.manager [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] with excutils.save_and_reraise_exception(): [ 657.897486] env[62066]: ERROR nova.compute.manager [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 657.897486] env[62066]: ERROR nova.compute.manager [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] self.force_reraise() [ 657.897486] env[62066]: ERROR nova.compute.manager [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 657.897486] env[62066]: ERROR nova.compute.manager [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] raise self.value [ 657.897486] env[62066]: ERROR nova.compute.manager [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 657.897486] env[62066]: ERROR nova.compute.manager [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] updated_port = self._update_port( [ 657.897486] env[62066]: ERROR nova.compute.manager [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 657.897486] env[62066]: ERROR nova.compute.manager [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] _ensure_no_port_binding_failure(port) [ 657.897486] env[62066]: ERROR nova.compute.manager [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 657.897486] env[62066]: ERROR nova.compute.manager [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] raise exception.PortBindingFailed(port_id=port['id']) [ 657.897879] env[62066]: ERROR nova.compute.manager [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] nova.exception.PortBindingFailed: Binding failed for port d8e26205-c295-47d1-b068-e154150e67e5, please check neutron logs for more information. [ 657.897879] env[62066]: ERROR nova.compute.manager [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] [ 657.897879] env[62066]: INFO nova.compute.manager [None req-ef1737fc-6565-4a6f-bd10-02fc3b07f848 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] Terminating instance [ 657.897879] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ef1737fc-6565-4a6f-bd10-02fc3b07f848 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Acquiring lock "refresh_cache-97e51e6c-9a3a-4b68-b737-d48090a22b1b" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 658.039842] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3515d98a-2869-41bb-bf8e-e38d995a4d22 tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 658.274422] env[62066]: DEBUG nova.network.neutron [None req-9371bde4-46e3-4272-a1fe-5125bccbd2c7 tempest-ServersTestJSON-680983284 tempest-ServersTestJSON-680983284-project-member] [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 658.279158] env[62066]: DEBUG nova.network.neutron [req-f24d14dc-6793-4a92-a328-78b7d9e779a4 req-9620fccf-17e7-4c7e-9816-05d54253fd1d service nova] [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 658.374127] env[62066]: DEBUG nova.network.neutron [None req-9371bde4-46e3-4272-a1fe-5125bccbd2c7 tempest-ServersTestJSON-680983284 tempest-ServersTestJSON-680983284-project-member] [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 658.475982] env[62066]: DEBUG nova.network.neutron [req-f24d14dc-6793-4a92-a328-78b7d9e779a4 req-9620fccf-17e7-4c7e-9816-05d54253fd1d service nova] [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 658.772670] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1916a05b-9dc9-4f04-9264-e49fd20c2e72 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.780822] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac023315-783c-4a29-bdd7-45c091b564a1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.819861] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32d795f5-dd4a-41ea-a6c7-a03205c4f9d8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.828893] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb7b16ad-c259-4a89-9ba9-0d258afe2702 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.851935] env[62066]: DEBUG nova.compute.provider_tree [None req-7abf52e5-ee7c-4f92-a757-ffa8ca735a57 tempest-ServerActionsTestJSON-164481335 tempest-ServerActionsTestJSON-164481335-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 658.877133] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9371bde4-46e3-4272-a1fe-5125bccbd2c7 tempest-ServersTestJSON-680983284 tempest-ServersTestJSON-680983284-project-member] Releasing lock "refresh_cache-0040f46d-9a60-4fa1-8a00-c08022e17df0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 658.877380] env[62066]: DEBUG nova.compute.manager [None req-9371bde4-46e3-4272-a1fe-5125bccbd2c7 tempest-ServersTestJSON-680983284 tempest-ServersTestJSON-680983284-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 658.877564] env[62066]: DEBUG nova.compute.manager [None req-9371bde4-46e3-4272-a1fe-5125bccbd2c7 tempest-ServersTestJSON-680983284 tempest-ServersTestJSON-680983284-project-member] [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 658.877729] env[62066]: DEBUG nova.network.neutron [None req-9371bde4-46e3-4272-a1fe-5125bccbd2c7 tempest-ServersTestJSON-680983284 tempest-ServersTestJSON-680983284-project-member] [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 658.894072] env[62066]: DEBUG nova.network.neutron [None req-9371bde4-46e3-4272-a1fe-5125bccbd2c7 tempest-ServersTestJSON-680983284 tempest-ServersTestJSON-680983284-project-member] [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 658.981268] env[62066]: DEBUG oslo_concurrency.lockutils [req-f24d14dc-6793-4a92-a328-78b7d9e779a4 req-9620fccf-17e7-4c7e-9816-05d54253fd1d service nova] Releasing lock "refresh_cache-97e51e6c-9a3a-4b68-b737-d48090a22b1b" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 658.981469] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ef1737fc-6565-4a6f-bd10-02fc3b07f848 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Acquired lock "refresh_cache-97e51e6c-9a3a-4b68-b737-d48090a22b1b" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 658.981649] env[62066]: DEBUG nova.network.neutron [None req-ef1737fc-6565-4a6f-bd10-02fc3b07f848 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 659.353061] env[62066]: DEBUG nova.scheduler.client.report [None req-7abf52e5-ee7c-4f92-a757-ffa8ca735a57 tempest-ServerActionsTestJSON-164481335 tempest-ServerActionsTestJSON-164481335-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 659.396770] env[62066]: DEBUG nova.network.neutron [None req-9371bde4-46e3-4272-a1fe-5125bccbd2c7 tempest-ServersTestJSON-680983284 tempest-ServersTestJSON-680983284-project-member] [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 659.504496] env[62066]: DEBUG nova.network.neutron [None req-ef1737fc-6565-4a6f-bd10-02fc3b07f848 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 659.594507] env[62066]: DEBUG nova.network.neutron [None req-ef1737fc-6565-4a6f-bd10-02fc3b07f848 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 659.773178] env[62066]: DEBUG nova.compute.manager [req-c692b3dd-798d-4664-a1fe-935e8dae32a2 req-40c47a2e-d7b4-42a7-af32-3a398915d954 service nova] [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] Received event network-vif-deleted-d8e26205-c295-47d1-b068-e154150e67e5 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 659.857607] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7abf52e5-ee7c-4f92-a757-ffa8ca735a57 tempest-ServerActionsTestJSON-164481335 tempest-ServerActionsTestJSON-164481335-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.102s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 659.858523] env[62066]: ERROR nova.compute.manager [None req-7abf52e5-ee7c-4f92-a757-ffa8ca735a57 tempest-ServerActionsTestJSON-164481335 tempest-ServerActionsTestJSON-164481335-project-member] [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port c80661ff-38e1-4e3e-9647-43d6309380db, please check neutron logs for more information. [ 659.858523] env[62066]: ERROR nova.compute.manager [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] Traceback (most recent call last): [ 659.858523] env[62066]: ERROR nova.compute.manager [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 659.858523] env[62066]: ERROR nova.compute.manager [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] self.driver.spawn(context, instance, image_meta, [ 659.858523] env[62066]: ERROR nova.compute.manager [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 659.858523] env[62066]: ERROR nova.compute.manager [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] self._vmops.spawn(context, instance, image_meta, injected_files, [ 659.858523] env[62066]: ERROR nova.compute.manager [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 659.858523] env[62066]: ERROR nova.compute.manager [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] vm_ref = self.build_virtual_machine(instance, [ 659.858523] env[62066]: ERROR nova.compute.manager [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 659.858523] env[62066]: ERROR nova.compute.manager [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] vif_infos = vmwarevif.get_vif_info(self._session, [ 659.858523] env[62066]: ERROR nova.compute.manager [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 659.858889] env[62066]: ERROR nova.compute.manager [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] for vif in network_info: [ 659.858889] env[62066]: ERROR nova.compute.manager [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 659.858889] env[62066]: ERROR nova.compute.manager [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] return self._sync_wrapper(fn, *args, **kwargs) [ 659.858889] env[62066]: ERROR nova.compute.manager [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 659.858889] env[62066]: ERROR nova.compute.manager [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] self.wait() [ 659.858889] env[62066]: ERROR nova.compute.manager [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 659.858889] env[62066]: ERROR nova.compute.manager [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] self[:] = self._gt.wait() [ 659.858889] env[62066]: ERROR nova.compute.manager [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 659.858889] env[62066]: ERROR nova.compute.manager [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] return self._exit_event.wait() [ 659.858889] env[62066]: ERROR nova.compute.manager [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 659.858889] env[62066]: ERROR nova.compute.manager [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] result = hub.switch() [ 659.858889] env[62066]: ERROR nova.compute.manager [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 659.858889] env[62066]: ERROR nova.compute.manager [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] return self.greenlet.switch() [ 659.859232] env[62066]: ERROR nova.compute.manager [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 659.859232] env[62066]: ERROR nova.compute.manager [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] result = function(*args, **kwargs) [ 659.859232] env[62066]: ERROR nova.compute.manager [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 659.859232] env[62066]: ERROR nova.compute.manager [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] return func(*args, **kwargs) [ 659.859232] env[62066]: ERROR nova.compute.manager [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 659.859232] env[62066]: ERROR nova.compute.manager [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] raise e [ 659.859232] env[62066]: ERROR nova.compute.manager [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 659.859232] env[62066]: ERROR nova.compute.manager [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] nwinfo = self.network_api.allocate_for_instance( [ 659.859232] env[62066]: ERROR nova.compute.manager [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 659.859232] env[62066]: ERROR nova.compute.manager [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] created_port_ids = self._update_ports_for_instance( [ 659.859232] env[62066]: ERROR nova.compute.manager [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 659.859232] env[62066]: ERROR nova.compute.manager [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] with excutils.save_and_reraise_exception(): [ 659.859232] env[62066]: ERROR nova.compute.manager [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 659.859549] env[62066]: ERROR nova.compute.manager [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] self.force_reraise() [ 659.859549] env[62066]: ERROR nova.compute.manager [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 659.859549] env[62066]: ERROR nova.compute.manager [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] raise self.value [ 659.859549] env[62066]: ERROR nova.compute.manager [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 659.859549] env[62066]: ERROR nova.compute.manager [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] updated_port = self._update_port( [ 659.859549] env[62066]: ERROR nova.compute.manager [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 659.859549] env[62066]: ERROR nova.compute.manager [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] _ensure_no_port_binding_failure(port) [ 659.859549] env[62066]: ERROR nova.compute.manager [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 659.859549] env[62066]: ERROR nova.compute.manager [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] raise exception.PortBindingFailed(port_id=port['id']) [ 659.859549] env[62066]: ERROR nova.compute.manager [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] nova.exception.PortBindingFailed: Binding failed for port c80661ff-38e1-4e3e-9647-43d6309380db, please check neutron logs for more information. [ 659.859549] env[62066]: ERROR nova.compute.manager [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] [ 659.859826] env[62066]: DEBUG nova.compute.utils [None req-7abf52e5-ee7c-4f92-a757-ffa8ca735a57 tempest-ServerActionsTestJSON-164481335 tempest-ServerActionsTestJSON-164481335-project-member] [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] Binding failed for port c80661ff-38e1-4e3e-9647-43d6309380db, please check neutron logs for more information. {{(pid=62066) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 659.860519] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4d395966-f335-497d-b8a0-23e22d9b79f7 tempest-ServerActionsTestOtherA-1520864800 tempest-ServerActionsTestOtherA-1520864800-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 29.880s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 659.862047] env[62066]: INFO nova.compute.claims [None req-4d395966-f335-497d-b8a0-23e22d9b79f7 tempest-ServerActionsTestOtherA-1520864800 tempest-ServerActionsTestOtherA-1520864800-project-member] [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 659.864796] env[62066]: DEBUG nova.compute.manager [None req-7abf52e5-ee7c-4f92-a757-ffa8ca735a57 tempest-ServerActionsTestJSON-164481335 tempest-ServerActionsTestJSON-164481335-project-member] [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] Build of instance bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94 was re-scheduled: Binding failed for port c80661ff-38e1-4e3e-9647-43d6309380db, please check neutron logs for more information. {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 659.865245] env[62066]: DEBUG nova.compute.manager [None req-7abf52e5-ee7c-4f92-a757-ffa8ca735a57 tempest-ServerActionsTestJSON-164481335 tempest-ServerActionsTestJSON-164481335-project-member] [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] Unplugging VIFs for instance {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 659.865469] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7abf52e5-ee7c-4f92-a757-ffa8ca735a57 tempest-ServerActionsTestJSON-164481335 tempest-ServerActionsTestJSON-164481335-project-member] Acquiring lock "refresh_cache-bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 659.865614] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7abf52e5-ee7c-4f92-a757-ffa8ca735a57 tempest-ServerActionsTestJSON-164481335 tempest-ServerActionsTestJSON-164481335-project-member] Acquired lock "refresh_cache-bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 659.865798] env[62066]: DEBUG nova.network.neutron [None req-7abf52e5-ee7c-4f92-a757-ffa8ca735a57 tempest-ServerActionsTestJSON-164481335 tempest-ServerActionsTestJSON-164481335-project-member] [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 659.899626] env[62066]: INFO nova.compute.manager [None req-9371bde4-46e3-4272-a1fe-5125bccbd2c7 tempest-ServersTestJSON-680983284 tempest-ServersTestJSON-680983284-project-member] [instance: 0040f46d-9a60-4fa1-8a00-c08022e17df0] Took 1.02 seconds to deallocate network for instance. [ 660.100016] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ef1737fc-6565-4a6f-bd10-02fc3b07f848 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Releasing lock "refresh_cache-97e51e6c-9a3a-4b68-b737-d48090a22b1b" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 660.100016] env[62066]: DEBUG nova.compute.manager [None req-ef1737fc-6565-4a6f-bd10-02fc3b07f848 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 660.100016] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-ef1737fc-6565-4a6f-bd10-02fc3b07f848 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 660.100016] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8fd10114-974d-4e82-b0ad-4b7f1292e6eb {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.108080] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1e8cc0c-a0c6-4be5-9f4c-a663c45207bf {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.131231] env[62066]: WARNING nova.virt.vmwareapi.vmops [None req-ef1737fc-6565-4a6f-bd10-02fc3b07f848 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 97e51e6c-9a3a-4b68-b737-d48090a22b1b could not be found. [ 660.131475] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-ef1737fc-6565-4a6f-bd10-02fc3b07f848 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 660.131661] env[62066]: INFO nova.compute.manager [None req-ef1737fc-6565-4a6f-bd10-02fc3b07f848 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] Took 0.03 seconds to destroy the instance on the hypervisor. [ 660.131917] env[62066]: DEBUG oslo.service.loopingcall [None req-ef1737fc-6565-4a6f-bd10-02fc3b07f848 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 660.132163] env[62066]: DEBUG nova.compute.manager [-] [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 660.132260] env[62066]: DEBUG nova.network.neutron [-] [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 660.151221] env[62066]: DEBUG nova.network.neutron [-] [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 660.391537] env[62066]: DEBUG nova.network.neutron [None req-7abf52e5-ee7c-4f92-a757-ffa8ca735a57 tempest-ServerActionsTestJSON-164481335 tempest-ServerActionsTestJSON-164481335-project-member] [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 660.521915] env[62066]: DEBUG nova.network.neutron [None req-7abf52e5-ee7c-4f92-a757-ffa8ca735a57 tempest-ServerActionsTestJSON-164481335 tempest-ServerActionsTestJSON-164481335-project-member] [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 660.654160] env[62066]: DEBUG nova.network.neutron [-] [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 660.928832] env[62066]: INFO nova.scheduler.client.report [None req-9371bde4-46e3-4272-a1fe-5125bccbd2c7 tempest-ServersTestJSON-680983284 tempest-ServersTestJSON-680983284-project-member] Deleted allocations for instance 0040f46d-9a60-4fa1-8a00-c08022e17df0 [ 661.025537] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7abf52e5-ee7c-4f92-a757-ffa8ca735a57 tempest-ServerActionsTestJSON-164481335 tempest-ServerActionsTestJSON-164481335-project-member] Releasing lock "refresh_cache-bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 661.025537] env[62066]: DEBUG nova.compute.manager [None req-7abf52e5-ee7c-4f92-a757-ffa8ca735a57 tempest-ServerActionsTestJSON-164481335 tempest-ServerActionsTestJSON-164481335-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 661.025799] env[62066]: DEBUG nova.compute.manager [None req-7abf52e5-ee7c-4f92-a757-ffa8ca735a57 tempest-ServerActionsTestJSON-164481335 tempest-ServerActionsTestJSON-164481335-project-member] [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 661.025863] env[62066]: DEBUG nova.network.neutron [None req-7abf52e5-ee7c-4f92-a757-ffa8ca735a57 tempest-ServerActionsTestJSON-164481335 tempest-ServerActionsTestJSON-164481335-project-member] [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 661.042407] env[62066]: DEBUG nova.network.neutron [None req-7abf52e5-ee7c-4f92-a757-ffa8ca735a57 tempest-ServerActionsTestJSON-164481335 tempest-ServerActionsTestJSON-164481335-project-member] [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 661.156905] env[62066]: INFO nova.compute.manager [-] [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] Took 1.02 seconds to deallocate network for instance. [ 661.159392] env[62066]: DEBUG nova.compute.claims [None req-ef1737fc-6565-4a6f-bd10-02fc3b07f848 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] Aborting claim: {{(pid=62066) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 661.159568] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ef1737fc-6565-4a6f-bd10-02fc3b07f848 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 661.260485] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83ae1d5e-fb86-4fad-b7eb-85452397c6aa {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.268105] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1cb40eb4-ee20-434e-a837-f11dd0aebecf {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.300539] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5d6ab88-8bd0-4865-920c-590777c0a859 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.307905] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cec2c257-344c-41b4-b363-8a9962139948 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.321431] env[62066]: DEBUG nova.compute.provider_tree [None req-4d395966-f335-497d-b8a0-23e22d9b79f7 tempest-ServerActionsTestOtherA-1520864800 tempest-ServerActionsTestOtherA-1520864800-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 661.439049] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9371bde4-46e3-4272-a1fe-5125bccbd2c7 tempest-ServersTestJSON-680983284 tempest-ServersTestJSON-680983284-project-member] Lock "0040f46d-9a60-4fa1-8a00-c08022e17df0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 130.027s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 661.545587] env[62066]: DEBUG nova.network.neutron [None req-7abf52e5-ee7c-4f92-a757-ffa8ca735a57 tempest-ServerActionsTestJSON-164481335 tempest-ServerActionsTestJSON-164481335-project-member] [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 661.824670] env[62066]: DEBUG nova.scheduler.client.report [None req-4d395966-f335-497d-b8a0-23e22d9b79f7 tempest-ServerActionsTestOtherA-1520864800 tempest-ServerActionsTestOtherA-1520864800-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 661.941861] env[62066]: DEBUG nova.compute.manager [None req-60a20a80-792e-4a03-84da-c17f8b28425d tempest-InstanceActionsNegativeTestJSON-1895685779 tempest-InstanceActionsNegativeTestJSON-1895685779-project-member] [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 662.048370] env[62066]: INFO nova.compute.manager [None req-7abf52e5-ee7c-4f92-a757-ffa8ca735a57 tempest-ServerActionsTestJSON-164481335 tempest-ServerActionsTestJSON-164481335-project-member] [instance: bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94] Took 1.02 seconds to deallocate network for instance. [ 662.331026] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4d395966-f335-497d-b8a0-23e22d9b79f7 tempest-ServerActionsTestOtherA-1520864800 tempest-ServerActionsTestOtherA-1520864800-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.470s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 662.331026] env[62066]: DEBUG nova.compute.manager [None req-4d395966-f335-497d-b8a0-23e22d9b79f7 tempest-ServerActionsTestOtherA-1520864800 tempest-ServerActionsTestOtherA-1520864800-project-member] [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 662.333456] env[62066]: DEBUG oslo_concurrency.lockutils [None req-29878964-568e-4c92-80e6-64ee3e8e11bb tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 29.087s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 662.463669] env[62066]: DEBUG oslo_concurrency.lockutils [None req-60a20a80-792e-4a03-84da-c17f8b28425d tempest-InstanceActionsNegativeTestJSON-1895685779 tempest-InstanceActionsNegativeTestJSON-1895685779-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 662.838471] env[62066]: DEBUG nova.compute.utils [None req-4d395966-f335-497d-b8a0-23e22d9b79f7 tempest-ServerActionsTestOtherA-1520864800 tempest-ServerActionsTestOtherA-1520864800-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 662.843032] env[62066]: DEBUG nova.compute.manager [None req-4d395966-f335-497d-b8a0-23e22d9b79f7 tempest-ServerActionsTestOtherA-1520864800 tempest-ServerActionsTestOtherA-1520864800-project-member] [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 662.843032] env[62066]: DEBUG nova.network.neutron [None req-4d395966-f335-497d-b8a0-23e22d9b79f7 tempest-ServerActionsTestOtherA-1520864800 tempest-ServerActionsTestOtherA-1520864800-project-member] [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 662.936675] env[62066]: DEBUG nova.policy [None req-4d395966-f335-497d-b8a0-23e22d9b79f7 tempest-ServerActionsTestOtherA-1520864800 tempest-ServerActionsTestOtherA-1520864800-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '146cfa31e9f8481c9a6002a2791bd754', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4f41ce3de03640e49bb23d3678c36a48', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 663.081461] env[62066]: INFO nova.scheduler.client.report [None req-7abf52e5-ee7c-4f92-a757-ffa8ca735a57 tempest-ServerActionsTestJSON-164481335 tempest-ServerActionsTestJSON-164481335-project-member] Deleted allocations for instance bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94 [ 663.315567] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-695ca9fa-c2b9-429f-b803-01e624e0281b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.324036] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38c465b4-4c4c-4e34-88b7-79c3580762de {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.367991] env[62066]: DEBUG nova.compute.manager [None req-4d395966-f335-497d-b8a0-23e22d9b79f7 tempest-ServerActionsTestOtherA-1520864800 tempest-ServerActionsTestOtherA-1520864800-project-member] [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 663.371910] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0846cfcd-9b33-42a5-a019-b1927571ee00 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.382768] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88ba17a9-3d68-4010-8787-55615858a15a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.398338] env[62066]: DEBUG nova.compute.provider_tree [None req-29878964-568e-4c92-80e6-64ee3e8e11bb tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 663.592765] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7abf52e5-ee7c-4f92-a757-ffa8ca735a57 tempest-ServerActionsTestJSON-164481335 tempest-ServerActionsTestJSON-164481335-project-member] Lock "bda8d4eb-2fe2-40ac-9e5f-e7a8f8acdd94" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 131.517s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 663.682092] env[62066]: DEBUG nova.network.neutron [None req-4d395966-f335-497d-b8a0-23e22d9b79f7 tempest-ServerActionsTestOtherA-1520864800 tempest-ServerActionsTestOtherA-1520864800-project-member] [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] Successfully created port: 22ad3b3e-efb9-4f6f-a5a5-e9bcbbbf5ec3 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 663.905018] env[62066]: DEBUG nova.scheduler.client.report [None req-29878964-568e-4c92-80e6-64ee3e8e11bb tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 664.096280] env[62066]: DEBUG nova.compute.manager [None req-19be2ac0-adbe-403f-ab5a-6d1895be1574 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 664.381711] env[62066]: DEBUG nova.compute.manager [None req-4d395966-f335-497d-b8a0-23e22d9b79f7 tempest-ServerActionsTestOtherA-1520864800 tempest-ServerActionsTestOtherA-1520864800-project-member] [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 664.410291] env[62066]: DEBUG oslo_concurrency.lockutils [None req-29878964-568e-4c92-80e6-64ee3e8e11bb tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.077s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 664.410935] env[62066]: ERROR nova.compute.manager [None req-29878964-568e-4c92-80e6-64ee3e8e11bb tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port d26b2533-b09b-4f8f-a9a3-31ed85d82aa4, please check neutron logs for more information. [ 664.410935] env[62066]: ERROR nova.compute.manager [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] Traceback (most recent call last): [ 664.410935] env[62066]: ERROR nova.compute.manager [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 664.410935] env[62066]: ERROR nova.compute.manager [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] self.driver.spawn(context, instance, image_meta, [ 664.410935] env[62066]: ERROR nova.compute.manager [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 664.410935] env[62066]: ERROR nova.compute.manager [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] self._vmops.spawn(context, instance, image_meta, injected_files, [ 664.410935] env[62066]: ERROR nova.compute.manager [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 664.410935] env[62066]: ERROR nova.compute.manager [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] vm_ref = self.build_virtual_machine(instance, [ 664.410935] env[62066]: ERROR nova.compute.manager [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 664.410935] env[62066]: ERROR nova.compute.manager [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] vif_infos = vmwarevif.get_vif_info(self._session, [ 664.410935] env[62066]: ERROR nova.compute.manager [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 664.411258] env[62066]: ERROR nova.compute.manager [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] for vif in network_info: [ 664.411258] env[62066]: ERROR nova.compute.manager [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 664.411258] env[62066]: ERROR nova.compute.manager [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] return self._sync_wrapper(fn, *args, **kwargs) [ 664.411258] env[62066]: ERROR nova.compute.manager [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 664.411258] env[62066]: ERROR nova.compute.manager [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] self.wait() [ 664.411258] env[62066]: ERROR nova.compute.manager [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 664.411258] env[62066]: ERROR nova.compute.manager [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] self[:] = self._gt.wait() [ 664.411258] env[62066]: ERROR nova.compute.manager [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 664.411258] env[62066]: ERROR nova.compute.manager [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] return self._exit_event.wait() [ 664.411258] env[62066]: ERROR nova.compute.manager [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 664.411258] env[62066]: ERROR nova.compute.manager [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] result = hub.switch() [ 664.411258] env[62066]: ERROR nova.compute.manager [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 664.411258] env[62066]: ERROR nova.compute.manager [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] return self.greenlet.switch() [ 664.411576] env[62066]: ERROR nova.compute.manager [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 664.411576] env[62066]: ERROR nova.compute.manager [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] result = function(*args, **kwargs) [ 664.411576] env[62066]: ERROR nova.compute.manager [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 664.411576] env[62066]: ERROR nova.compute.manager [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] return func(*args, **kwargs) [ 664.411576] env[62066]: ERROR nova.compute.manager [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 664.411576] env[62066]: ERROR nova.compute.manager [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] raise e [ 664.411576] env[62066]: ERROR nova.compute.manager [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 664.411576] env[62066]: ERROR nova.compute.manager [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] nwinfo = self.network_api.allocate_for_instance( [ 664.411576] env[62066]: ERROR nova.compute.manager [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 664.411576] env[62066]: ERROR nova.compute.manager [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] created_port_ids = self._update_ports_for_instance( [ 664.411576] env[62066]: ERROR nova.compute.manager [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 664.411576] env[62066]: ERROR nova.compute.manager [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] with excutils.save_and_reraise_exception(): [ 664.411576] env[62066]: ERROR nova.compute.manager [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 664.411901] env[62066]: ERROR nova.compute.manager [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] self.force_reraise() [ 664.411901] env[62066]: ERROR nova.compute.manager [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 664.411901] env[62066]: ERROR nova.compute.manager [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] raise self.value [ 664.411901] env[62066]: ERROR nova.compute.manager [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 664.411901] env[62066]: ERROR nova.compute.manager [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] updated_port = self._update_port( [ 664.411901] env[62066]: ERROR nova.compute.manager [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 664.411901] env[62066]: ERROR nova.compute.manager [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] _ensure_no_port_binding_failure(port) [ 664.411901] env[62066]: ERROR nova.compute.manager [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 664.411901] env[62066]: ERROR nova.compute.manager [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] raise exception.PortBindingFailed(port_id=port['id']) [ 664.411901] env[62066]: ERROR nova.compute.manager [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] nova.exception.PortBindingFailed: Binding failed for port d26b2533-b09b-4f8f-a9a3-31ed85d82aa4, please check neutron logs for more information. [ 664.411901] env[62066]: ERROR nova.compute.manager [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] [ 664.412184] env[62066]: DEBUG nova.compute.utils [None req-29878964-568e-4c92-80e6-64ee3e8e11bb tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] Binding failed for port d26b2533-b09b-4f8f-a9a3-31ed85d82aa4, please check neutron logs for more information. {{(pid=62066) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 664.412946] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cf333fac-d104-4122-9317-eab87c4499f0 tempest-ServerDiagnosticsTest-1737887255 tempest-ServerDiagnosticsTest-1737887255-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 29.362s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 664.414430] env[62066]: INFO nova.compute.claims [None req-cf333fac-d104-4122-9317-eab87c4499f0 tempest-ServerDiagnosticsTest-1737887255 tempest-ServerDiagnosticsTest-1737887255-project-member] [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 664.420769] env[62066]: DEBUG nova.compute.manager [None req-29878964-568e-4c92-80e6-64ee3e8e11bb tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] Build of instance 0e53c28c-a440-4e0e-9a4d-ff1d52557752 was re-scheduled: Binding failed for port d26b2533-b09b-4f8f-a9a3-31ed85d82aa4, please check neutron logs for more information. {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 664.421233] env[62066]: DEBUG nova.compute.manager [None req-29878964-568e-4c92-80e6-64ee3e8e11bb tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] Unplugging VIFs for instance {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 664.421526] env[62066]: DEBUG oslo_concurrency.lockutils [None req-29878964-568e-4c92-80e6-64ee3e8e11bb tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] Acquiring lock "refresh_cache-0e53c28c-a440-4e0e-9a4d-ff1d52557752" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 664.421601] env[62066]: DEBUG oslo_concurrency.lockutils [None req-29878964-568e-4c92-80e6-64ee3e8e11bb tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] Acquired lock "refresh_cache-0e53c28c-a440-4e0e-9a4d-ff1d52557752" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 664.421757] env[62066]: DEBUG nova.network.neutron [None req-29878964-568e-4c92-80e6-64ee3e8e11bb tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 664.428352] env[62066]: DEBUG nova.virt.hardware [None req-4d395966-f335-497d-b8a0-23e22d9b79f7 tempest-ServerActionsTestOtherA-1520864800 tempest-ServerActionsTestOtherA-1520864800-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-23T13:40:41Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-23T13:40:23Z,direct_url=,disk_format='vmdk',id=50ff584c-3b50-4395-af07-3e66769bc9f7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='eb52f7069a374c61ae946f052007c6d9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-23T13:40:24Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 664.428352] env[62066]: DEBUG nova.virt.hardware [None req-4d395966-f335-497d-b8a0-23e22d9b79f7 tempest-ServerActionsTestOtherA-1520864800 tempest-ServerActionsTestOtherA-1520864800-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 664.428352] env[62066]: DEBUG nova.virt.hardware [None req-4d395966-f335-497d-b8a0-23e22d9b79f7 tempest-ServerActionsTestOtherA-1520864800 tempest-ServerActionsTestOtherA-1520864800-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 664.428480] env[62066]: DEBUG nova.virt.hardware [None req-4d395966-f335-497d-b8a0-23e22d9b79f7 tempest-ServerActionsTestOtherA-1520864800 tempest-ServerActionsTestOtherA-1520864800-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 664.428480] env[62066]: DEBUG nova.virt.hardware [None req-4d395966-f335-497d-b8a0-23e22d9b79f7 tempest-ServerActionsTestOtherA-1520864800 tempest-ServerActionsTestOtherA-1520864800-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 664.428480] env[62066]: DEBUG nova.virt.hardware [None req-4d395966-f335-497d-b8a0-23e22d9b79f7 tempest-ServerActionsTestOtherA-1520864800 tempest-ServerActionsTestOtherA-1520864800-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 664.428480] env[62066]: DEBUG nova.virt.hardware [None req-4d395966-f335-497d-b8a0-23e22d9b79f7 tempest-ServerActionsTestOtherA-1520864800 tempest-ServerActionsTestOtherA-1520864800-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 664.428480] env[62066]: DEBUG nova.virt.hardware [None req-4d395966-f335-497d-b8a0-23e22d9b79f7 tempest-ServerActionsTestOtherA-1520864800 tempest-ServerActionsTestOtherA-1520864800-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 664.428665] env[62066]: DEBUG nova.virt.hardware [None req-4d395966-f335-497d-b8a0-23e22d9b79f7 tempest-ServerActionsTestOtherA-1520864800 tempest-ServerActionsTestOtherA-1520864800-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 664.428665] env[62066]: DEBUG nova.virt.hardware [None req-4d395966-f335-497d-b8a0-23e22d9b79f7 tempest-ServerActionsTestOtherA-1520864800 tempest-ServerActionsTestOtherA-1520864800-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 664.428665] env[62066]: DEBUG nova.virt.hardware [None req-4d395966-f335-497d-b8a0-23e22d9b79f7 tempest-ServerActionsTestOtherA-1520864800 tempest-ServerActionsTestOtherA-1520864800-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 664.428665] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ac99d19-cdbd-4658-a7d0-0c58223ecd8a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.440012] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-682570a9-f6a0-480d-91d8-ea713b989a2e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.455152] env[62066]: DEBUG nova.network.neutron [None req-29878964-568e-4c92-80e6-64ee3e8e11bb tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 664.627970] env[62066]: DEBUG oslo_concurrency.lockutils [None req-19be2ac0-adbe-403f-ab5a-6d1895be1574 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 664.690370] env[62066]: DEBUG nova.network.neutron [None req-29878964-568e-4c92-80e6-64ee3e8e11bb tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 665.195389] env[62066]: DEBUG oslo_concurrency.lockutils [None req-29878964-568e-4c92-80e6-64ee3e8e11bb tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] Releasing lock "refresh_cache-0e53c28c-a440-4e0e-9a4d-ff1d52557752" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 665.195389] env[62066]: DEBUG nova.compute.manager [None req-29878964-568e-4c92-80e6-64ee3e8e11bb tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 665.195389] env[62066]: DEBUG nova.compute.manager [None req-29878964-568e-4c92-80e6-64ee3e8e11bb tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 665.195389] env[62066]: DEBUG nova.network.neutron [None req-29878964-568e-4c92-80e6-64ee3e8e11bb tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 665.217400] env[62066]: DEBUG nova.network.neutron [None req-29878964-568e-4c92-80e6-64ee3e8e11bb tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 665.386555] env[62066]: DEBUG nova.compute.manager [req-a0ec73de-c751-4dd4-a1ce-9e98922f25ec req-2b4c7d36-9cb8-4b33-837d-d7a793347e85 service nova] [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] Received event network-changed-22ad3b3e-efb9-4f6f-a5a5-e9bcbbbf5ec3 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 665.386796] env[62066]: DEBUG nova.compute.manager [req-a0ec73de-c751-4dd4-a1ce-9e98922f25ec req-2b4c7d36-9cb8-4b33-837d-d7a793347e85 service nova] [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] Refreshing instance network info cache due to event network-changed-22ad3b3e-efb9-4f6f-a5a5-e9bcbbbf5ec3. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 665.387025] env[62066]: DEBUG oslo_concurrency.lockutils [req-a0ec73de-c751-4dd4-a1ce-9e98922f25ec req-2b4c7d36-9cb8-4b33-837d-d7a793347e85 service nova] Acquiring lock "refresh_cache-56984993-a8b1-464e-b20c-79ffe67d2491" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 665.387191] env[62066]: DEBUG oslo_concurrency.lockutils [req-a0ec73de-c751-4dd4-a1ce-9e98922f25ec req-2b4c7d36-9cb8-4b33-837d-d7a793347e85 service nova] Acquired lock "refresh_cache-56984993-a8b1-464e-b20c-79ffe67d2491" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 665.387354] env[62066]: DEBUG nova.network.neutron [req-a0ec73de-c751-4dd4-a1ce-9e98922f25ec req-2b4c7d36-9cb8-4b33-837d-d7a793347e85 service nova] [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] Refreshing network info cache for port 22ad3b3e-efb9-4f6f-a5a5-e9bcbbbf5ec3 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 665.561358] env[62066]: ERROR nova.compute.manager [None req-4d395966-f335-497d-b8a0-23e22d9b79f7 tempest-ServerActionsTestOtherA-1520864800 tempest-ServerActionsTestOtherA-1520864800-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 22ad3b3e-efb9-4f6f-a5a5-e9bcbbbf5ec3, please check neutron logs for more information. [ 665.561358] env[62066]: ERROR nova.compute.manager Traceback (most recent call last): [ 665.561358] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 665.561358] env[62066]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 665.561358] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 665.561358] env[62066]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 665.561358] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 665.561358] env[62066]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 665.561358] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 665.561358] env[62066]: ERROR nova.compute.manager self.force_reraise() [ 665.561358] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 665.561358] env[62066]: ERROR nova.compute.manager raise self.value [ 665.561358] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 665.561358] env[62066]: ERROR nova.compute.manager updated_port = self._update_port( [ 665.561358] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 665.561358] env[62066]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 665.562530] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 665.562530] env[62066]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 665.562530] env[62066]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 22ad3b3e-efb9-4f6f-a5a5-e9bcbbbf5ec3, please check neutron logs for more information. [ 665.562530] env[62066]: ERROR nova.compute.manager [ 665.562530] env[62066]: Traceback (most recent call last): [ 665.562530] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 665.562530] env[62066]: listener.cb(fileno) [ 665.562530] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 665.562530] env[62066]: result = function(*args, **kwargs) [ 665.562530] env[62066]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 665.562530] env[62066]: return func(*args, **kwargs) [ 665.562530] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 665.562530] env[62066]: raise e [ 665.562530] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 665.562530] env[62066]: nwinfo = self.network_api.allocate_for_instance( [ 665.562530] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 665.562530] env[62066]: created_port_ids = self._update_ports_for_instance( [ 665.562530] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 665.562530] env[62066]: with excutils.save_and_reraise_exception(): [ 665.562530] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 665.562530] env[62066]: self.force_reraise() [ 665.562530] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 665.562530] env[62066]: raise self.value [ 665.562530] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 665.562530] env[62066]: updated_port = self._update_port( [ 665.562530] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 665.562530] env[62066]: _ensure_no_port_binding_failure(port) [ 665.562530] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 665.562530] env[62066]: raise exception.PortBindingFailed(port_id=port['id']) [ 665.563742] env[62066]: nova.exception.PortBindingFailed: Binding failed for port 22ad3b3e-efb9-4f6f-a5a5-e9bcbbbf5ec3, please check neutron logs for more information. [ 665.563742] env[62066]: Removing descriptor: 20 [ 665.563742] env[62066]: ERROR nova.compute.manager [None req-4d395966-f335-497d-b8a0-23e22d9b79f7 tempest-ServerActionsTestOtherA-1520864800 tempest-ServerActionsTestOtherA-1520864800-project-member] [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 22ad3b3e-efb9-4f6f-a5a5-e9bcbbbf5ec3, please check neutron logs for more information. [ 665.563742] env[62066]: ERROR nova.compute.manager [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] Traceback (most recent call last): [ 665.563742] env[62066]: ERROR nova.compute.manager [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 665.563742] env[62066]: ERROR nova.compute.manager [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] yield resources [ 665.563742] env[62066]: ERROR nova.compute.manager [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 665.563742] env[62066]: ERROR nova.compute.manager [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] self.driver.spawn(context, instance, image_meta, [ 665.563742] env[62066]: ERROR nova.compute.manager [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 665.563742] env[62066]: ERROR nova.compute.manager [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] self._vmops.spawn(context, instance, image_meta, injected_files, [ 665.563742] env[62066]: ERROR nova.compute.manager [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 665.563742] env[62066]: ERROR nova.compute.manager [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] vm_ref = self.build_virtual_machine(instance, [ 665.564266] env[62066]: ERROR nova.compute.manager [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 665.564266] env[62066]: ERROR nova.compute.manager [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] vif_infos = vmwarevif.get_vif_info(self._session, [ 665.564266] env[62066]: ERROR nova.compute.manager [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 665.564266] env[62066]: ERROR nova.compute.manager [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] for vif in network_info: [ 665.564266] env[62066]: ERROR nova.compute.manager [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 665.564266] env[62066]: ERROR nova.compute.manager [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] return self._sync_wrapper(fn, *args, **kwargs) [ 665.564266] env[62066]: ERROR nova.compute.manager [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 665.564266] env[62066]: ERROR nova.compute.manager [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] self.wait() [ 665.564266] env[62066]: ERROR nova.compute.manager [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 665.564266] env[62066]: ERROR nova.compute.manager [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] self[:] = self._gt.wait() [ 665.564266] env[62066]: ERROR nova.compute.manager [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 665.564266] env[62066]: ERROR nova.compute.manager [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] return self._exit_event.wait() [ 665.564266] env[62066]: ERROR nova.compute.manager [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 665.564833] env[62066]: ERROR nova.compute.manager [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] result = hub.switch() [ 665.564833] env[62066]: ERROR nova.compute.manager [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 665.564833] env[62066]: ERROR nova.compute.manager [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] return self.greenlet.switch() [ 665.564833] env[62066]: ERROR nova.compute.manager [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 665.564833] env[62066]: ERROR nova.compute.manager [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] result = function(*args, **kwargs) [ 665.564833] env[62066]: ERROR nova.compute.manager [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 665.564833] env[62066]: ERROR nova.compute.manager [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] return func(*args, **kwargs) [ 665.564833] env[62066]: ERROR nova.compute.manager [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 665.564833] env[62066]: ERROR nova.compute.manager [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] raise e [ 665.564833] env[62066]: ERROR nova.compute.manager [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 665.564833] env[62066]: ERROR nova.compute.manager [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] nwinfo = self.network_api.allocate_for_instance( [ 665.564833] env[62066]: ERROR nova.compute.manager [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 665.564833] env[62066]: ERROR nova.compute.manager [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] created_port_ids = self._update_ports_for_instance( [ 665.565464] env[62066]: ERROR nova.compute.manager [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 665.565464] env[62066]: ERROR nova.compute.manager [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] with excutils.save_and_reraise_exception(): [ 665.565464] env[62066]: ERROR nova.compute.manager [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 665.565464] env[62066]: ERROR nova.compute.manager [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] self.force_reraise() [ 665.565464] env[62066]: ERROR nova.compute.manager [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 665.565464] env[62066]: ERROR nova.compute.manager [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] raise self.value [ 665.565464] env[62066]: ERROR nova.compute.manager [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 665.565464] env[62066]: ERROR nova.compute.manager [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] updated_port = self._update_port( [ 665.565464] env[62066]: ERROR nova.compute.manager [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 665.565464] env[62066]: ERROR nova.compute.manager [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] _ensure_no_port_binding_failure(port) [ 665.565464] env[62066]: ERROR nova.compute.manager [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 665.565464] env[62066]: ERROR nova.compute.manager [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] raise exception.PortBindingFailed(port_id=port['id']) [ 665.565983] env[62066]: ERROR nova.compute.manager [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] nova.exception.PortBindingFailed: Binding failed for port 22ad3b3e-efb9-4f6f-a5a5-e9bcbbbf5ec3, please check neutron logs for more information. [ 665.565983] env[62066]: ERROR nova.compute.manager [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] [ 665.565983] env[62066]: INFO nova.compute.manager [None req-4d395966-f335-497d-b8a0-23e22d9b79f7 tempest-ServerActionsTestOtherA-1520864800 tempest-ServerActionsTestOtherA-1520864800-project-member] [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] Terminating instance [ 665.565983] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4d395966-f335-497d-b8a0-23e22d9b79f7 tempest-ServerActionsTestOtherA-1520864800 tempest-ServerActionsTestOtherA-1520864800-project-member] Acquiring lock "refresh_cache-56984993-a8b1-464e-b20c-79ffe67d2491" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 665.720243] env[62066]: DEBUG nova.network.neutron [None req-29878964-568e-4c92-80e6-64ee3e8e11bb tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 665.839701] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9fb456e-47ef-48d6-9837-fcd78cfb84bb {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.851135] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bbc2064-84f0-452f-a273-e853a7c773e8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.886067] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-798bc986-3fcc-4dd5-94d4-8366235a4f49 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.896601] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-922bd963-a356-4ee6-afe1-f3b7c4912815 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.912341] env[62066]: DEBUG nova.compute.provider_tree [None req-cf333fac-d104-4122-9317-eab87c4499f0 tempest-ServerDiagnosticsTest-1737887255 tempest-ServerDiagnosticsTest-1737887255-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 665.914305] env[62066]: DEBUG nova.network.neutron [req-a0ec73de-c751-4dd4-a1ce-9e98922f25ec req-2b4c7d36-9cb8-4b33-837d-d7a793347e85 service nova] [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 666.032413] env[62066]: DEBUG nova.network.neutron [req-a0ec73de-c751-4dd4-a1ce-9e98922f25ec req-2b4c7d36-9cb8-4b33-837d-d7a793347e85 service nova] [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 666.226691] env[62066]: INFO nova.compute.manager [None req-29878964-568e-4c92-80e6-64ee3e8e11bb tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] [instance: 0e53c28c-a440-4e0e-9a4d-ff1d52557752] Took 1.03 seconds to deallocate network for instance. [ 666.418587] env[62066]: DEBUG nova.scheduler.client.report [None req-cf333fac-d104-4122-9317-eab87c4499f0 tempest-ServerDiagnosticsTest-1737887255 tempest-ServerDiagnosticsTest-1737887255-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 666.537080] env[62066]: DEBUG oslo_concurrency.lockutils [req-a0ec73de-c751-4dd4-a1ce-9e98922f25ec req-2b4c7d36-9cb8-4b33-837d-d7a793347e85 service nova] Releasing lock "refresh_cache-56984993-a8b1-464e-b20c-79ffe67d2491" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 666.537542] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4d395966-f335-497d-b8a0-23e22d9b79f7 tempest-ServerActionsTestOtherA-1520864800 tempest-ServerActionsTestOtherA-1520864800-project-member] Acquired lock "refresh_cache-56984993-a8b1-464e-b20c-79ffe67d2491" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 666.537724] env[62066]: DEBUG nova.network.neutron [None req-4d395966-f335-497d-b8a0-23e22d9b79f7 tempest-ServerActionsTestOtherA-1520864800 tempest-ServerActionsTestOtherA-1520864800-project-member] [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 666.926079] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cf333fac-d104-4122-9317-eab87c4499f0 tempest-ServerDiagnosticsTest-1737887255 tempest-ServerDiagnosticsTest-1737887255-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.513s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 666.927127] env[62066]: DEBUG nova.compute.manager [None req-cf333fac-d104-4122-9317-eab87c4499f0 tempest-ServerDiagnosticsTest-1737887255 tempest-ServerDiagnosticsTest-1737887255-project-member] [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 666.929335] env[62066]: DEBUG oslo_concurrency.lockutils [None req-67a0ccea-4cde-4de9-9aa4-1bb67e83772e tempest-ServersNegativeTestJSON-789906455 tempest-ServersNegativeTestJSON-789906455-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.409s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 666.930820] env[62066]: INFO nova.compute.claims [None req-67a0ccea-4cde-4de9-9aa4-1bb67e83772e tempest-ServersNegativeTestJSON-789906455 tempest-ServersNegativeTestJSON-789906455-project-member] [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 667.067558] env[62066]: DEBUG nova.network.neutron [None req-4d395966-f335-497d-b8a0-23e22d9b79f7 tempest-ServerActionsTestOtherA-1520864800 tempest-ServerActionsTestOtherA-1520864800-project-member] [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 667.228188] env[62066]: DEBUG nova.network.neutron [None req-4d395966-f335-497d-b8a0-23e22d9b79f7 tempest-ServerActionsTestOtherA-1520864800 tempest-ServerActionsTestOtherA-1520864800-project-member] [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 667.303191] env[62066]: INFO nova.scheduler.client.report [None req-29878964-568e-4c92-80e6-64ee3e8e11bb tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] Deleted allocations for instance 0e53c28c-a440-4e0e-9a4d-ff1d52557752 [ 667.435381] env[62066]: DEBUG nova.compute.utils [None req-cf333fac-d104-4122-9317-eab87c4499f0 tempest-ServerDiagnosticsTest-1737887255 tempest-ServerDiagnosticsTest-1737887255-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 667.437108] env[62066]: DEBUG nova.compute.manager [None req-cf333fac-d104-4122-9317-eab87c4499f0 tempest-ServerDiagnosticsTest-1737887255 tempest-ServerDiagnosticsTest-1737887255-project-member] [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 667.439934] env[62066]: DEBUG nova.network.neutron [None req-cf333fac-d104-4122-9317-eab87c4499f0 tempest-ServerDiagnosticsTest-1737887255 tempest-ServerDiagnosticsTest-1737887255-project-member] [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 667.550735] env[62066]: DEBUG nova.policy [None req-cf333fac-d104-4122-9317-eab87c4499f0 tempest-ServerDiagnosticsTest-1737887255 tempest-ServerDiagnosticsTest-1737887255-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '45c26bf770eb4baa99d61e7d522c92a8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '268920dfe77c40daa178c7b9560d9089', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 667.657703] env[62066]: DEBUG nova.compute.manager [req-a103de9e-2e7a-4f4f-a614-fb0dd6917d6f req-07736306-4431-4d05-8565-3902db08babb service nova] [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] Received event network-vif-deleted-22ad3b3e-efb9-4f6f-a5a5-e9bcbbbf5ec3 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 667.731419] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4d395966-f335-497d-b8a0-23e22d9b79f7 tempest-ServerActionsTestOtherA-1520864800 tempest-ServerActionsTestOtherA-1520864800-project-member] Releasing lock "refresh_cache-56984993-a8b1-464e-b20c-79ffe67d2491" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 667.731855] env[62066]: DEBUG nova.compute.manager [None req-4d395966-f335-497d-b8a0-23e22d9b79f7 tempest-ServerActionsTestOtherA-1520864800 tempest-ServerActionsTestOtherA-1520864800-project-member] [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 667.732063] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-4d395966-f335-497d-b8a0-23e22d9b79f7 tempest-ServerActionsTestOtherA-1520864800 tempest-ServerActionsTestOtherA-1520864800-project-member] [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 667.732756] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-43ddc389-5914-4537-9f6c-32dbe4bda38c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.743576] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0a4da3e-e1dd-4f77-a3c2-a5e27c5c34c7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.765501] env[62066]: WARNING nova.virt.vmwareapi.vmops [None req-4d395966-f335-497d-b8a0-23e22d9b79f7 tempest-ServerActionsTestOtherA-1520864800 tempest-ServerActionsTestOtherA-1520864800-project-member] [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 56984993-a8b1-464e-b20c-79ffe67d2491 could not be found. [ 667.765821] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-4d395966-f335-497d-b8a0-23e22d9b79f7 tempest-ServerActionsTestOtherA-1520864800 tempest-ServerActionsTestOtherA-1520864800-project-member] [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 667.766062] env[62066]: INFO nova.compute.manager [None req-4d395966-f335-497d-b8a0-23e22d9b79f7 tempest-ServerActionsTestOtherA-1520864800 tempest-ServerActionsTestOtherA-1520864800-project-member] [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] Took 0.03 seconds to destroy the instance on the hypervisor. [ 667.766391] env[62066]: DEBUG oslo.service.loopingcall [None req-4d395966-f335-497d-b8a0-23e22d9b79f7 tempest-ServerActionsTestOtherA-1520864800 tempest-ServerActionsTestOtherA-1520864800-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 667.766625] env[62066]: DEBUG nova.compute.manager [-] [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 667.766730] env[62066]: DEBUG nova.network.neutron [-] [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 667.783187] env[62066]: DEBUG nova.network.neutron [-] [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 667.812993] env[62066]: DEBUG oslo_concurrency.lockutils [None req-29878964-568e-4c92-80e6-64ee3e8e11bb tempest-DeleteServersAdminTestJSON-2048021236 tempest-DeleteServersAdminTestJSON-2048021236-project-member] Lock "0e53c28c-a440-4e0e-9a4d-ff1d52557752" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 135.231s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 667.942563] env[62066]: DEBUG nova.compute.manager [None req-cf333fac-d104-4122-9317-eab87c4499f0 tempest-ServerDiagnosticsTest-1737887255 tempest-ServerDiagnosticsTest-1737887255-project-member] [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 668.285400] env[62066]: DEBUG nova.network.neutron [-] [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 668.307798] env[62066]: DEBUG nova.network.neutron [None req-cf333fac-d104-4122-9317-eab87c4499f0 tempest-ServerDiagnosticsTest-1737887255 tempest-ServerDiagnosticsTest-1737887255-project-member] [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] Successfully created port: 70d1fb9d-0a6e-46b4-a478-297d09fbe121 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 668.314786] env[62066]: DEBUG nova.compute.manager [None req-96f72479-6fed-4ff5-9d17-40fa20c1f416 tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 668.470386] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46ba6f22-a703-4f4d-8de3-c4aeffaf812d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.477294] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17dd8fae-27e9-453c-a1b5-15c49a75766d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.516194] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-454c77d7-c397-40b9-80b3-e57b3cb3f740 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.523725] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bddc5b5b-77fc-4413-a294-8152a31b0d08 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.540798] env[62066]: DEBUG nova.compute.provider_tree [None req-67a0ccea-4cde-4de9-9aa4-1bb67e83772e tempest-ServersNegativeTestJSON-789906455 tempest-ServersNegativeTestJSON-789906455-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 668.789059] env[62066]: INFO nova.compute.manager [-] [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] Took 1.02 seconds to deallocate network for instance. [ 668.792992] env[62066]: DEBUG nova.compute.claims [None req-4d395966-f335-497d-b8a0-23e22d9b79f7 tempest-ServerActionsTestOtherA-1520864800 tempest-ServerActionsTestOtherA-1520864800-project-member] [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] Aborting claim: {{(pid=62066) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 668.793701] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4d395966-f335-497d-b8a0-23e22d9b79f7 tempest-ServerActionsTestOtherA-1520864800 tempest-ServerActionsTestOtherA-1520864800-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 668.838737] env[62066]: DEBUG oslo_concurrency.lockutils [None req-96f72479-6fed-4ff5-9d17-40fa20c1f416 tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 668.956555] env[62066]: DEBUG nova.compute.manager [None req-cf333fac-d104-4122-9317-eab87c4499f0 tempest-ServerDiagnosticsTest-1737887255 tempest-ServerDiagnosticsTest-1737887255-project-member] [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 668.981957] env[62066]: DEBUG nova.virt.hardware [None req-cf333fac-d104-4122-9317-eab87c4499f0 tempest-ServerDiagnosticsTest-1737887255 tempest-ServerDiagnosticsTest-1737887255-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-23T13:40:41Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-23T13:40:23Z,direct_url=,disk_format='vmdk',id=50ff584c-3b50-4395-af07-3e66769bc9f7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='eb52f7069a374c61ae946f052007c6d9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-23T13:40:24Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 668.982390] env[62066]: DEBUG nova.virt.hardware [None req-cf333fac-d104-4122-9317-eab87c4499f0 tempest-ServerDiagnosticsTest-1737887255 tempest-ServerDiagnosticsTest-1737887255-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 668.982555] env[62066]: DEBUG nova.virt.hardware [None req-cf333fac-d104-4122-9317-eab87c4499f0 tempest-ServerDiagnosticsTest-1737887255 tempest-ServerDiagnosticsTest-1737887255-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 668.982878] env[62066]: DEBUG nova.virt.hardware [None req-cf333fac-d104-4122-9317-eab87c4499f0 tempest-ServerDiagnosticsTest-1737887255 tempest-ServerDiagnosticsTest-1737887255-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 668.982878] env[62066]: DEBUG nova.virt.hardware [None req-cf333fac-d104-4122-9317-eab87c4499f0 tempest-ServerDiagnosticsTest-1737887255 tempest-ServerDiagnosticsTest-1737887255-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 668.983235] env[62066]: DEBUG nova.virt.hardware [None req-cf333fac-d104-4122-9317-eab87c4499f0 tempest-ServerDiagnosticsTest-1737887255 tempest-ServerDiagnosticsTest-1737887255-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 668.983473] env[62066]: DEBUG nova.virt.hardware [None req-cf333fac-d104-4122-9317-eab87c4499f0 tempest-ServerDiagnosticsTest-1737887255 tempest-ServerDiagnosticsTest-1737887255-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 668.983654] env[62066]: DEBUG nova.virt.hardware [None req-cf333fac-d104-4122-9317-eab87c4499f0 tempest-ServerDiagnosticsTest-1737887255 tempest-ServerDiagnosticsTest-1737887255-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 668.983804] env[62066]: DEBUG nova.virt.hardware [None req-cf333fac-d104-4122-9317-eab87c4499f0 tempest-ServerDiagnosticsTest-1737887255 tempest-ServerDiagnosticsTest-1737887255-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 668.983966] env[62066]: DEBUG nova.virt.hardware [None req-cf333fac-d104-4122-9317-eab87c4499f0 tempest-ServerDiagnosticsTest-1737887255 tempest-ServerDiagnosticsTest-1737887255-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 668.984154] env[62066]: DEBUG nova.virt.hardware [None req-cf333fac-d104-4122-9317-eab87c4499f0 tempest-ServerDiagnosticsTest-1737887255 tempest-ServerDiagnosticsTest-1737887255-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 668.985276] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-455916b8-3f1d-4bef-80d2-e7e531f13b13 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.993429] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7406b6b6-8000-4972-8f02-3ddd23f7edb6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.043859] env[62066]: DEBUG nova.scheduler.client.report [None req-67a0ccea-4cde-4de9-9aa4-1bb67e83772e tempest-ServersNegativeTestJSON-789906455 tempest-ServersNegativeTestJSON-789906455-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 669.552637] env[62066]: DEBUG oslo_concurrency.lockutils [None req-67a0ccea-4cde-4de9-9aa4-1bb67e83772e tempest-ServersNegativeTestJSON-789906455 tempest-ServersNegativeTestJSON-789906455-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.623s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 669.553919] env[62066]: DEBUG nova.compute.manager [None req-67a0ccea-4cde-4de9-9aa4-1bb67e83772e tempest-ServersNegativeTestJSON-789906455 tempest-ServersNegativeTestJSON-789906455-project-member] [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 669.557771] env[62066]: DEBUG oslo_concurrency.lockutils [None req-2c107b45-3cbf-4be9-af5a-796fbdb9b5fe tempest-ImagesNegativeTestJSON-881899573 tempest-ImagesNegativeTestJSON-881899573-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 27.287s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 669.731147] env[62066]: DEBUG nova.compute.manager [req-8183304d-f97e-46af-b30b-7e0a91cf5391 req-28cb3e63-4f5b-4e58-b354-874f70e62889 service nova] [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] Received event network-changed-70d1fb9d-0a6e-46b4-a478-297d09fbe121 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 669.731147] env[62066]: DEBUG nova.compute.manager [req-8183304d-f97e-46af-b30b-7e0a91cf5391 req-28cb3e63-4f5b-4e58-b354-874f70e62889 service nova] [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] Refreshing instance network info cache due to event network-changed-70d1fb9d-0a6e-46b4-a478-297d09fbe121. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 669.731147] env[62066]: DEBUG oslo_concurrency.lockutils [req-8183304d-f97e-46af-b30b-7e0a91cf5391 req-28cb3e63-4f5b-4e58-b354-874f70e62889 service nova] Acquiring lock "refresh_cache-2cf18a04-6bf9-4f47-a920-2c568207057d" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 669.731147] env[62066]: DEBUG oslo_concurrency.lockutils [req-8183304d-f97e-46af-b30b-7e0a91cf5391 req-28cb3e63-4f5b-4e58-b354-874f70e62889 service nova] Acquired lock "refresh_cache-2cf18a04-6bf9-4f47-a920-2c568207057d" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 669.731147] env[62066]: DEBUG nova.network.neutron [req-8183304d-f97e-46af-b30b-7e0a91cf5391 req-28cb3e63-4f5b-4e58-b354-874f70e62889 service nova] [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] Refreshing network info cache for port 70d1fb9d-0a6e-46b4-a478-297d09fbe121 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 669.978593] env[62066]: ERROR nova.compute.manager [None req-cf333fac-d104-4122-9317-eab87c4499f0 tempest-ServerDiagnosticsTest-1737887255 tempest-ServerDiagnosticsTest-1737887255-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 70d1fb9d-0a6e-46b4-a478-297d09fbe121, please check neutron logs for more information. [ 669.978593] env[62066]: ERROR nova.compute.manager Traceback (most recent call last): [ 669.978593] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 669.978593] env[62066]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 669.978593] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 669.978593] env[62066]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 669.978593] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 669.978593] env[62066]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 669.978593] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 669.978593] env[62066]: ERROR nova.compute.manager self.force_reraise() [ 669.978593] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 669.978593] env[62066]: ERROR nova.compute.manager raise self.value [ 669.978593] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 669.978593] env[62066]: ERROR nova.compute.manager updated_port = self._update_port( [ 669.978593] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 669.978593] env[62066]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 669.979947] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 669.979947] env[62066]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 669.979947] env[62066]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 70d1fb9d-0a6e-46b4-a478-297d09fbe121, please check neutron logs for more information. [ 669.979947] env[62066]: ERROR nova.compute.manager [ 669.979947] env[62066]: Traceback (most recent call last): [ 669.979947] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 669.979947] env[62066]: listener.cb(fileno) [ 669.979947] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 669.979947] env[62066]: result = function(*args, **kwargs) [ 669.979947] env[62066]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 669.979947] env[62066]: return func(*args, **kwargs) [ 669.979947] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 669.979947] env[62066]: raise e [ 669.979947] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 669.979947] env[62066]: nwinfo = self.network_api.allocate_for_instance( [ 669.979947] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 669.979947] env[62066]: created_port_ids = self._update_ports_for_instance( [ 669.979947] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 669.979947] env[62066]: with excutils.save_and_reraise_exception(): [ 669.979947] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 669.979947] env[62066]: self.force_reraise() [ 669.979947] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 669.979947] env[62066]: raise self.value [ 669.979947] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 669.979947] env[62066]: updated_port = self._update_port( [ 669.979947] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 669.979947] env[62066]: _ensure_no_port_binding_failure(port) [ 669.979947] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 669.979947] env[62066]: raise exception.PortBindingFailed(port_id=port['id']) [ 669.980743] env[62066]: nova.exception.PortBindingFailed: Binding failed for port 70d1fb9d-0a6e-46b4-a478-297d09fbe121, please check neutron logs for more information. [ 669.980743] env[62066]: Removing descriptor: 20 [ 669.980743] env[62066]: ERROR nova.compute.manager [None req-cf333fac-d104-4122-9317-eab87c4499f0 tempest-ServerDiagnosticsTest-1737887255 tempest-ServerDiagnosticsTest-1737887255-project-member] [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 70d1fb9d-0a6e-46b4-a478-297d09fbe121, please check neutron logs for more information. [ 669.980743] env[62066]: ERROR nova.compute.manager [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] Traceback (most recent call last): [ 669.980743] env[62066]: ERROR nova.compute.manager [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 669.980743] env[62066]: ERROR nova.compute.manager [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] yield resources [ 669.980743] env[62066]: ERROR nova.compute.manager [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 669.980743] env[62066]: ERROR nova.compute.manager [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] self.driver.spawn(context, instance, image_meta, [ 669.980743] env[62066]: ERROR nova.compute.manager [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 669.980743] env[62066]: ERROR nova.compute.manager [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 669.980743] env[62066]: ERROR nova.compute.manager [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 669.980743] env[62066]: ERROR nova.compute.manager [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] vm_ref = self.build_virtual_machine(instance, [ 669.981092] env[62066]: ERROR nova.compute.manager [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 669.981092] env[62066]: ERROR nova.compute.manager [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] vif_infos = vmwarevif.get_vif_info(self._session, [ 669.981092] env[62066]: ERROR nova.compute.manager [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 669.981092] env[62066]: ERROR nova.compute.manager [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] for vif in network_info: [ 669.981092] env[62066]: ERROR nova.compute.manager [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 669.981092] env[62066]: ERROR nova.compute.manager [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] return self._sync_wrapper(fn, *args, **kwargs) [ 669.981092] env[62066]: ERROR nova.compute.manager [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 669.981092] env[62066]: ERROR nova.compute.manager [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] self.wait() [ 669.981092] env[62066]: ERROR nova.compute.manager [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 669.981092] env[62066]: ERROR nova.compute.manager [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] self[:] = self._gt.wait() [ 669.981092] env[62066]: ERROR nova.compute.manager [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 669.981092] env[62066]: ERROR nova.compute.manager [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] return self._exit_event.wait() [ 669.981092] env[62066]: ERROR nova.compute.manager [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 669.981560] env[62066]: ERROR nova.compute.manager [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] result = hub.switch() [ 669.981560] env[62066]: ERROR nova.compute.manager [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 669.981560] env[62066]: ERROR nova.compute.manager [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] return self.greenlet.switch() [ 669.981560] env[62066]: ERROR nova.compute.manager [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 669.981560] env[62066]: ERROR nova.compute.manager [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] result = function(*args, **kwargs) [ 669.981560] env[62066]: ERROR nova.compute.manager [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 669.981560] env[62066]: ERROR nova.compute.manager [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] return func(*args, **kwargs) [ 669.981560] env[62066]: ERROR nova.compute.manager [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 669.981560] env[62066]: ERROR nova.compute.manager [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] raise e [ 669.981560] env[62066]: ERROR nova.compute.manager [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 669.981560] env[62066]: ERROR nova.compute.manager [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] nwinfo = self.network_api.allocate_for_instance( [ 669.981560] env[62066]: ERROR nova.compute.manager [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 669.981560] env[62066]: ERROR nova.compute.manager [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] created_port_ids = self._update_ports_for_instance( [ 669.981906] env[62066]: ERROR nova.compute.manager [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 669.981906] env[62066]: ERROR nova.compute.manager [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] with excutils.save_and_reraise_exception(): [ 669.981906] env[62066]: ERROR nova.compute.manager [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 669.981906] env[62066]: ERROR nova.compute.manager [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] self.force_reraise() [ 669.981906] env[62066]: ERROR nova.compute.manager [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 669.981906] env[62066]: ERROR nova.compute.manager [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] raise self.value [ 669.981906] env[62066]: ERROR nova.compute.manager [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 669.981906] env[62066]: ERROR nova.compute.manager [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] updated_port = self._update_port( [ 669.981906] env[62066]: ERROR nova.compute.manager [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 669.981906] env[62066]: ERROR nova.compute.manager [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] _ensure_no_port_binding_failure(port) [ 669.981906] env[62066]: ERROR nova.compute.manager [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 669.981906] env[62066]: ERROR nova.compute.manager [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] raise exception.PortBindingFailed(port_id=port['id']) [ 669.982329] env[62066]: ERROR nova.compute.manager [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] nova.exception.PortBindingFailed: Binding failed for port 70d1fb9d-0a6e-46b4-a478-297d09fbe121, please check neutron logs for more information. [ 669.982329] env[62066]: ERROR nova.compute.manager [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] [ 669.982329] env[62066]: INFO nova.compute.manager [None req-cf333fac-d104-4122-9317-eab87c4499f0 tempest-ServerDiagnosticsTest-1737887255 tempest-ServerDiagnosticsTest-1737887255-project-member] [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] Terminating instance [ 669.982329] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cf333fac-d104-4122-9317-eab87c4499f0 tempest-ServerDiagnosticsTest-1737887255 tempest-ServerDiagnosticsTest-1737887255-project-member] Acquiring lock "refresh_cache-2cf18a04-6bf9-4f47-a920-2c568207057d" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 670.062525] env[62066]: DEBUG nova.compute.utils [None req-67a0ccea-4cde-4de9-9aa4-1bb67e83772e tempest-ServersNegativeTestJSON-789906455 tempest-ServersNegativeTestJSON-789906455-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 670.068720] env[62066]: DEBUG nova.compute.manager [None req-67a0ccea-4cde-4de9-9aa4-1bb67e83772e tempest-ServersNegativeTestJSON-789906455 tempest-ServersNegativeTestJSON-789906455-project-member] [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 670.068903] env[62066]: DEBUG nova.network.neutron [None req-67a0ccea-4cde-4de9-9aa4-1bb67e83772e tempest-ServersNegativeTestJSON-789906455 tempest-ServersNegativeTestJSON-789906455-project-member] [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 670.156301] env[62066]: DEBUG nova.policy [None req-67a0ccea-4cde-4de9-9aa4-1bb67e83772e tempest-ServersNegativeTestJSON-789906455 tempest-ServersNegativeTestJSON-789906455-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0916e3d886714bd8b10525f8e3999a67', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'af608eea2b6a4a119c46174a936e5dd1', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 670.255032] env[62066]: DEBUG nova.network.neutron [req-8183304d-f97e-46af-b30b-7e0a91cf5391 req-28cb3e63-4f5b-4e58-b354-874f70e62889 service nova] [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 670.331088] env[62066]: DEBUG nova.network.neutron [req-8183304d-f97e-46af-b30b-7e0a91cf5391 req-28cb3e63-4f5b-4e58-b354-874f70e62889 service nova] [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 670.534903] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f35d877e-b0f3-4961-b448-371244a0e3c3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.544136] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86ab1884-9e0e-42d2-b28d-8b29e75c9d77 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.574169] env[62066]: DEBUG nova.compute.manager [None req-67a0ccea-4cde-4de9-9aa4-1bb67e83772e tempest-ServersNegativeTestJSON-789906455 tempest-ServersNegativeTestJSON-789906455-project-member] [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 670.577356] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ea25a56-4dcb-4c97-8f3e-4a6478c877cc {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.584847] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8081712a-2f80-42d1-8003-a303bd331996 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.599514] env[62066]: DEBUG nova.compute.provider_tree [None req-2c107b45-3cbf-4be9-af5a-796fbdb9b5fe tempest-ImagesNegativeTestJSON-881899573 tempest-ImagesNegativeTestJSON-881899573-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 670.619898] env[62066]: DEBUG nova.network.neutron [None req-67a0ccea-4cde-4de9-9aa4-1bb67e83772e tempest-ServersNegativeTestJSON-789906455 tempest-ServersNegativeTestJSON-789906455-project-member] [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] Successfully created port: ea8fd014-c428-41e1-a6bb-7aba32299177 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 670.683960] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c70bcb7f-78ff-424d-8121-684c45ecd12b tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Acquiring lock "02fb3e92-5dd6-4b1e-a6e4-d60d3fc85b07" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 670.685045] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c70bcb7f-78ff-424d-8121-684c45ecd12b tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Lock "02fb3e92-5dd6-4b1e-a6e4-d60d3fc85b07" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 670.834189] env[62066]: DEBUG oslo_concurrency.lockutils [req-8183304d-f97e-46af-b30b-7e0a91cf5391 req-28cb3e63-4f5b-4e58-b354-874f70e62889 service nova] Releasing lock "refresh_cache-2cf18a04-6bf9-4f47-a920-2c568207057d" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 670.835679] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cf333fac-d104-4122-9317-eab87c4499f0 tempest-ServerDiagnosticsTest-1737887255 tempest-ServerDiagnosticsTest-1737887255-project-member] Acquired lock "refresh_cache-2cf18a04-6bf9-4f47-a920-2c568207057d" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 670.835939] env[62066]: DEBUG nova.network.neutron [None req-cf333fac-d104-4122-9317-eab87c4499f0 tempest-ServerDiagnosticsTest-1737887255 tempest-ServerDiagnosticsTest-1737887255-project-member] [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 671.102319] env[62066]: DEBUG nova.scheduler.client.report [None req-2c107b45-3cbf-4be9-af5a-796fbdb9b5fe tempest-ImagesNegativeTestJSON-881899573 tempest-ImagesNegativeTestJSON-881899573-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 671.355707] env[62066]: DEBUG nova.network.neutron [None req-cf333fac-d104-4122-9317-eab87c4499f0 tempest-ServerDiagnosticsTest-1737887255 tempest-ServerDiagnosticsTest-1737887255-project-member] [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 671.413333] env[62066]: DEBUG nova.network.neutron [None req-cf333fac-d104-4122-9317-eab87c4499f0 tempest-ServerDiagnosticsTest-1737887255 tempest-ServerDiagnosticsTest-1737887255-project-member] [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 671.590230] env[62066]: DEBUG nova.compute.manager [None req-67a0ccea-4cde-4de9-9aa4-1bb67e83772e tempest-ServersNegativeTestJSON-789906455 tempest-ServersNegativeTestJSON-789906455-project-member] [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 671.611251] env[62066]: DEBUG oslo_concurrency.lockutils [None req-2c107b45-3cbf-4be9-af5a-796fbdb9b5fe tempest-ImagesNegativeTestJSON-881899573 tempest-ImagesNegativeTestJSON-881899573-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.052s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 671.611938] env[62066]: ERROR nova.compute.manager [None req-2c107b45-3cbf-4be9-af5a-796fbdb9b5fe tempest-ImagesNegativeTestJSON-881899573 tempest-ImagesNegativeTestJSON-881899573-project-member] [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port aa9ed432-13ef-49d1-94be-5f4666d101ef, please check neutron logs for more information. [ 671.611938] env[62066]: ERROR nova.compute.manager [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] Traceback (most recent call last): [ 671.611938] env[62066]: ERROR nova.compute.manager [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 671.611938] env[62066]: ERROR nova.compute.manager [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] self.driver.spawn(context, instance, image_meta, [ 671.611938] env[62066]: ERROR nova.compute.manager [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 671.611938] env[62066]: ERROR nova.compute.manager [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] self._vmops.spawn(context, instance, image_meta, injected_files, [ 671.611938] env[62066]: ERROR nova.compute.manager [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 671.611938] env[62066]: ERROR nova.compute.manager [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] vm_ref = self.build_virtual_machine(instance, [ 671.611938] env[62066]: ERROR nova.compute.manager [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 671.611938] env[62066]: ERROR nova.compute.manager [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] vif_infos = vmwarevif.get_vif_info(self._session, [ 671.611938] env[62066]: ERROR nova.compute.manager [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 671.612268] env[62066]: ERROR nova.compute.manager [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] for vif in network_info: [ 671.612268] env[62066]: ERROR nova.compute.manager [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 671.612268] env[62066]: ERROR nova.compute.manager [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] return self._sync_wrapper(fn, *args, **kwargs) [ 671.612268] env[62066]: ERROR nova.compute.manager [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 671.612268] env[62066]: ERROR nova.compute.manager [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] self.wait() [ 671.612268] env[62066]: ERROR nova.compute.manager [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 671.612268] env[62066]: ERROR nova.compute.manager [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] self[:] = self._gt.wait() [ 671.612268] env[62066]: ERROR nova.compute.manager [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 671.612268] env[62066]: ERROR nova.compute.manager [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] return self._exit_event.wait() [ 671.612268] env[62066]: ERROR nova.compute.manager [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 671.612268] env[62066]: ERROR nova.compute.manager [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] result = hub.switch() [ 671.612268] env[62066]: ERROR nova.compute.manager [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 671.612268] env[62066]: ERROR nova.compute.manager [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] return self.greenlet.switch() [ 671.612591] env[62066]: ERROR nova.compute.manager [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 671.612591] env[62066]: ERROR nova.compute.manager [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] result = function(*args, **kwargs) [ 671.612591] env[62066]: ERROR nova.compute.manager [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 671.612591] env[62066]: ERROR nova.compute.manager [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] return func(*args, **kwargs) [ 671.612591] env[62066]: ERROR nova.compute.manager [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 671.612591] env[62066]: ERROR nova.compute.manager [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] raise e [ 671.612591] env[62066]: ERROR nova.compute.manager [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 671.612591] env[62066]: ERROR nova.compute.manager [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] nwinfo = self.network_api.allocate_for_instance( [ 671.612591] env[62066]: ERROR nova.compute.manager [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 671.612591] env[62066]: ERROR nova.compute.manager [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] created_port_ids = self._update_ports_for_instance( [ 671.612591] env[62066]: ERROR nova.compute.manager [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 671.612591] env[62066]: ERROR nova.compute.manager [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] with excutils.save_and_reraise_exception(): [ 671.612591] env[62066]: ERROR nova.compute.manager [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 671.612915] env[62066]: ERROR nova.compute.manager [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] self.force_reraise() [ 671.612915] env[62066]: ERROR nova.compute.manager [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 671.612915] env[62066]: ERROR nova.compute.manager [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] raise self.value [ 671.612915] env[62066]: ERROR nova.compute.manager [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 671.612915] env[62066]: ERROR nova.compute.manager [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] updated_port = self._update_port( [ 671.612915] env[62066]: ERROR nova.compute.manager [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 671.612915] env[62066]: ERROR nova.compute.manager [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] _ensure_no_port_binding_failure(port) [ 671.612915] env[62066]: ERROR nova.compute.manager [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 671.612915] env[62066]: ERROR nova.compute.manager [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] raise exception.PortBindingFailed(port_id=port['id']) [ 671.612915] env[62066]: ERROR nova.compute.manager [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] nova.exception.PortBindingFailed: Binding failed for port aa9ed432-13ef-49d1-94be-5f4666d101ef, please check neutron logs for more information. [ 671.612915] env[62066]: ERROR nova.compute.manager [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] [ 671.613250] env[62066]: DEBUG nova.compute.utils [None req-2c107b45-3cbf-4be9-af5a-796fbdb9b5fe tempest-ImagesNegativeTestJSON-881899573 tempest-ImagesNegativeTestJSON-881899573-project-member] [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] Binding failed for port aa9ed432-13ef-49d1-94be-5f4666d101ef, please check neutron logs for more information. {{(pid=62066) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 671.615621] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b96d7111-6e60-4fd2-a38a-3863242412fb tempest-ServerActionsTestOtherB-426172427 tempest-ServerActionsTestOtherB-426172427-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.995s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 671.617425] env[62066]: INFO nova.compute.claims [None req-b96d7111-6e60-4fd2-a38a-3863242412fb tempest-ServerActionsTestOtherB-426172427 tempest-ServerActionsTestOtherB-426172427-project-member] [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 671.620752] env[62066]: DEBUG nova.compute.manager [None req-2c107b45-3cbf-4be9-af5a-796fbdb9b5fe tempest-ImagesNegativeTestJSON-881899573 tempest-ImagesNegativeTestJSON-881899573-project-member] [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] Build of instance cb319720-29ab-4ff2-a71e-f77ae8c85735 was re-scheduled: Binding failed for port aa9ed432-13ef-49d1-94be-5f4666d101ef, please check neutron logs for more information. {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 671.620873] env[62066]: DEBUG nova.compute.manager [None req-2c107b45-3cbf-4be9-af5a-796fbdb9b5fe tempest-ImagesNegativeTestJSON-881899573 tempest-ImagesNegativeTestJSON-881899573-project-member] [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] Unplugging VIFs for instance {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 671.621745] env[62066]: DEBUG oslo_concurrency.lockutils [None req-2c107b45-3cbf-4be9-af5a-796fbdb9b5fe tempest-ImagesNegativeTestJSON-881899573 tempest-ImagesNegativeTestJSON-881899573-project-member] Acquiring lock "refresh_cache-cb319720-29ab-4ff2-a71e-f77ae8c85735" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 671.621933] env[62066]: DEBUG oslo_concurrency.lockutils [None req-2c107b45-3cbf-4be9-af5a-796fbdb9b5fe tempest-ImagesNegativeTestJSON-881899573 tempest-ImagesNegativeTestJSON-881899573-project-member] Acquired lock "refresh_cache-cb319720-29ab-4ff2-a71e-f77ae8c85735" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 671.622139] env[62066]: DEBUG nova.network.neutron [None req-2c107b45-3cbf-4be9-af5a-796fbdb9b5fe tempest-ImagesNegativeTestJSON-881899573 tempest-ImagesNegativeTestJSON-881899573-project-member] [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 671.629224] env[62066]: DEBUG nova.virt.hardware [None req-67a0ccea-4cde-4de9-9aa4-1bb67e83772e tempest-ServersNegativeTestJSON-789906455 tempest-ServersNegativeTestJSON-789906455-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-23T13:40:41Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-23T13:40:23Z,direct_url=,disk_format='vmdk',id=50ff584c-3b50-4395-af07-3e66769bc9f7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='eb52f7069a374c61ae946f052007c6d9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-23T13:40:24Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 671.629456] env[62066]: DEBUG nova.virt.hardware [None req-67a0ccea-4cde-4de9-9aa4-1bb67e83772e tempest-ServersNegativeTestJSON-789906455 tempest-ServersNegativeTestJSON-789906455-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 671.629610] env[62066]: DEBUG nova.virt.hardware [None req-67a0ccea-4cde-4de9-9aa4-1bb67e83772e tempest-ServersNegativeTestJSON-789906455 tempest-ServersNegativeTestJSON-789906455-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 671.629781] env[62066]: DEBUG nova.virt.hardware [None req-67a0ccea-4cde-4de9-9aa4-1bb67e83772e tempest-ServersNegativeTestJSON-789906455 tempest-ServersNegativeTestJSON-789906455-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 671.629920] env[62066]: DEBUG nova.virt.hardware [None req-67a0ccea-4cde-4de9-9aa4-1bb67e83772e tempest-ServersNegativeTestJSON-789906455 tempest-ServersNegativeTestJSON-789906455-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 671.630086] env[62066]: DEBUG nova.virt.hardware [None req-67a0ccea-4cde-4de9-9aa4-1bb67e83772e tempest-ServersNegativeTestJSON-789906455 tempest-ServersNegativeTestJSON-789906455-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 671.630293] env[62066]: DEBUG nova.virt.hardware [None req-67a0ccea-4cde-4de9-9aa4-1bb67e83772e tempest-ServersNegativeTestJSON-789906455 tempest-ServersNegativeTestJSON-789906455-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 671.630445] env[62066]: DEBUG nova.virt.hardware [None req-67a0ccea-4cde-4de9-9aa4-1bb67e83772e tempest-ServersNegativeTestJSON-789906455 tempest-ServersNegativeTestJSON-789906455-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 671.630626] env[62066]: DEBUG nova.virt.hardware [None req-67a0ccea-4cde-4de9-9aa4-1bb67e83772e tempest-ServersNegativeTestJSON-789906455 tempest-ServersNegativeTestJSON-789906455-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 671.630783] env[62066]: DEBUG nova.virt.hardware [None req-67a0ccea-4cde-4de9-9aa4-1bb67e83772e tempest-ServersNegativeTestJSON-789906455 tempest-ServersNegativeTestJSON-789906455-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 671.630973] env[62066]: DEBUG nova.virt.hardware [None req-67a0ccea-4cde-4de9-9aa4-1bb67e83772e tempest-ServersNegativeTestJSON-789906455 tempest-ServersNegativeTestJSON-789906455-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 671.631859] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8dad7f28-5bd0-46b6-94bc-be7f664bf067 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.641706] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9eca8fe-5276-4c32-8173-b7df60c28def {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.742583] env[62066]: ERROR nova.compute.manager [None req-67a0ccea-4cde-4de9-9aa4-1bb67e83772e tempest-ServersNegativeTestJSON-789906455 tempest-ServersNegativeTestJSON-789906455-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port ea8fd014-c428-41e1-a6bb-7aba32299177, please check neutron logs for more information. [ 671.742583] env[62066]: ERROR nova.compute.manager Traceback (most recent call last): [ 671.742583] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 671.742583] env[62066]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 671.742583] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 671.742583] env[62066]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 671.742583] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 671.742583] env[62066]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 671.742583] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 671.742583] env[62066]: ERROR nova.compute.manager self.force_reraise() [ 671.742583] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 671.742583] env[62066]: ERROR nova.compute.manager raise self.value [ 671.742583] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 671.742583] env[62066]: ERROR nova.compute.manager updated_port = self._update_port( [ 671.742583] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 671.742583] env[62066]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 671.743118] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 671.743118] env[62066]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 671.743118] env[62066]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port ea8fd014-c428-41e1-a6bb-7aba32299177, please check neutron logs for more information. [ 671.743118] env[62066]: ERROR nova.compute.manager [ 671.743118] env[62066]: Traceback (most recent call last): [ 671.743118] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 671.743118] env[62066]: listener.cb(fileno) [ 671.743118] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 671.743118] env[62066]: result = function(*args, **kwargs) [ 671.743118] env[62066]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 671.743118] env[62066]: return func(*args, **kwargs) [ 671.743118] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 671.743118] env[62066]: raise e [ 671.743118] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 671.743118] env[62066]: nwinfo = self.network_api.allocate_for_instance( [ 671.743118] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 671.743118] env[62066]: created_port_ids = self._update_ports_for_instance( [ 671.743118] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 671.743118] env[62066]: with excutils.save_and_reraise_exception(): [ 671.743118] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 671.743118] env[62066]: self.force_reraise() [ 671.743118] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 671.743118] env[62066]: raise self.value [ 671.743118] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 671.743118] env[62066]: updated_port = self._update_port( [ 671.743118] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 671.743118] env[62066]: _ensure_no_port_binding_failure(port) [ 671.743118] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 671.743118] env[62066]: raise exception.PortBindingFailed(port_id=port['id']) [ 671.744182] env[62066]: nova.exception.PortBindingFailed: Binding failed for port ea8fd014-c428-41e1-a6bb-7aba32299177, please check neutron logs for more information. [ 671.744182] env[62066]: Removing descriptor: 20 [ 671.744182] env[62066]: ERROR nova.compute.manager [None req-67a0ccea-4cde-4de9-9aa4-1bb67e83772e tempest-ServersNegativeTestJSON-789906455 tempest-ServersNegativeTestJSON-789906455-project-member] [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port ea8fd014-c428-41e1-a6bb-7aba32299177, please check neutron logs for more information. [ 671.744182] env[62066]: ERROR nova.compute.manager [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] Traceback (most recent call last): [ 671.744182] env[62066]: ERROR nova.compute.manager [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 671.744182] env[62066]: ERROR nova.compute.manager [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] yield resources [ 671.744182] env[62066]: ERROR nova.compute.manager [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 671.744182] env[62066]: ERROR nova.compute.manager [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] self.driver.spawn(context, instance, image_meta, [ 671.744182] env[62066]: ERROR nova.compute.manager [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 671.744182] env[62066]: ERROR nova.compute.manager [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] self._vmops.spawn(context, instance, image_meta, injected_files, [ 671.744182] env[62066]: ERROR nova.compute.manager [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 671.744182] env[62066]: ERROR nova.compute.manager [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] vm_ref = self.build_virtual_machine(instance, [ 671.744556] env[62066]: ERROR nova.compute.manager [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 671.744556] env[62066]: ERROR nova.compute.manager [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] vif_infos = vmwarevif.get_vif_info(self._session, [ 671.744556] env[62066]: ERROR nova.compute.manager [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 671.744556] env[62066]: ERROR nova.compute.manager [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] for vif in network_info: [ 671.744556] env[62066]: ERROR nova.compute.manager [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 671.744556] env[62066]: ERROR nova.compute.manager [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] return self._sync_wrapper(fn, *args, **kwargs) [ 671.744556] env[62066]: ERROR nova.compute.manager [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 671.744556] env[62066]: ERROR nova.compute.manager [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] self.wait() [ 671.744556] env[62066]: ERROR nova.compute.manager [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 671.744556] env[62066]: ERROR nova.compute.manager [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] self[:] = self._gt.wait() [ 671.744556] env[62066]: ERROR nova.compute.manager [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 671.744556] env[62066]: ERROR nova.compute.manager [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] return self._exit_event.wait() [ 671.744556] env[62066]: ERROR nova.compute.manager [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 671.744937] env[62066]: ERROR nova.compute.manager [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] result = hub.switch() [ 671.744937] env[62066]: ERROR nova.compute.manager [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 671.744937] env[62066]: ERROR nova.compute.manager [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] return self.greenlet.switch() [ 671.744937] env[62066]: ERROR nova.compute.manager [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 671.744937] env[62066]: ERROR nova.compute.manager [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] result = function(*args, **kwargs) [ 671.744937] env[62066]: ERROR nova.compute.manager [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 671.744937] env[62066]: ERROR nova.compute.manager [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] return func(*args, **kwargs) [ 671.744937] env[62066]: ERROR nova.compute.manager [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 671.744937] env[62066]: ERROR nova.compute.manager [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] raise e [ 671.744937] env[62066]: ERROR nova.compute.manager [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 671.744937] env[62066]: ERROR nova.compute.manager [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] nwinfo = self.network_api.allocate_for_instance( [ 671.744937] env[62066]: ERROR nova.compute.manager [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 671.744937] env[62066]: ERROR nova.compute.manager [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] created_port_ids = self._update_ports_for_instance( [ 671.745328] env[62066]: ERROR nova.compute.manager [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 671.745328] env[62066]: ERROR nova.compute.manager [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] with excutils.save_and_reraise_exception(): [ 671.745328] env[62066]: ERROR nova.compute.manager [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 671.745328] env[62066]: ERROR nova.compute.manager [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] self.force_reraise() [ 671.745328] env[62066]: ERROR nova.compute.manager [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 671.745328] env[62066]: ERROR nova.compute.manager [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] raise self.value [ 671.745328] env[62066]: ERROR nova.compute.manager [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 671.745328] env[62066]: ERROR nova.compute.manager [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] updated_port = self._update_port( [ 671.745328] env[62066]: ERROR nova.compute.manager [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 671.745328] env[62066]: ERROR nova.compute.manager [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] _ensure_no_port_binding_failure(port) [ 671.745328] env[62066]: ERROR nova.compute.manager [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 671.745328] env[62066]: ERROR nova.compute.manager [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] raise exception.PortBindingFailed(port_id=port['id']) [ 671.745698] env[62066]: ERROR nova.compute.manager [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] nova.exception.PortBindingFailed: Binding failed for port ea8fd014-c428-41e1-a6bb-7aba32299177, please check neutron logs for more information. [ 671.745698] env[62066]: ERROR nova.compute.manager [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] [ 671.745698] env[62066]: INFO nova.compute.manager [None req-67a0ccea-4cde-4de9-9aa4-1bb67e83772e tempest-ServersNegativeTestJSON-789906455 tempest-ServersNegativeTestJSON-789906455-project-member] [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] Terminating instance [ 671.746612] env[62066]: DEBUG oslo_concurrency.lockutils [None req-67a0ccea-4cde-4de9-9aa4-1bb67e83772e tempest-ServersNegativeTestJSON-789906455 tempest-ServersNegativeTestJSON-789906455-project-member] Acquiring lock "refresh_cache-5bbdf138-0aaf-42d1-b5a1-4f7ebc650393" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 671.746815] env[62066]: DEBUG oslo_concurrency.lockutils [None req-67a0ccea-4cde-4de9-9aa4-1bb67e83772e tempest-ServersNegativeTestJSON-789906455 tempest-ServersNegativeTestJSON-789906455-project-member] Acquired lock "refresh_cache-5bbdf138-0aaf-42d1-b5a1-4f7ebc650393" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 671.746963] env[62066]: DEBUG nova.network.neutron [None req-67a0ccea-4cde-4de9-9aa4-1bb67e83772e tempest-ServersNegativeTestJSON-789906455 tempest-ServersNegativeTestJSON-789906455-project-member] [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 671.901050] env[62066]: DEBUG nova.compute.manager [req-df554add-17a4-4166-a6c3-7220e35f5a47 req-92dbf117-93de-42f0-a4b0-f6f047bfb5a6 service nova] [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] Received event network-vif-deleted-70d1fb9d-0a6e-46b4-a478-297d09fbe121 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 671.901050] env[62066]: DEBUG nova.compute.manager [req-df554add-17a4-4166-a6c3-7220e35f5a47 req-92dbf117-93de-42f0-a4b0-f6f047bfb5a6 service nova] [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] Received event network-changed-ea8fd014-c428-41e1-a6bb-7aba32299177 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 671.901050] env[62066]: DEBUG nova.compute.manager [req-df554add-17a4-4166-a6c3-7220e35f5a47 req-92dbf117-93de-42f0-a4b0-f6f047bfb5a6 service nova] [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] Refreshing instance network info cache due to event network-changed-ea8fd014-c428-41e1-a6bb-7aba32299177. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 671.901245] env[62066]: DEBUG oslo_concurrency.lockutils [req-df554add-17a4-4166-a6c3-7220e35f5a47 req-92dbf117-93de-42f0-a4b0-f6f047bfb5a6 service nova] Acquiring lock "refresh_cache-5bbdf138-0aaf-42d1-b5a1-4f7ebc650393" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 671.915686] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cf333fac-d104-4122-9317-eab87c4499f0 tempest-ServerDiagnosticsTest-1737887255 tempest-ServerDiagnosticsTest-1737887255-project-member] Releasing lock "refresh_cache-2cf18a04-6bf9-4f47-a920-2c568207057d" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 671.916146] env[62066]: DEBUG nova.compute.manager [None req-cf333fac-d104-4122-9317-eab87c4499f0 tempest-ServerDiagnosticsTest-1737887255 tempest-ServerDiagnosticsTest-1737887255-project-member] [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 671.916346] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-cf333fac-d104-4122-9317-eab87c4499f0 tempest-ServerDiagnosticsTest-1737887255 tempest-ServerDiagnosticsTest-1737887255-project-member] [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 671.917199] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-02ea0e11-7f75-40e5-8f6b-ddf1f3edd36d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.925962] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2f842b0-210c-48a4-8e15-a33d515cce8f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.947209] env[62066]: WARNING nova.virt.vmwareapi.vmops [None req-cf333fac-d104-4122-9317-eab87c4499f0 tempest-ServerDiagnosticsTest-1737887255 tempest-ServerDiagnosticsTest-1737887255-project-member] [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 2cf18a04-6bf9-4f47-a920-2c568207057d could not be found. [ 671.947444] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-cf333fac-d104-4122-9317-eab87c4499f0 tempest-ServerDiagnosticsTest-1737887255 tempest-ServerDiagnosticsTest-1737887255-project-member] [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 671.947623] env[62066]: INFO nova.compute.manager [None req-cf333fac-d104-4122-9317-eab87c4499f0 tempest-ServerDiagnosticsTest-1737887255 tempest-ServerDiagnosticsTest-1737887255-project-member] [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] Took 0.03 seconds to destroy the instance on the hypervisor. [ 671.947872] env[62066]: DEBUG oslo.service.loopingcall [None req-cf333fac-d104-4122-9317-eab87c4499f0 tempest-ServerDiagnosticsTest-1737887255 tempest-ServerDiagnosticsTest-1737887255-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 671.948106] env[62066]: DEBUG nova.compute.manager [-] [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 671.948202] env[62066]: DEBUG nova.network.neutron [-] [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 671.975403] env[62066]: DEBUG nova.network.neutron [-] [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 672.146957] env[62066]: DEBUG nova.network.neutron [None req-2c107b45-3cbf-4be9-af5a-796fbdb9b5fe tempest-ImagesNegativeTestJSON-881899573 tempest-ImagesNegativeTestJSON-881899573-project-member] [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 672.251872] env[62066]: DEBUG nova.network.neutron [None req-2c107b45-3cbf-4be9-af5a-796fbdb9b5fe tempest-ImagesNegativeTestJSON-881899573 tempest-ImagesNegativeTestJSON-881899573-project-member] [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 672.274070] env[62066]: DEBUG nova.network.neutron [None req-67a0ccea-4cde-4de9-9aa4-1bb67e83772e tempest-ServersNegativeTestJSON-789906455 tempest-ServersNegativeTestJSON-789906455-project-member] [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 672.350943] env[62066]: DEBUG nova.network.neutron [None req-67a0ccea-4cde-4de9-9aa4-1bb67e83772e tempest-ServersNegativeTestJSON-789906455 tempest-ServersNegativeTestJSON-789906455-project-member] [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 672.478510] env[62066]: DEBUG nova.network.neutron [-] [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 672.756318] env[62066]: DEBUG oslo_concurrency.lockutils [None req-2c107b45-3cbf-4be9-af5a-796fbdb9b5fe tempest-ImagesNegativeTestJSON-881899573 tempest-ImagesNegativeTestJSON-881899573-project-member] Releasing lock "refresh_cache-cb319720-29ab-4ff2-a71e-f77ae8c85735" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 672.756318] env[62066]: DEBUG nova.compute.manager [None req-2c107b45-3cbf-4be9-af5a-796fbdb9b5fe tempest-ImagesNegativeTestJSON-881899573 tempest-ImagesNegativeTestJSON-881899573-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 672.756318] env[62066]: DEBUG nova.compute.manager [None req-2c107b45-3cbf-4be9-af5a-796fbdb9b5fe tempest-ImagesNegativeTestJSON-881899573 tempest-ImagesNegativeTestJSON-881899573-project-member] [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 672.756318] env[62066]: DEBUG nova.network.neutron [None req-2c107b45-3cbf-4be9-af5a-796fbdb9b5fe tempest-ImagesNegativeTestJSON-881899573 tempest-ImagesNegativeTestJSON-881899573-project-member] [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 672.776418] env[62066]: DEBUG nova.network.neutron [None req-2c107b45-3cbf-4be9-af5a-796fbdb9b5fe tempest-ImagesNegativeTestJSON-881899573 tempest-ImagesNegativeTestJSON-881899573-project-member] [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 672.853240] env[62066]: DEBUG oslo_concurrency.lockutils [None req-67a0ccea-4cde-4de9-9aa4-1bb67e83772e tempest-ServersNegativeTestJSON-789906455 tempest-ServersNegativeTestJSON-789906455-project-member] Releasing lock "refresh_cache-5bbdf138-0aaf-42d1-b5a1-4f7ebc650393" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 672.853646] env[62066]: DEBUG nova.compute.manager [None req-67a0ccea-4cde-4de9-9aa4-1bb67e83772e tempest-ServersNegativeTestJSON-789906455 tempest-ServersNegativeTestJSON-789906455-project-member] [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 672.853833] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-67a0ccea-4cde-4de9-9aa4-1bb67e83772e tempest-ServersNegativeTestJSON-789906455 tempest-ServersNegativeTestJSON-789906455-project-member] [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 672.854828] env[62066]: DEBUG oslo_concurrency.lockutils [req-df554add-17a4-4166-a6c3-7220e35f5a47 req-92dbf117-93de-42f0-a4b0-f6f047bfb5a6 service nova] Acquired lock "refresh_cache-5bbdf138-0aaf-42d1-b5a1-4f7ebc650393" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 672.854828] env[62066]: DEBUG nova.network.neutron [req-df554add-17a4-4166-a6c3-7220e35f5a47 req-92dbf117-93de-42f0-a4b0-f6f047bfb5a6 service nova] [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] Refreshing network info cache for port ea8fd014-c428-41e1-a6bb-7aba32299177 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 672.855614] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-10e22dbf-b519-43a3-b9d5-8ae88bc2b80f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.865944] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8b165c7-2d5c-4645-965f-1b573646bd1e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.891611] env[62066]: WARNING nova.virt.vmwareapi.vmops [None req-67a0ccea-4cde-4de9-9aa4-1bb67e83772e tempest-ServersNegativeTestJSON-789906455 tempest-ServersNegativeTestJSON-789906455-project-member] [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393 could not be found. [ 672.891839] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-67a0ccea-4cde-4de9-9aa4-1bb67e83772e tempest-ServersNegativeTestJSON-789906455 tempest-ServersNegativeTestJSON-789906455-project-member] [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 672.892025] env[62066]: INFO nova.compute.manager [None req-67a0ccea-4cde-4de9-9aa4-1bb67e83772e tempest-ServersNegativeTestJSON-789906455 tempest-ServersNegativeTestJSON-789906455-project-member] [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] Took 0.04 seconds to destroy the instance on the hypervisor. [ 672.892272] env[62066]: DEBUG oslo.service.loopingcall [None req-67a0ccea-4cde-4de9-9aa4-1bb67e83772e tempest-ServersNegativeTestJSON-789906455 tempest-ServersNegativeTestJSON-789906455-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 672.894704] env[62066]: DEBUG nova.compute.manager [-] [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 672.895025] env[62066]: DEBUG nova.network.neutron [-] [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 672.918283] env[62066]: DEBUG nova.network.neutron [-] [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 672.981207] env[62066]: INFO nova.compute.manager [-] [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] Took 1.03 seconds to deallocate network for instance. [ 672.985273] env[62066]: DEBUG nova.compute.claims [None req-cf333fac-d104-4122-9317-eab87c4499f0 tempest-ServerDiagnosticsTest-1737887255 tempest-ServerDiagnosticsTest-1737887255-project-member] [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] Aborting claim: {{(pid=62066) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 672.986027] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cf333fac-d104-4122-9317-eab87c4499f0 tempest-ServerDiagnosticsTest-1737887255 tempest-ServerDiagnosticsTest-1737887255-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 673.084594] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f427e16-72bb-4fd3-b290-6ad2c7cd7e10 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.092524] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa8aaae8-dd71-4371-b91e-7956bbe86348 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.121217] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a9d0242-1086-42fe-9c38-23f081ff9e25 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.128310] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7cfe8a0b-c9ec-47f7-84b0-f0916dbe5a8f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.141087] env[62066]: DEBUG nova.compute.provider_tree [None req-b96d7111-6e60-4fd2-a38a-3863242412fb tempest-ServerActionsTestOtherB-426172427 tempest-ServerActionsTestOtherB-426172427-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 673.279614] env[62066]: DEBUG nova.network.neutron [None req-2c107b45-3cbf-4be9-af5a-796fbdb9b5fe tempest-ImagesNegativeTestJSON-881899573 tempest-ImagesNegativeTestJSON-881899573-project-member] [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 673.385674] env[62066]: DEBUG nova.network.neutron [req-df554add-17a4-4166-a6c3-7220e35f5a47 req-92dbf117-93de-42f0-a4b0-f6f047bfb5a6 service nova] [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 673.421238] env[62066]: DEBUG nova.network.neutron [-] [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 673.502673] env[62066]: DEBUG nova.network.neutron [req-df554add-17a4-4166-a6c3-7220e35f5a47 req-92dbf117-93de-42f0-a4b0-f6f047bfb5a6 service nova] [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 673.648048] env[62066]: DEBUG nova.scheduler.client.report [None req-b96d7111-6e60-4fd2-a38a-3863242412fb tempest-ServerActionsTestOtherB-426172427 tempest-ServerActionsTestOtherB-426172427-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 673.782668] env[62066]: INFO nova.compute.manager [None req-2c107b45-3cbf-4be9-af5a-796fbdb9b5fe tempest-ImagesNegativeTestJSON-881899573 tempest-ImagesNegativeTestJSON-881899573-project-member] [instance: cb319720-29ab-4ff2-a71e-f77ae8c85735] Took 1.03 seconds to deallocate network for instance. [ 673.925432] env[62066]: INFO nova.compute.manager [-] [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] Took 1.03 seconds to deallocate network for instance. [ 673.927371] env[62066]: DEBUG nova.compute.claims [None req-67a0ccea-4cde-4de9-9aa4-1bb67e83772e tempest-ServersNegativeTestJSON-789906455 tempest-ServersNegativeTestJSON-789906455-project-member] [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] Aborting claim: {{(pid=62066) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 673.927371] env[62066]: DEBUG oslo_concurrency.lockutils [None req-67a0ccea-4cde-4de9-9aa4-1bb67e83772e tempest-ServersNegativeTestJSON-789906455 tempest-ServersNegativeTestJSON-789906455-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 673.957506] env[62066]: DEBUG nova.compute.manager [req-eddc60c8-3308-4791-93ad-510cb51add6a req-adbcf2db-e91e-48c3-8bc7-8c81bb145462 service nova] [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] Received event network-vif-deleted-ea8fd014-c428-41e1-a6bb-7aba32299177 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 674.007346] env[62066]: DEBUG oslo_concurrency.lockutils [req-df554add-17a4-4166-a6c3-7220e35f5a47 req-92dbf117-93de-42f0-a4b0-f6f047bfb5a6 service nova] Releasing lock "refresh_cache-5bbdf138-0aaf-42d1-b5a1-4f7ebc650393" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 674.152636] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b96d7111-6e60-4fd2-a38a-3863242412fb tempest-ServerActionsTestOtherB-426172427 tempest-ServerActionsTestOtherB-426172427-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.534s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 674.152636] env[62066]: DEBUG nova.compute.manager [None req-b96d7111-6e60-4fd2-a38a-3863242412fb tempest-ServerActionsTestOtherB-426172427 tempest-ServerActionsTestOtherB-426172427-project-member] [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 674.158306] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b5125342-7692-4177-9e88-9f6d8ef9f1dc tempest-ServersAaction247Test-1482016775 tempest-ServersAaction247Test-1482016775-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 27.519s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 674.158306] env[62066]: DEBUG nova.objects.instance [None req-b5125342-7692-4177-9e88-9f6d8ef9f1dc tempest-ServersAaction247Test-1482016775 tempest-ServersAaction247Test-1482016775-project-member] Lazy-loading 'resources' on Instance uuid 5d54c3fd-b239-4965-b7a3-5909e8de8bc0 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 674.663019] env[62066]: DEBUG nova.compute.utils [None req-b96d7111-6e60-4fd2-a38a-3863242412fb tempest-ServerActionsTestOtherB-426172427 tempest-ServerActionsTestOtherB-426172427-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 674.664900] env[62066]: DEBUG nova.compute.manager [None req-b96d7111-6e60-4fd2-a38a-3863242412fb tempest-ServerActionsTestOtherB-426172427 tempest-ServerActionsTestOtherB-426172427-project-member] [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 674.665647] env[62066]: DEBUG nova.network.neutron [None req-b96d7111-6e60-4fd2-a38a-3863242412fb tempest-ServerActionsTestOtherB-426172427 tempest-ServerActionsTestOtherB-426172427-project-member] [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 674.733193] env[62066]: DEBUG nova.policy [None req-b96d7111-6e60-4fd2-a38a-3863242412fb tempest-ServerActionsTestOtherB-426172427 tempest-ServerActionsTestOtherB-426172427-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '64cc8214951a4ba9a013936b3d3f74c6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b0ea5ddf21fb4a02aed89f371ee441d5', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 674.836286] env[62066]: INFO nova.scheduler.client.report [None req-2c107b45-3cbf-4be9-af5a-796fbdb9b5fe tempest-ImagesNegativeTestJSON-881899573 tempest-ImagesNegativeTestJSON-881899573-project-member] Deleted allocations for instance cb319720-29ab-4ff2-a71e-f77ae8c85735 [ 675.141427] env[62066]: DEBUG nova.network.neutron [None req-b96d7111-6e60-4fd2-a38a-3863242412fb tempest-ServerActionsTestOtherB-426172427 tempest-ServerActionsTestOtherB-426172427-project-member] [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] Successfully created port: 9380aec8-fb60-43d1-b0f9-2068fbbc6c28 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 675.165426] env[62066]: DEBUG nova.compute.manager [None req-b96d7111-6e60-4fd2-a38a-3863242412fb tempest-ServerActionsTestOtherB-426172427 tempest-ServerActionsTestOtherB-426172427-project-member] [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 675.173602] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1433814-ac2e-49a5-ac92-8d93317a4221 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.181751] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e7fde91-9658-4804-bb96-38f3e3fcf967 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.214349] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2256494f-204e-430f-845b-8d3b3eaff0f3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.223793] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c18dc538-419a-4ed0-9c83-ae2d4d213841 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.236341] env[62066]: DEBUG nova.compute.provider_tree [None req-b5125342-7692-4177-9e88-9f6d8ef9f1dc tempest-ServersAaction247Test-1482016775 tempest-ServersAaction247Test-1482016775-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 675.350998] env[62066]: DEBUG oslo_concurrency.lockutils [None req-2c107b45-3cbf-4be9-af5a-796fbdb9b5fe tempest-ImagesNegativeTestJSON-881899573 tempest-ImagesNegativeTestJSON-881899573-project-member] Lock "cb319720-29ab-4ff2-a71e-f77ae8c85735" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 139.438s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 675.740873] env[62066]: DEBUG nova.scheduler.client.report [None req-b5125342-7692-4177-9e88-9f6d8ef9f1dc tempest-ServersAaction247Test-1482016775 tempest-ServersAaction247Test-1482016775-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 675.856648] env[62066]: DEBUG nova.compute.manager [None req-b2b06eaa-4237-476b-9167-ed8be81ba3b9 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 676.041593] env[62066]: DEBUG nova.compute.manager [req-66274895-b199-4f89-bb81-3f31e1eb18e8 req-6da99013-f757-402e-ac11-727c24da45b8 service nova] [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] Received event network-changed-9380aec8-fb60-43d1-b0f9-2068fbbc6c28 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 676.041593] env[62066]: DEBUG nova.compute.manager [req-66274895-b199-4f89-bb81-3f31e1eb18e8 req-6da99013-f757-402e-ac11-727c24da45b8 service nova] [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] Refreshing instance network info cache due to event network-changed-9380aec8-fb60-43d1-b0f9-2068fbbc6c28. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 676.041593] env[62066]: DEBUG oslo_concurrency.lockutils [req-66274895-b199-4f89-bb81-3f31e1eb18e8 req-6da99013-f757-402e-ac11-727c24da45b8 service nova] Acquiring lock "refresh_cache-e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 676.041593] env[62066]: DEBUG oslo_concurrency.lockutils [req-66274895-b199-4f89-bb81-3f31e1eb18e8 req-6da99013-f757-402e-ac11-727c24da45b8 service nova] Acquired lock "refresh_cache-e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 676.049062] env[62066]: DEBUG nova.network.neutron [req-66274895-b199-4f89-bb81-3f31e1eb18e8 req-6da99013-f757-402e-ac11-727c24da45b8 service nova] [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] Refreshing network info cache for port 9380aec8-fb60-43d1-b0f9-2068fbbc6c28 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 676.148854] env[62066]: ERROR nova.compute.manager [None req-b96d7111-6e60-4fd2-a38a-3863242412fb tempest-ServerActionsTestOtherB-426172427 tempest-ServerActionsTestOtherB-426172427-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 9380aec8-fb60-43d1-b0f9-2068fbbc6c28, please check neutron logs for more information. [ 676.148854] env[62066]: ERROR nova.compute.manager Traceback (most recent call last): [ 676.148854] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 676.148854] env[62066]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 676.148854] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 676.148854] env[62066]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 676.148854] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 676.148854] env[62066]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 676.148854] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 676.148854] env[62066]: ERROR nova.compute.manager self.force_reraise() [ 676.148854] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 676.148854] env[62066]: ERROR nova.compute.manager raise self.value [ 676.148854] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 676.148854] env[62066]: ERROR nova.compute.manager updated_port = self._update_port( [ 676.148854] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 676.148854] env[62066]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 676.149315] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 676.149315] env[62066]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 676.149315] env[62066]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 9380aec8-fb60-43d1-b0f9-2068fbbc6c28, please check neutron logs for more information. [ 676.149315] env[62066]: ERROR nova.compute.manager [ 676.149315] env[62066]: Traceback (most recent call last): [ 676.149315] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 676.149315] env[62066]: listener.cb(fileno) [ 676.149315] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 676.149315] env[62066]: result = function(*args, **kwargs) [ 676.149315] env[62066]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 676.149315] env[62066]: return func(*args, **kwargs) [ 676.149315] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 676.149315] env[62066]: raise e [ 676.149315] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 676.149315] env[62066]: nwinfo = self.network_api.allocate_for_instance( [ 676.149315] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 676.149315] env[62066]: created_port_ids = self._update_ports_for_instance( [ 676.149315] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 676.149315] env[62066]: with excutils.save_and_reraise_exception(): [ 676.149315] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 676.149315] env[62066]: self.force_reraise() [ 676.149315] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 676.149315] env[62066]: raise self.value [ 676.149315] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 676.149315] env[62066]: updated_port = self._update_port( [ 676.149315] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 676.149315] env[62066]: _ensure_no_port_binding_failure(port) [ 676.149315] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 676.149315] env[62066]: raise exception.PortBindingFailed(port_id=port['id']) [ 676.150298] env[62066]: nova.exception.PortBindingFailed: Binding failed for port 9380aec8-fb60-43d1-b0f9-2068fbbc6c28, please check neutron logs for more information. [ 676.150298] env[62066]: Removing descriptor: 15 [ 676.178584] env[62066]: DEBUG nova.compute.manager [None req-b96d7111-6e60-4fd2-a38a-3863242412fb tempest-ServerActionsTestOtherB-426172427 tempest-ServerActionsTestOtherB-426172427-project-member] [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 676.204252] env[62066]: DEBUG nova.virt.hardware [None req-b96d7111-6e60-4fd2-a38a-3863242412fb tempest-ServerActionsTestOtherB-426172427 tempest-ServerActionsTestOtherB-426172427-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-23T13:40:41Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-23T13:40:23Z,direct_url=,disk_format='vmdk',id=50ff584c-3b50-4395-af07-3e66769bc9f7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='eb52f7069a374c61ae946f052007c6d9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-23T13:40:24Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 676.204497] env[62066]: DEBUG nova.virt.hardware [None req-b96d7111-6e60-4fd2-a38a-3863242412fb tempest-ServerActionsTestOtherB-426172427 tempest-ServerActionsTestOtherB-426172427-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 676.204652] env[62066]: DEBUG nova.virt.hardware [None req-b96d7111-6e60-4fd2-a38a-3863242412fb tempest-ServerActionsTestOtherB-426172427 tempest-ServerActionsTestOtherB-426172427-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 676.204832] env[62066]: DEBUG nova.virt.hardware [None req-b96d7111-6e60-4fd2-a38a-3863242412fb tempest-ServerActionsTestOtherB-426172427 tempest-ServerActionsTestOtherB-426172427-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 676.205011] env[62066]: DEBUG nova.virt.hardware [None req-b96d7111-6e60-4fd2-a38a-3863242412fb tempest-ServerActionsTestOtherB-426172427 tempest-ServerActionsTestOtherB-426172427-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 676.205135] env[62066]: DEBUG nova.virt.hardware [None req-b96d7111-6e60-4fd2-a38a-3863242412fb tempest-ServerActionsTestOtherB-426172427 tempest-ServerActionsTestOtherB-426172427-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 676.205344] env[62066]: DEBUG nova.virt.hardware [None req-b96d7111-6e60-4fd2-a38a-3863242412fb tempest-ServerActionsTestOtherB-426172427 tempest-ServerActionsTestOtherB-426172427-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 676.205501] env[62066]: DEBUG nova.virt.hardware [None req-b96d7111-6e60-4fd2-a38a-3863242412fb tempest-ServerActionsTestOtherB-426172427 tempest-ServerActionsTestOtherB-426172427-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 676.205666] env[62066]: DEBUG nova.virt.hardware [None req-b96d7111-6e60-4fd2-a38a-3863242412fb tempest-ServerActionsTestOtherB-426172427 tempest-ServerActionsTestOtherB-426172427-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 676.205824] env[62066]: DEBUG nova.virt.hardware [None req-b96d7111-6e60-4fd2-a38a-3863242412fb tempest-ServerActionsTestOtherB-426172427 tempest-ServerActionsTestOtherB-426172427-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 676.206094] env[62066]: DEBUG nova.virt.hardware [None req-b96d7111-6e60-4fd2-a38a-3863242412fb tempest-ServerActionsTestOtherB-426172427 tempest-ServerActionsTestOtherB-426172427-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 676.207016] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9108dc2-6ac5-4cbf-ab1b-977c31d8fe89 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.215372] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d184245-9627-4ed3-829b-6da4bfe6227b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.230519] env[62066]: ERROR nova.compute.manager [None req-b96d7111-6e60-4fd2-a38a-3863242412fb tempest-ServerActionsTestOtherB-426172427 tempest-ServerActionsTestOtherB-426172427-project-member] [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 9380aec8-fb60-43d1-b0f9-2068fbbc6c28, please check neutron logs for more information. [ 676.230519] env[62066]: ERROR nova.compute.manager [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] Traceback (most recent call last): [ 676.230519] env[62066]: ERROR nova.compute.manager [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 676.230519] env[62066]: ERROR nova.compute.manager [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] yield resources [ 676.230519] env[62066]: ERROR nova.compute.manager [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 676.230519] env[62066]: ERROR nova.compute.manager [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] self.driver.spawn(context, instance, image_meta, [ 676.230519] env[62066]: ERROR nova.compute.manager [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 676.230519] env[62066]: ERROR nova.compute.manager [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] self._vmops.spawn(context, instance, image_meta, injected_files, [ 676.230519] env[62066]: ERROR nova.compute.manager [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 676.230519] env[62066]: ERROR nova.compute.manager [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] vm_ref = self.build_virtual_machine(instance, [ 676.230519] env[62066]: ERROR nova.compute.manager [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 676.231144] env[62066]: ERROR nova.compute.manager [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] vif_infos = vmwarevif.get_vif_info(self._session, [ 676.231144] env[62066]: ERROR nova.compute.manager [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 676.231144] env[62066]: ERROR nova.compute.manager [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] for vif in network_info: [ 676.231144] env[62066]: ERROR nova.compute.manager [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 676.231144] env[62066]: ERROR nova.compute.manager [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] return self._sync_wrapper(fn, *args, **kwargs) [ 676.231144] env[62066]: ERROR nova.compute.manager [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 676.231144] env[62066]: ERROR nova.compute.manager [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] self.wait() [ 676.231144] env[62066]: ERROR nova.compute.manager [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 676.231144] env[62066]: ERROR nova.compute.manager [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] self[:] = self._gt.wait() [ 676.231144] env[62066]: ERROR nova.compute.manager [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 676.231144] env[62066]: ERROR nova.compute.manager [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] return self._exit_event.wait() [ 676.231144] env[62066]: ERROR nova.compute.manager [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 676.231144] env[62066]: ERROR nova.compute.manager [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] current.throw(*self._exc) [ 676.232614] env[62066]: ERROR nova.compute.manager [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 676.232614] env[62066]: ERROR nova.compute.manager [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] result = function(*args, **kwargs) [ 676.232614] env[62066]: ERROR nova.compute.manager [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 676.232614] env[62066]: ERROR nova.compute.manager [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] return func(*args, **kwargs) [ 676.232614] env[62066]: ERROR nova.compute.manager [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 676.232614] env[62066]: ERROR nova.compute.manager [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] raise e [ 676.232614] env[62066]: ERROR nova.compute.manager [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 676.232614] env[62066]: ERROR nova.compute.manager [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] nwinfo = self.network_api.allocate_for_instance( [ 676.232614] env[62066]: ERROR nova.compute.manager [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 676.232614] env[62066]: ERROR nova.compute.manager [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] created_port_ids = self._update_ports_for_instance( [ 676.232614] env[62066]: ERROR nova.compute.manager [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 676.232614] env[62066]: ERROR nova.compute.manager [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] with excutils.save_and_reraise_exception(): [ 676.232614] env[62066]: ERROR nova.compute.manager [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 676.232985] env[62066]: ERROR nova.compute.manager [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] self.force_reraise() [ 676.232985] env[62066]: ERROR nova.compute.manager [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 676.232985] env[62066]: ERROR nova.compute.manager [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] raise self.value [ 676.232985] env[62066]: ERROR nova.compute.manager [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 676.232985] env[62066]: ERROR nova.compute.manager [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] updated_port = self._update_port( [ 676.232985] env[62066]: ERROR nova.compute.manager [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 676.232985] env[62066]: ERROR nova.compute.manager [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] _ensure_no_port_binding_failure(port) [ 676.232985] env[62066]: ERROR nova.compute.manager [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 676.232985] env[62066]: ERROR nova.compute.manager [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] raise exception.PortBindingFailed(port_id=port['id']) [ 676.232985] env[62066]: ERROR nova.compute.manager [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] nova.exception.PortBindingFailed: Binding failed for port 9380aec8-fb60-43d1-b0f9-2068fbbc6c28, please check neutron logs for more information. [ 676.232985] env[62066]: ERROR nova.compute.manager [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] [ 676.232985] env[62066]: INFO nova.compute.manager [None req-b96d7111-6e60-4fd2-a38a-3863242412fb tempest-ServerActionsTestOtherB-426172427 tempest-ServerActionsTestOtherB-426172427-project-member] [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] Terminating instance [ 676.233361] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b96d7111-6e60-4fd2-a38a-3863242412fb tempest-ServerActionsTestOtherB-426172427 tempest-ServerActionsTestOtherB-426172427-project-member] Acquiring lock "refresh_cache-e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 676.247965] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b5125342-7692-4177-9e88-9f6d8ef9f1dc tempest-ServersAaction247Test-1482016775 tempest-ServersAaction247Test-1482016775-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.091s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 676.250560] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3515d98a-2869-41bb-bf8e-e38d995a4d22 tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.211s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 676.252277] env[62066]: INFO nova.compute.claims [None req-3515d98a-2869-41bb-bf8e-e38d995a4d22 tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 676.278734] env[62066]: INFO nova.scheduler.client.report [None req-b5125342-7692-4177-9e88-9f6d8ef9f1dc tempest-ServersAaction247Test-1482016775 tempest-ServersAaction247Test-1482016775-project-member] Deleted allocations for instance 5d54c3fd-b239-4965-b7a3-5909e8de8bc0 [ 676.382016] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b2b06eaa-4237-476b-9167-ed8be81ba3b9 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 676.568135] env[62066]: DEBUG nova.network.neutron [req-66274895-b199-4f89-bb81-3f31e1eb18e8 req-6da99013-f757-402e-ac11-727c24da45b8 service nova] [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 676.667742] env[62066]: DEBUG nova.network.neutron [req-66274895-b199-4f89-bb81-3f31e1eb18e8 req-6da99013-f757-402e-ac11-727c24da45b8 service nova] [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 676.785727] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b5125342-7692-4177-9e88-9f6d8ef9f1dc tempest-ServersAaction247Test-1482016775 tempest-ServersAaction247Test-1482016775-project-member] Lock "5d54c3fd-b239-4965-b7a3-5909e8de8bc0" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 33.836s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 677.170344] env[62066]: DEBUG oslo_concurrency.lockutils [req-66274895-b199-4f89-bb81-3f31e1eb18e8 req-6da99013-f757-402e-ac11-727c24da45b8 service nova] Releasing lock "refresh_cache-e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 677.171544] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b96d7111-6e60-4fd2-a38a-3863242412fb tempest-ServerActionsTestOtherB-426172427 tempest-ServerActionsTestOtherB-426172427-project-member] Acquired lock "refresh_cache-e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 677.171544] env[62066]: DEBUG nova.network.neutron [None req-b96d7111-6e60-4fd2-a38a-3863242412fb tempest-ServerActionsTestOtherB-426172427 tempest-ServerActionsTestOtherB-426172427-project-member] [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 677.611039] env[62066]: DEBUG oslo_concurrency.lockutils [None req-816774ed-d63b-4dd8-94be-e4c70c8f10e7 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Acquiring lock "459f432b-e817-405f-9c3e-dfa5d957788c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 677.611277] env[62066]: DEBUG oslo_concurrency.lockutils [None req-816774ed-d63b-4dd8-94be-e4c70c8f10e7 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Lock "459f432b-e817-405f-9c3e-dfa5d957788c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 677.689750] env[62066]: DEBUG nova.network.neutron [None req-b96d7111-6e60-4fd2-a38a-3863242412fb tempest-ServerActionsTestOtherB-426172427 tempest-ServerActionsTestOtherB-426172427-project-member] [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 677.749453] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-588a2f43-641c-4a73-8b58-309e68bf998f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.764172] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae94b98f-a531-4c86-acb9-cb1ccccd3f01 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.795675] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c409e823-edca-4e87-aa93-4389d8e737cc {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.802954] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c3b5416-88b9-4b65-bfac-434643a9a758 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.807275] env[62066]: DEBUG nova.network.neutron [None req-b96d7111-6e60-4fd2-a38a-3863242412fb tempest-ServerActionsTestOtherB-426172427 tempest-ServerActionsTestOtherB-426172427-project-member] [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 677.817593] env[62066]: DEBUG nova.compute.provider_tree [None req-3515d98a-2869-41bb-bf8e-e38d995a4d22 tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 678.107282] env[62066]: DEBUG nova.compute.manager [req-c4dfec5a-ea11-4809-a8de-4c97d7b04f3d req-366f930c-d277-4b9c-89f9-c1e820814f40 service nova] [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] Received event network-vif-deleted-9380aec8-fb60-43d1-b0f9-2068fbbc6c28 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 678.309999] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b96d7111-6e60-4fd2-a38a-3863242412fb tempest-ServerActionsTestOtherB-426172427 tempest-ServerActionsTestOtherB-426172427-project-member] Releasing lock "refresh_cache-e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 678.310257] env[62066]: DEBUG nova.compute.manager [None req-b96d7111-6e60-4fd2-a38a-3863242412fb tempest-ServerActionsTestOtherB-426172427 tempest-ServerActionsTestOtherB-426172427-project-member] [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 678.310257] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-b96d7111-6e60-4fd2-a38a-3863242412fb tempest-ServerActionsTestOtherB-426172427 tempest-ServerActionsTestOtherB-426172427-project-member] [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 678.310921] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7f3566b1-95d0-418f-84ed-5bec9aaa37b1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.321122] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c56b4e99-5702-439d-a69f-ee0f76925c78 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.335823] env[62066]: DEBUG nova.scheduler.client.report [None req-3515d98a-2869-41bb-bf8e-e38d995a4d22 tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 678.351382] env[62066]: WARNING nova.virt.vmwareapi.vmops [None req-b96d7111-6e60-4fd2-a38a-3863242412fb tempest-ServerActionsTestOtherB-426172427 tempest-ServerActionsTestOtherB-426172427-project-member] [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8 could not be found. [ 678.351382] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-b96d7111-6e60-4fd2-a38a-3863242412fb tempest-ServerActionsTestOtherB-426172427 tempest-ServerActionsTestOtherB-426172427-project-member] [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 678.351382] env[62066]: INFO nova.compute.manager [None req-b96d7111-6e60-4fd2-a38a-3863242412fb tempest-ServerActionsTestOtherB-426172427 tempest-ServerActionsTestOtherB-426172427-project-member] [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] Took 0.04 seconds to destroy the instance on the hypervisor. [ 678.351591] env[62066]: DEBUG oslo.service.loopingcall [None req-b96d7111-6e60-4fd2-a38a-3863242412fb tempest-ServerActionsTestOtherB-426172427 tempest-ServerActionsTestOtherB-426172427-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 678.351734] env[62066]: DEBUG nova.compute.manager [-] [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 678.351734] env[62066]: DEBUG nova.network.neutron [-] [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 678.373515] env[62066]: DEBUG nova.network.neutron [-] [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 678.841927] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3515d98a-2869-41bb-bf8e-e38d995a4d22 tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.591s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 678.842059] env[62066]: DEBUG nova.compute.manager [None req-3515d98a-2869-41bb-bf8e-e38d995a4d22 tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 678.844549] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ef1737fc-6565-4a6f-bd10-02fc3b07f848 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.685s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 678.876688] env[62066]: DEBUG nova.network.neutron [-] [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 679.349525] env[62066]: DEBUG nova.compute.utils [None req-3515d98a-2869-41bb-bf8e-e38d995a4d22 tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 679.354026] env[62066]: DEBUG nova.compute.manager [None req-3515d98a-2869-41bb-bf8e-e38d995a4d22 tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 679.354259] env[62066]: DEBUG nova.network.neutron [None req-3515d98a-2869-41bb-bf8e-e38d995a4d22 tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 679.383662] env[62066]: INFO nova.compute.manager [-] [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] Took 1.03 seconds to deallocate network for instance. [ 679.387384] env[62066]: DEBUG nova.compute.claims [None req-b96d7111-6e60-4fd2-a38a-3863242412fb tempest-ServerActionsTestOtherB-426172427 tempest-ServerActionsTestOtherB-426172427-project-member] [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] Aborting claim: {{(pid=62066) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 679.387802] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b96d7111-6e60-4fd2-a38a-3863242412fb tempest-ServerActionsTestOtherB-426172427 tempest-ServerActionsTestOtherB-426172427-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 679.408498] env[62066]: DEBUG nova.policy [None req-3515d98a-2869-41bb-bf8e-e38d995a4d22 tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '16c8fcc7c0b140f48716a8a00da46a3b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a9b239f4cb3242d0bd1dcd98a2a3bfc0', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 679.759270] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-babf2841-4a0e-48b7-b6f2-9d2e10d9d401 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.767350] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a183403-704c-4a25-94df-d286f1c5e331 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.798081] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ed1b2a5-2153-4707-a6d7-63e08fdcd01e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.805093] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6ac06ec-fa30-4f49-8f76-2e25a6b09e16 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.818290] env[62066]: DEBUG nova.compute.provider_tree [None req-ef1737fc-6565-4a6f-bd10-02fc3b07f848 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 679.855318] env[62066]: DEBUG nova.compute.manager [None req-3515d98a-2869-41bb-bf8e-e38d995a4d22 tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 679.874958] env[62066]: DEBUG nova.network.neutron [None req-3515d98a-2869-41bb-bf8e-e38d995a4d22 tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] Successfully created port: e43c5325-d9ab-45b3-9e9e-e4736fb5735f {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 680.320898] env[62066]: DEBUG nova.scheduler.client.report [None req-ef1737fc-6565-4a6f-bd10-02fc3b07f848 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 680.829020] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ef1737fc-6565-4a6f-bd10-02fc3b07f848 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.982s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 680.829020] env[62066]: ERROR nova.compute.manager [None req-ef1737fc-6565-4a6f-bd10-02fc3b07f848 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port d8e26205-c295-47d1-b068-e154150e67e5, please check neutron logs for more information. [ 680.829020] env[62066]: ERROR nova.compute.manager [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] Traceback (most recent call last): [ 680.829020] env[62066]: ERROR nova.compute.manager [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 680.829020] env[62066]: ERROR nova.compute.manager [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] self.driver.spawn(context, instance, image_meta, [ 680.829020] env[62066]: ERROR nova.compute.manager [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 680.829020] env[62066]: ERROR nova.compute.manager [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 680.829020] env[62066]: ERROR nova.compute.manager [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 680.829020] env[62066]: ERROR nova.compute.manager [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] vm_ref = self.build_virtual_machine(instance, [ 680.829535] env[62066]: ERROR nova.compute.manager [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 680.829535] env[62066]: ERROR nova.compute.manager [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] vif_infos = vmwarevif.get_vif_info(self._session, [ 680.829535] env[62066]: ERROR nova.compute.manager [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 680.829535] env[62066]: ERROR nova.compute.manager [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] for vif in network_info: [ 680.829535] env[62066]: ERROR nova.compute.manager [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 680.829535] env[62066]: ERROR nova.compute.manager [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] return self._sync_wrapper(fn, *args, **kwargs) [ 680.829535] env[62066]: ERROR nova.compute.manager [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 680.829535] env[62066]: ERROR nova.compute.manager [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] self.wait() [ 680.829535] env[62066]: ERROR nova.compute.manager [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 680.829535] env[62066]: ERROR nova.compute.manager [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] self[:] = self._gt.wait() [ 680.829535] env[62066]: ERROR nova.compute.manager [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 680.829535] env[62066]: ERROR nova.compute.manager [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] return self._exit_event.wait() [ 680.829535] env[62066]: ERROR nova.compute.manager [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 680.829940] env[62066]: ERROR nova.compute.manager [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] result = hub.switch() [ 680.829940] env[62066]: ERROR nova.compute.manager [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 680.829940] env[62066]: ERROR nova.compute.manager [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] return self.greenlet.switch() [ 680.829940] env[62066]: ERROR nova.compute.manager [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 680.829940] env[62066]: ERROR nova.compute.manager [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] result = function(*args, **kwargs) [ 680.829940] env[62066]: ERROR nova.compute.manager [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 680.829940] env[62066]: ERROR nova.compute.manager [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] return func(*args, **kwargs) [ 680.829940] env[62066]: ERROR nova.compute.manager [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 680.829940] env[62066]: ERROR nova.compute.manager [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] raise e [ 680.829940] env[62066]: ERROR nova.compute.manager [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 680.829940] env[62066]: ERROR nova.compute.manager [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] nwinfo = self.network_api.allocate_for_instance( [ 680.829940] env[62066]: ERROR nova.compute.manager [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 680.829940] env[62066]: ERROR nova.compute.manager [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] created_port_ids = self._update_ports_for_instance( [ 680.830294] env[62066]: ERROR nova.compute.manager [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 680.830294] env[62066]: ERROR nova.compute.manager [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] with excutils.save_and_reraise_exception(): [ 680.830294] env[62066]: ERROR nova.compute.manager [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 680.830294] env[62066]: ERROR nova.compute.manager [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] self.force_reraise() [ 680.830294] env[62066]: ERROR nova.compute.manager [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 680.830294] env[62066]: ERROR nova.compute.manager [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] raise self.value [ 680.830294] env[62066]: ERROR nova.compute.manager [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 680.830294] env[62066]: ERROR nova.compute.manager [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] updated_port = self._update_port( [ 680.830294] env[62066]: ERROR nova.compute.manager [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 680.830294] env[62066]: ERROR nova.compute.manager [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] _ensure_no_port_binding_failure(port) [ 680.830294] env[62066]: ERROR nova.compute.manager [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 680.830294] env[62066]: ERROR nova.compute.manager [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] raise exception.PortBindingFailed(port_id=port['id']) [ 680.830612] env[62066]: ERROR nova.compute.manager [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] nova.exception.PortBindingFailed: Binding failed for port d8e26205-c295-47d1-b068-e154150e67e5, please check neutron logs for more information. [ 680.830612] env[62066]: ERROR nova.compute.manager [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] [ 680.830612] env[62066]: DEBUG nova.compute.utils [None req-ef1737fc-6565-4a6f-bd10-02fc3b07f848 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] Binding failed for port d8e26205-c295-47d1-b068-e154150e67e5, please check neutron logs for more information. {{(pid=62066) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 680.834783] env[62066]: DEBUG oslo_concurrency.lockutils [None req-60a20a80-792e-4a03-84da-c17f8b28425d tempest-InstanceActionsNegativeTestJSON-1895685779 tempest-InstanceActionsNegativeTestJSON-1895685779-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.369s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 680.834783] env[62066]: INFO nova.compute.claims [None req-60a20a80-792e-4a03-84da-c17f8b28425d tempest-InstanceActionsNegativeTestJSON-1895685779 tempest-InstanceActionsNegativeTestJSON-1895685779-project-member] [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 680.840474] env[62066]: DEBUG nova.compute.manager [None req-ef1737fc-6565-4a6f-bd10-02fc3b07f848 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] Build of instance 97e51e6c-9a3a-4b68-b737-d48090a22b1b was re-scheduled: Binding failed for port d8e26205-c295-47d1-b068-e154150e67e5, please check neutron logs for more information. {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 680.840474] env[62066]: DEBUG nova.compute.manager [None req-ef1737fc-6565-4a6f-bd10-02fc3b07f848 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] Unplugging VIFs for instance {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 680.840474] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ef1737fc-6565-4a6f-bd10-02fc3b07f848 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Acquiring lock "refresh_cache-97e51e6c-9a3a-4b68-b737-d48090a22b1b" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 680.840474] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ef1737fc-6565-4a6f-bd10-02fc3b07f848 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Acquired lock "refresh_cache-97e51e6c-9a3a-4b68-b737-d48090a22b1b" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 680.840474] env[62066]: DEBUG nova.network.neutron [None req-ef1737fc-6565-4a6f-bd10-02fc3b07f848 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 680.841845] env[62066]: DEBUG nova.compute.manager [req-e21f211e-1c73-434f-a1bc-9fb09998cf8c req-514a3578-7a2f-49b4-8060-75e41b50f73c service nova] [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] Received event network-changed-e43c5325-d9ab-45b3-9e9e-e4736fb5735f {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 680.842082] env[62066]: DEBUG nova.compute.manager [req-e21f211e-1c73-434f-a1bc-9fb09998cf8c req-514a3578-7a2f-49b4-8060-75e41b50f73c service nova] [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] Refreshing instance network info cache due to event network-changed-e43c5325-d9ab-45b3-9e9e-e4736fb5735f. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 680.842420] env[62066]: DEBUG oslo_concurrency.lockutils [req-e21f211e-1c73-434f-a1bc-9fb09998cf8c req-514a3578-7a2f-49b4-8060-75e41b50f73c service nova] Acquiring lock "refresh_cache-e795e53e-af33-4d0c-8dbd-9c59f236acd7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 680.842613] env[62066]: DEBUG oslo_concurrency.lockutils [req-e21f211e-1c73-434f-a1bc-9fb09998cf8c req-514a3578-7a2f-49b4-8060-75e41b50f73c service nova] Acquired lock "refresh_cache-e795e53e-af33-4d0c-8dbd-9c59f236acd7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 680.843356] env[62066]: DEBUG nova.network.neutron [req-e21f211e-1c73-434f-a1bc-9fb09998cf8c req-514a3578-7a2f-49b4-8060-75e41b50f73c service nova] [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] Refreshing network info cache for port e43c5325-d9ab-45b3-9e9e-e4736fb5735f {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 680.869521] env[62066]: DEBUG nova.compute.manager [None req-3515d98a-2869-41bb-bf8e-e38d995a4d22 tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 680.903208] env[62066]: DEBUG nova.virt.hardware [None req-3515d98a-2869-41bb-bf8e-e38d995a4d22 tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-23T13:40:41Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-23T13:40:23Z,direct_url=,disk_format='vmdk',id=50ff584c-3b50-4395-af07-3e66769bc9f7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='eb52f7069a374c61ae946f052007c6d9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-23T13:40:24Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 680.903358] env[62066]: DEBUG nova.virt.hardware [None req-3515d98a-2869-41bb-bf8e-e38d995a4d22 tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 680.903433] env[62066]: DEBUG nova.virt.hardware [None req-3515d98a-2869-41bb-bf8e-e38d995a4d22 tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 680.903611] env[62066]: DEBUG nova.virt.hardware [None req-3515d98a-2869-41bb-bf8e-e38d995a4d22 tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 680.903751] env[62066]: DEBUG nova.virt.hardware [None req-3515d98a-2869-41bb-bf8e-e38d995a4d22 tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 680.903891] env[62066]: DEBUG nova.virt.hardware [None req-3515d98a-2869-41bb-bf8e-e38d995a4d22 tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 680.904484] env[62066]: DEBUG nova.virt.hardware [None req-3515d98a-2869-41bb-bf8e-e38d995a4d22 tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 680.905045] env[62066]: DEBUG nova.virt.hardware [None req-3515d98a-2869-41bb-bf8e-e38d995a4d22 tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 680.905318] env[62066]: DEBUG nova.virt.hardware [None req-3515d98a-2869-41bb-bf8e-e38d995a4d22 tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 680.905543] env[62066]: DEBUG nova.virt.hardware [None req-3515d98a-2869-41bb-bf8e-e38d995a4d22 tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 680.906605] env[62066]: DEBUG nova.virt.hardware [None req-3515d98a-2869-41bb-bf8e-e38d995a4d22 tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 680.907868] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48dc2ab9-1198-4dc7-9609-a6b498f7c904 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.919022] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-528497ab-f3dc-4141-a052-03b502aed74c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.986547] env[62066]: ERROR nova.compute.manager [None req-3515d98a-2869-41bb-bf8e-e38d995a4d22 tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port e43c5325-d9ab-45b3-9e9e-e4736fb5735f, please check neutron logs for more information. [ 680.986547] env[62066]: ERROR nova.compute.manager Traceback (most recent call last): [ 680.986547] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 680.986547] env[62066]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 680.986547] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 680.986547] env[62066]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 680.986547] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 680.986547] env[62066]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 680.986547] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 680.986547] env[62066]: ERROR nova.compute.manager self.force_reraise() [ 680.986547] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 680.986547] env[62066]: ERROR nova.compute.manager raise self.value [ 680.986547] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 680.986547] env[62066]: ERROR nova.compute.manager updated_port = self._update_port( [ 680.986547] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 680.986547] env[62066]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 680.987112] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 680.987112] env[62066]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 680.987112] env[62066]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port e43c5325-d9ab-45b3-9e9e-e4736fb5735f, please check neutron logs for more information. [ 680.987112] env[62066]: ERROR nova.compute.manager [ 680.987112] env[62066]: Traceback (most recent call last): [ 680.987112] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 680.987112] env[62066]: listener.cb(fileno) [ 680.987112] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 680.987112] env[62066]: result = function(*args, **kwargs) [ 680.987112] env[62066]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 680.987112] env[62066]: return func(*args, **kwargs) [ 680.987112] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 680.987112] env[62066]: raise e [ 680.987112] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 680.987112] env[62066]: nwinfo = self.network_api.allocate_for_instance( [ 680.987112] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 680.987112] env[62066]: created_port_ids = self._update_ports_for_instance( [ 680.987112] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 680.987112] env[62066]: with excutils.save_and_reraise_exception(): [ 680.987112] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 680.987112] env[62066]: self.force_reraise() [ 680.987112] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 680.987112] env[62066]: raise self.value [ 680.987112] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 680.987112] env[62066]: updated_port = self._update_port( [ 680.987112] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 680.987112] env[62066]: _ensure_no_port_binding_failure(port) [ 680.987112] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 680.987112] env[62066]: raise exception.PortBindingFailed(port_id=port['id']) [ 680.987861] env[62066]: nova.exception.PortBindingFailed: Binding failed for port e43c5325-d9ab-45b3-9e9e-e4736fb5735f, please check neutron logs for more information. [ 680.987861] env[62066]: Removing descriptor: 15 [ 680.987861] env[62066]: ERROR nova.compute.manager [None req-3515d98a-2869-41bb-bf8e-e38d995a4d22 tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port e43c5325-d9ab-45b3-9e9e-e4736fb5735f, please check neutron logs for more information. [ 680.987861] env[62066]: ERROR nova.compute.manager [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] Traceback (most recent call last): [ 680.987861] env[62066]: ERROR nova.compute.manager [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 680.987861] env[62066]: ERROR nova.compute.manager [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] yield resources [ 680.987861] env[62066]: ERROR nova.compute.manager [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 680.987861] env[62066]: ERROR nova.compute.manager [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] self.driver.spawn(context, instance, image_meta, [ 680.987861] env[62066]: ERROR nova.compute.manager [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 680.987861] env[62066]: ERROR nova.compute.manager [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] self._vmops.spawn(context, instance, image_meta, injected_files, [ 680.987861] env[62066]: ERROR nova.compute.manager [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 680.987861] env[62066]: ERROR nova.compute.manager [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] vm_ref = self.build_virtual_machine(instance, [ 680.988253] env[62066]: ERROR nova.compute.manager [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 680.988253] env[62066]: ERROR nova.compute.manager [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] vif_infos = vmwarevif.get_vif_info(self._session, [ 680.988253] env[62066]: ERROR nova.compute.manager [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 680.988253] env[62066]: ERROR nova.compute.manager [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] for vif in network_info: [ 680.988253] env[62066]: ERROR nova.compute.manager [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 680.988253] env[62066]: ERROR nova.compute.manager [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] return self._sync_wrapper(fn, *args, **kwargs) [ 680.988253] env[62066]: ERROR nova.compute.manager [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 680.988253] env[62066]: ERROR nova.compute.manager [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] self.wait() [ 680.988253] env[62066]: ERROR nova.compute.manager [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 680.988253] env[62066]: ERROR nova.compute.manager [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] self[:] = self._gt.wait() [ 680.988253] env[62066]: ERROR nova.compute.manager [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 680.988253] env[62066]: ERROR nova.compute.manager [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] return self._exit_event.wait() [ 680.988253] env[62066]: ERROR nova.compute.manager [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 680.988602] env[62066]: ERROR nova.compute.manager [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] result = hub.switch() [ 680.988602] env[62066]: ERROR nova.compute.manager [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 680.988602] env[62066]: ERROR nova.compute.manager [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] return self.greenlet.switch() [ 680.988602] env[62066]: ERROR nova.compute.manager [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 680.988602] env[62066]: ERROR nova.compute.manager [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] result = function(*args, **kwargs) [ 680.988602] env[62066]: ERROR nova.compute.manager [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 680.988602] env[62066]: ERROR nova.compute.manager [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] return func(*args, **kwargs) [ 680.988602] env[62066]: ERROR nova.compute.manager [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 680.988602] env[62066]: ERROR nova.compute.manager [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] raise e [ 680.988602] env[62066]: ERROR nova.compute.manager [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 680.988602] env[62066]: ERROR nova.compute.manager [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] nwinfo = self.network_api.allocate_for_instance( [ 680.988602] env[62066]: ERROR nova.compute.manager [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 680.988602] env[62066]: ERROR nova.compute.manager [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] created_port_ids = self._update_ports_for_instance( [ 680.989008] env[62066]: ERROR nova.compute.manager [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 680.989008] env[62066]: ERROR nova.compute.manager [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] with excutils.save_and_reraise_exception(): [ 680.989008] env[62066]: ERROR nova.compute.manager [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 680.989008] env[62066]: ERROR nova.compute.manager [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] self.force_reraise() [ 680.989008] env[62066]: ERROR nova.compute.manager [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 680.989008] env[62066]: ERROR nova.compute.manager [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] raise self.value [ 680.989008] env[62066]: ERROR nova.compute.manager [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 680.989008] env[62066]: ERROR nova.compute.manager [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] updated_port = self._update_port( [ 680.989008] env[62066]: ERROR nova.compute.manager [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 680.989008] env[62066]: ERROR nova.compute.manager [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] _ensure_no_port_binding_failure(port) [ 680.989008] env[62066]: ERROR nova.compute.manager [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 680.989008] env[62066]: ERROR nova.compute.manager [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] raise exception.PortBindingFailed(port_id=port['id']) [ 680.989472] env[62066]: ERROR nova.compute.manager [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] nova.exception.PortBindingFailed: Binding failed for port e43c5325-d9ab-45b3-9e9e-e4736fb5735f, please check neutron logs for more information. [ 680.989472] env[62066]: ERROR nova.compute.manager [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] [ 680.989472] env[62066]: INFO nova.compute.manager [None req-3515d98a-2869-41bb-bf8e-e38d995a4d22 tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] Terminating instance [ 680.991125] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3515d98a-2869-41bb-bf8e-e38d995a4d22 tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Acquiring lock "refresh_cache-e795e53e-af33-4d0c-8dbd-9c59f236acd7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 681.367860] env[62066]: DEBUG nova.network.neutron [None req-ef1737fc-6565-4a6f-bd10-02fc3b07f848 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 681.382682] env[62066]: DEBUG nova.network.neutron [req-e21f211e-1c73-434f-a1bc-9fb09998cf8c req-514a3578-7a2f-49b4-8060-75e41b50f73c service nova] [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 681.455727] env[62066]: DEBUG nova.network.neutron [None req-ef1737fc-6565-4a6f-bd10-02fc3b07f848 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 681.577780] env[62066]: DEBUG nova.network.neutron [req-e21f211e-1c73-434f-a1bc-9fb09998cf8c req-514a3578-7a2f-49b4-8060-75e41b50f73c service nova] [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 681.958094] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ef1737fc-6565-4a6f-bd10-02fc3b07f848 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Releasing lock "refresh_cache-97e51e6c-9a3a-4b68-b737-d48090a22b1b" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 681.958926] env[62066]: DEBUG nova.compute.manager [None req-ef1737fc-6565-4a6f-bd10-02fc3b07f848 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 681.958926] env[62066]: DEBUG nova.compute.manager [None req-ef1737fc-6565-4a6f-bd10-02fc3b07f848 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 681.958926] env[62066]: DEBUG nova.network.neutron [None req-ef1737fc-6565-4a6f-bd10-02fc3b07f848 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 681.981202] env[62066]: DEBUG nova.network.neutron [None req-ef1737fc-6565-4a6f-bd10-02fc3b07f848 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 682.080275] env[62066]: DEBUG oslo_concurrency.lockutils [req-e21f211e-1c73-434f-a1bc-9fb09998cf8c req-514a3578-7a2f-49b4-8060-75e41b50f73c service nova] Releasing lock "refresh_cache-e795e53e-af33-4d0c-8dbd-9c59f236acd7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 682.080742] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3515d98a-2869-41bb-bf8e-e38d995a4d22 tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Acquired lock "refresh_cache-e795e53e-af33-4d0c-8dbd-9c59f236acd7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 682.080924] env[62066]: DEBUG nova.network.neutron [None req-3515d98a-2869-41bb-bf8e-e38d995a4d22 tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 682.263576] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bbbc101-1444-4709-9697-997eec6793dd {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.275922] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fbca328-cfc9-4c41-971c-a607f6d07abb {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.311135] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8493b544-af56-4a61-95fd-23bf8f868378 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.318602] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74d5fdb0-6084-4eb2-81da-2598a3a77986 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.334120] env[62066]: DEBUG nova.compute.provider_tree [None req-60a20a80-792e-4a03-84da-c17f8b28425d tempest-InstanceActionsNegativeTestJSON-1895685779 tempest-InstanceActionsNegativeTestJSON-1895685779-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 682.483815] env[62066]: DEBUG nova.network.neutron [None req-ef1737fc-6565-4a6f-bd10-02fc3b07f848 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 682.602381] env[62066]: DEBUG nova.network.neutron [None req-3515d98a-2869-41bb-bf8e-e38d995a4d22 tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 682.689529] env[62066]: DEBUG nova.network.neutron [None req-3515d98a-2869-41bb-bf8e-e38d995a4d22 tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 682.838737] env[62066]: DEBUG nova.scheduler.client.report [None req-60a20a80-792e-4a03-84da-c17f8b28425d tempest-InstanceActionsNegativeTestJSON-1895685779 tempest-InstanceActionsNegativeTestJSON-1895685779-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 682.873826] env[62066]: DEBUG nova.compute.manager [req-c37a4161-92c6-49c5-b2d1-e8ebc6b68924 req-be169a68-5a1e-4b94-bf2e-49201dfd8cae service nova] [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] Received event network-vif-deleted-e43c5325-d9ab-45b3-9e9e-e4736fb5735f {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 682.877332] env[62066]: DEBUG oslo_concurrency.lockutils [None req-febd07a3-de81-4924-888e-f9c7d8ba2329 tempest-ServerMetadataNegativeTestJSON-1549470887 tempest-ServerMetadataNegativeTestJSON-1549470887-project-member] Acquiring lock "31d47299-83a3-4f27-aeb7-95c4cd36c5ac" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 682.877332] env[62066]: DEBUG oslo_concurrency.lockutils [None req-febd07a3-de81-4924-888e-f9c7d8ba2329 tempest-ServerMetadataNegativeTestJSON-1549470887 tempest-ServerMetadataNegativeTestJSON-1549470887-project-member] Lock "31d47299-83a3-4f27-aeb7-95c4cd36c5ac" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 682.987599] env[62066]: INFO nova.compute.manager [None req-ef1737fc-6565-4a6f-bd10-02fc3b07f848 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 97e51e6c-9a3a-4b68-b737-d48090a22b1b] Took 1.03 seconds to deallocate network for instance. [ 683.192433] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3515d98a-2869-41bb-bf8e-e38d995a4d22 tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Releasing lock "refresh_cache-e795e53e-af33-4d0c-8dbd-9c59f236acd7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 683.192869] env[62066]: DEBUG nova.compute.manager [None req-3515d98a-2869-41bb-bf8e-e38d995a4d22 tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 683.193079] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-3515d98a-2869-41bb-bf8e-e38d995a4d22 tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 683.193796] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c81d2b33-f91c-4015-b887-1e1b4f434525 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.203252] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-805bd811-7c22-456b-beda-0b964518bc5f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.224443] env[62066]: WARNING nova.virt.vmwareapi.vmops [None req-3515d98a-2869-41bb-bf8e-e38d995a4d22 tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance e795e53e-af33-4d0c-8dbd-9c59f236acd7 could not be found. [ 683.224755] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-3515d98a-2869-41bb-bf8e-e38d995a4d22 tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 683.224964] env[62066]: INFO nova.compute.manager [None req-3515d98a-2869-41bb-bf8e-e38d995a4d22 tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] Took 0.03 seconds to destroy the instance on the hypervisor. [ 683.225226] env[62066]: DEBUG oslo.service.loopingcall [None req-3515d98a-2869-41bb-bf8e-e38d995a4d22 tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 683.225508] env[62066]: DEBUG nova.compute.manager [-] [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 683.225627] env[62066]: DEBUG nova.network.neutron [-] [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 683.248391] env[62066]: DEBUG nova.network.neutron [-] [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 683.351202] env[62066]: DEBUG oslo_concurrency.lockutils [None req-60a20a80-792e-4a03-84da-c17f8b28425d tempest-InstanceActionsNegativeTestJSON-1895685779 tempest-InstanceActionsNegativeTestJSON-1895685779-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.516s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 683.351202] env[62066]: DEBUG nova.compute.manager [None req-60a20a80-792e-4a03-84da-c17f8b28425d tempest-InstanceActionsNegativeTestJSON-1895685779 tempest-InstanceActionsNegativeTestJSON-1895685779-project-member] [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 683.352351] env[62066]: DEBUG oslo_concurrency.lockutils [None req-19be2ac0-adbe-403f-ab5a-6d1895be1574 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.724s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 683.356242] env[62066]: INFO nova.compute.claims [None req-19be2ac0-adbe-403f-ab5a-6d1895be1574 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 683.751135] env[62066]: DEBUG nova.network.neutron [-] [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 683.853795] env[62066]: DEBUG nova.compute.utils [None req-60a20a80-792e-4a03-84da-c17f8b28425d tempest-InstanceActionsNegativeTestJSON-1895685779 tempest-InstanceActionsNegativeTestJSON-1895685779-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 683.855267] env[62066]: DEBUG nova.compute.manager [None req-60a20a80-792e-4a03-84da-c17f8b28425d tempest-InstanceActionsNegativeTestJSON-1895685779 tempest-InstanceActionsNegativeTestJSON-1895685779-project-member] [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 683.855373] env[62066]: DEBUG nova.network.neutron [None req-60a20a80-792e-4a03-84da-c17f8b28425d tempest-InstanceActionsNegativeTestJSON-1895685779 tempest-InstanceActionsNegativeTestJSON-1895685779-project-member] [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 683.920967] env[62066]: DEBUG nova.policy [None req-60a20a80-792e-4a03-84da-c17f8b28425d tempest-InstanceActionsNegativeTestJSON-1895685779 tempest-InstanceActionsNegativeTestJSON-1895685779-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '113a1bfba03a4105adc913e46c6eef40', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8cae4b68afcf46448d05d484906f48de', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 684.018575] env[62066]: INFO nova.scheduler.client.report [None req-ef1737fc-6565-4a6f-bd10-02fc3b07f848 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Deleted allocations for instance 97e51e6c-9a3a-4b68-b737-d48090a22b1b [ 684.253471] env[62066]: INFO nova.compute.manager [-] [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] Took 1.03 seconds to deallocate network for instance. [ 684.256485] env[62066]: DEBUG nova.compute.claims [None req-3515d98a-2869-41bb-bf8e-e38d995a4d22 tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] Aborting claim: {{(pid=62066) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 684.256693] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3515d98a-2869-41bb-bf8e-e38d995a4d22 tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 684.361133] env[62066]: DEBUG nova.compute.manager [None req-60a20a80-792e-4a03-84da-c17f8b28425d tempest-InstanceActionsNegativeTestJSON-1895685779 tempest-InstanceActionsNegativeTestJSON-1895685779-project-member] [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 684.436136] env[62066]: DEBUG nova.network.neutron [None req-60a20a80-792e-4a03-84da-c17f8b28425d tempest-InstanceActionsNegativeTestJSON-1895685779 tempest-InstanceActionsNegativeTestJSON-1895685779-project-member] [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] Successfully created port: df0ac18e-7f9a-44a1-98ac-693de6d5fa5f {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 684.527597] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ef1737fc-6565-4a6f-bd10-02fc3b07f848 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Lock "97e51e6c-9a3a-4b68-b737-d48090a22b1b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 145.620s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 684.778267] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf618da9-a7e0-4d20-b7c8-c39075e63cf6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.787248] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86bf4fab-04c4-4ca7-b320-5a40c25e00c3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.822391] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67a8c09b-5d77-43bd-b0a8-c861da7fd349 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.830798] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-850fd598-5b46-4cde-a71a-5e1ac5028575 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.844996] env[62066]: DEBUG nova.compute.provider_tree [None req-19be2ac0-adbe-403f-ab5a-6d1895be1574 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 685.033926] env[62066]: DEBUG nova.compute.manager [None req-84147b21-d606-4daf-b0b9-f581072be1cb tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 685.348135] env[62066]: DEBUG nova.scheduler.client.report [None req-19be2ac0-adbe-403f-ab5a-6d1895be1574 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 685.375691] env[62066]: DEBUG nova.compute.manager [None req-60a20a80-792e-4a03-84da-c17f8b28425d tempest-InstanceActionsNegativeTestJSON-1895685779 tempest-InstanceActionsNegativeTestJSON-1895685779-project-member] [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 685.397254] env[62066]: DEBUG nova.virt.hardware [None req-60a20a80-792e-4a03-84da-c17f8b28425d tempest-InstanceActionsNegativeTestJSON-1895685779 tempest-InstanceActionsNegativeTestJSON-1895685779-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-23T13:40:41Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-23T13:40:23Z,direct_url=,disk_format='vmdk',id=50ff584c-3b50-4395-af07-3e66769bc9f7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='eb52f7069a374c61ae946f052007c6d9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-23T13:40:24Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 685.397254] env[62066]: DEBUG nova.virt.hardware [None req-60a20a80-792e-4a03-84da-c17f8b28425d tempest-InstanceActionsNegativeTestJSON-1895685779 tempest-InstanceActionsNegativeTestJSON-1895685779-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 685.397439] env[62066]: DEBUG nova.virt.hardware [None req-60a20a80-792e-4a03-84da-c17f8b28425d tempest-InstanceActionsNegativeTestJSON-1895685779 tempest-InstanceActionsNegativeTestJSON-1895685779-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 685.397819] env[62066]: DEBUG nova.virt.hardware [None req-60a20a80-792e-4a03-84da-c17f8b28425d tempest-InstanceActionsNegativeTestJSON-1895685779 tempest-InstanceActionsNegativeTestJSON-1895685779-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 685.397819] env[62066]: DEBUG nova.virt.hardware [None req-60a20a80-792e-4a03-84da-c17f8b28425d tempest-InstanceActionsNegativeTestJSON-1895685779 tempest-InstanceActionsNegativeTestJSON-1895685779-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 685.397819] env[62066]: DEBUG nova.virt.hardware [None req-60a20a80-792e-4a03-84da-c17f8b28425d tempest-InstanceActionsNegativeTestJSON-1895685779 tempest-InstanceActionsNegativeTestJSON-1895685779-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 685.398021] env[62066]: DEBUG nova.virt.hardware [None req-60a20a80-792e-4a03-84da-c17f8b28425d tempest-InstanceActionsNegativeTestJSON-1895685779 tempest-InstanceActionsNegativeTestJSON-1895685779-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 685.398198] env[62066]: DEBUG nova.virt.hardware [None req-60a20a80-792e-4a03-84da-c17f8b28425d tempest-InstanceActionsNegativeTestJSON-1895685779 tempest-InstanceActionsNegativeTestJSON-1895685779-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 685.398377] env[62066]: DEBUG nova.virt.hardware [None req-60a20a80-792e-4a03-84da-c17f8b28425d tempest-InstanceActionsNegativeTestJSON-1895685779 tempest-InstanceActionsNegativeTestJSON-1895685779-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 685.398543] env[62066]: DEBUG nova.virt.hardware [None req-60a20a80-792e-4a03-84da-c17f8b28425d tempest-InstanceActionsNegativeTestJSON-1895685779 tempest-InstanceActionsNegativeTestJSON-1895685779-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 685.398713] env[62066]: DEBUG nova.virt.hardware [None req-60a20a80-792e-4a03-84da-c17f8b28425d tempest-InstanceActionsNegativeTestJSON-1895685779 tempest-InstanceActionsNegativeTestJSON-1895685779-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 685.399852] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3007f1d8-f6cb-454b-8c5f-7e2ce3279d04 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.407310] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b1f72e7-cf9c-4652-9e45-5c2e0dcc1da6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.566645] env[62066]: DEBUG oslo_concurrency.lockutils [None req-84147b21-d606-4daf-b0b9-f581072be1cb tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 685.753139] env[62066]: DEBUG nova.compute.manager [req-48b1c930-aa92-480a-962b-956c877bd190 req-9904d57d-2136-414c-b64a-a16c3c5c2eee service nova] [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] Received event network-changed-df0ac18e-7f9a-44a1-98ac-693de6d5fa5f {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 685.753139] env[62066]: DEBUG nova.compute.manager [req-48b1c930-aa92-480a-962b-956c877bd190 req-9904d57d-2136-414c-b64a-a16c3c5c2eee service nova] [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] Refreshing instance network info cache due to event network-changed-df0ac18e-7f9a-44a1-98ac-693de6d5fa5f. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 685.753808] env[62066]: DEBUG oslo_concurrency.lockutils [req-48b1c930-aa92-480a-962b-956c877bd190 req-9904d57d-2136-414c-b64a-a16c3c5c2eee service nova] Acquiring lock "refresh_cache-a690ad15-2878-4fa7-a810-d155ce6b1a8c" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 685.754264] env[62066]: DEBUG oslo_concurrency.lockutils [req-48b1c930-aa92-480a-962b-956c877bd190 req-9904d57d-2136-414c-b64a-a16c3c5c2eee service nova] Acquired lock "refresh_cache-a690ad15-2878-4fa7-a810-d155ce6b1a8c" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 685.754997] env[62066]: DEBUG nova.network.neutron [req-48b1c930-aa92-480a-962b-956c877bd190 req-9904d57d-2136-414c-b64a-a16c3c5c2eee service nova] [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] Refreshing network info cache for port df0ac18e-7f9a-44a1-98ac-693de6d5fa5f {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 685.827646] env[62066]: ERROR nova.compute.manager [None req-60a20a80-792e-4a03-84da-c17f8b28425d tempest-InstanceActionsNegativeTestJSON-1895685779 tempest-InstanceActionsNegativeTestJSON-1895685779-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port df0ac18e-7f9a-44a1-98ac-693de6d5fa5f, please check neutron logs for more information. [ 685.827646] env[62066]: ERROR nova.compute.manager Traceback (most recent call last): [ 685.827646] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 685.827646] env[62066]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 685.827646] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 685.827646] env[62066]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 685.827646] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 685.827646] env[62066]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 685.827646] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 685.827646] env[62066]: ERROR nova.compute.manager self.force_reraise() [ 685.827646] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 685.827646] env[62066]: ERROR nova.compute.manager raise self.value [ 685.827646] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 685.827646] env[62066]: ERROR nova.compute.manager updated_port = self._update_port( [ 685.827646] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 685.827646] env[62066]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 685.828243] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 685.828243] env[62066]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 685.828243] env[62066]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port df0ac18e-7f9a-44a1-98ac-693de6d5fa5f, please check neutron logs for more information. [ 685.828243] env[62066]: ERROR nova.compute.manager [ 685.828243] env[62066]: Traceback (most recent call last): [ 685.828243] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 685.828243] env[62066]: listener.cb(fileno) [ 685.828243] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 685.828243] env[62066]: result = function(*args, **kwargs) [ 685.828243] env[62066]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 685.828243] env[62066]: return func(*args, **kwargs) [ 685.828243] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 685.828243] env[62066]: raise e [ 685.828243] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 685.828243] env[62066]: nwinfo = self.network_api.allocate_for_instance( [ 685.828243] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 685.828243] env[62066]: created_port_ids = self._update_ports_for_instance( [ 685.828243] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 685.828243] env[62066]: with excutils.save_and_reraise_exception(): [ 685.828243] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 685.828243] env[62066]: self.force_reraise() [ 685.828243] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 685.828243] env[62066]: raise self.value [ 685.828243] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 685.828243] env[62066]: updated_port = self._update_port( [ 685.828243] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 685.828243] env[62066]: _ensure_no_port_binding_failure(port) [ 685.828243] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 685.828243] env[62066]: raise exception.PortBindingFailed(port_id=port['id']) [ 685.829272] env[62066]: nova.exception.PortBindingFailed: Binding failed for port df0ac18e-7f9a-44a1-98ac-693de6d5fa5f, please check neutron logs for more information. [ 685.829272] env[62066]: Removing descriptor: 20 [ 685.829272] env[62066]: ERROR nova.compute.manager [None req-60a20a80-792e-4a03-84da-c17f8b28425d tempest-InstanceActionsNegativeTestJSON-1895685779 tempest-InstanceActionsNegativeTestJSON-1895685779-project-member] [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port df0ac18e-7f9a-44a1-98ac-693de6d5fa5f, please check neutron logs for more information. [ 685.829272] env[62066]: ERROR nova.compute.manager [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] Traceback (most recent call last): [ 685.829272] env[62066]: ERROR nova.compute.manager [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 685.829272] env[62066]: ERROR nova.compute.manager [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] yield resources [ 685.829272] env[62066]: ERROR nova.compute.manager [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 685.829272] env[62066]: ERROR nova.compute.manager [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] self.driver.spawn(context, instance, image_meta, [ 685.829272] env[62066]: ERROR nova.compute.manager [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 685.829272] env[62066]: ERROR nova.compute.manager [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 685.829272] env[62066]: ERROR nova.compute.manager [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 685.829272] env[62066]: ERROR nova.compute.manager [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] vm_ref = self.build_virtual_machine(instance, [ 685.829790] env[62066]: ERROR nova.compute.manager [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 685.829790] env[62066]: ERROR nova.compute.manager [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] vif_infos = vmwarevif.get_vif_info(self._session, [ 685.829790] env[62066]: ERROR nova.compute.manager [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 685.829790] env[62066]: ERROR nova.compute.manager [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] for vif in network_info: [ 685.829790] env[62066]: ERROR nova.compute.manager [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 685.829790] env[62066]: ERROR nova.compute.manager [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] return self._sync_wrapper(fn, *args, **kwargs) [ 685.829790] env[62066]: ERROR nova.compute.manager [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 685.829790] env[62066]: ERROR nova.compute.manager [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] self.wait() [ 685.829790] env[62066]: ERROR nova.compute.manager [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 685.829790] env[62066]: ERROR nova.compute.manager [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] self[:] = self._gt.wait() [ 685.829790] env[62066]: ERROR nova.compute.manager [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 685.829790] env[62066]: ERROR nova.compute.manager [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] return self._exit_event.wait() [ 685.829790] env[62066]: ERROR nova.compute.manager [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 685.830240] env[62066]: ERROR nova.compute.manager [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] result = hub.switch() [ 685.830240] env[62066]: ERROR nova.compute.manager [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 685.830240] env[62066]: ERROR nova.compute.manager [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] return self.greenlet.switch() [ 685.830240] env[62066]: ERROR nova.compute.manager [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 685.830240] env[62066]: ERROR nova.compute.manager [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] result = function(*args, **kwargs) [ 685.830240] env[62066]: ERROR nova.compute.manager [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 685.830240] env[62066]: ERROR nova.compute.manager [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] return func(*args, **kwargs) [ 685.830240] env[62066]: ERROR nova.compute.manager [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 685.830240] env[62066]: ERROR nova.compute.manager [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] raise e [ 685.830240] env[62066]: ERROR nova.compute.manager [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 685.830240] env[62066]: ERROR nova.compute.manager [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] nwinfo = self.network_api.allocate_for_instance( [ 685.830240] env[62066]: ERROR nova.compute.manager [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 685.830240] env[62066]: ERROR nova.compute.manager [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] created_port_ids = self._update_ports_for_instance( [ 685.830662] env[62066]: ERROR nova.compute.manager [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 685.830662] env[62066]: ERROR nova.compute.manager [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] with excutils.save_and_reraise_exception(): [ 685.830662] env[62066]: ERROR nova.compute.manager [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 685.830662] env[62066]: ERROR nova.compute.manager [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] self.force_reraise() [ 685.830662] env[62066]: ERROR nova.compute.manager [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 685.830662] env[62066]: ERROR nova.compute.manager [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] raise self.value [ 685.830662] env[62066]: ERROR nova.compute.manager [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 685.830662] env[62066]: ERROR nova.compute.manager [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] updated_port = self._update_port( [ 685.830662] env[62066]: ERROR nova.compute.manager [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 685.830662] env[62066]: ERROR nova.compute.manager [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] _ensure_no_port_binding_failure(port) [ 685.830662] env[62066]: ERROR nova.compute.manager [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 685.830662] env[62066]: ERROR nova.compute.manager [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] raise exception.PortBindingFailed(port_id=port['id']) [ 685.831133] env[62066]: ERROR nova.compute.manager [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] nova.exception.PortBindingFailed: Binding failed for port df0ac18e-7f9a-44a1-98ac-693de6d5fa5f, please check neutron logs for more information. [ 685.831133] env[62066]: ERROR nova.compute.manager [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] [ 685.831133] env[62066]: INFO nova.compute.manager [None req-60a20a80-792e-4a03-84da-c17f8b28425d tempest-InstanceActionsNegativeTestJSON-1895685779 tempest-InstanceActionsNegativeTestJSON-1895685779-project-member] [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] Terminating instance [ 685.831133] env[62066]: DEBUG oslo_concurrency.lockutils [None req-60a20a80-792e-4a03-84da-c17f8b28425d tempest-InstanceActionsNegativeTestJSON-1895685779 tempest-InstanceActionsNegativeTestJSON-1895685779-project-member] Acquiring lock "refresh_cache-a690ad15-2878-4fa7-a810-d155ce6b1a8c" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 685.852757] env[62066]: DEBUG oslo_concurrency.lockutils [None req-19be2ac0-adbe-403f-ab5a-6d1895be1574 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.500s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 685.853230] env[62066]: DEBUG nova.compute.manager [None req-19be2ac0-adbe-403f-ab5a-6d1895be1574 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 685.855702] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4d395966-f335-497d-b8a0-23e22d9b79f7 tempest-ServerActionsTestOtherA-1520864800 tempest-ServerActionsTestOtherA-1520864800-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.062s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 686.091629] env[62066]: DEBUG oslo_concurrency.lockutils [None req-df48af02-4c4b-4364-8d38-efc343bf8241 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Acquiring lock "49f46244-34fa-48a1-95a2-8e95850f345d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 686.091629] env[62066]: DEBUG oslo_concurrency.lockutils [None req-df48af02-4c4b-4364-8d38-efc343bf8241 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Lock "49f46244-34fa-48a1-95a2-8e95850f345d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 686.273297] env[62066]: DEBUG nova.network.neutron [req-48b1c930-aa92-480a-962b-956c877bd190 req-9904d57d-2136-414c-b64a-a16c3c5c2eee service nova] [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 686.360134] env[62066]: DEBUG nova.network.neutron [req-48b1c930-aa92-480a-962b-956c877bd190 req-9904d57d-2136-414c-b64a-a16c3c5c2eee service nova] [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 686.362888] env[62066]: DEBUG nova.compute.utils [None req-19be2ac0-adbe-403f-ab5a-6d1895be1574 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 686.367515] env[62066]: DEBUG nova.compute.manager [None req-19be2ac0-adbe-403f-ab5a-6d1895be1574 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 686.367694] env[62066]: DEBUG nova.network.neutron [None req-19be2ac0-adbe-403f-ab5a-6d1895be1574 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 686.415879] env[62066]: DEBUG nova.policy [None req-19be2ac0-adbe-403f-ab5a-6d1895be1574 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fef942e00fa5418ca11215a109e5ca59', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c358d8ed80254a148ce1d26749e64951', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 686.717381] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44db522f-fdb7-4f25-a8fb-a06607814706 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.726737] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0059032-a6ef-439b-8534-b55e6c1562d5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.771085] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12e952ad-f8cf-48bd-9cc1-cff3d07e1d3a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.774613] env[62066]: DEBUG nova.network.neutron [None req-19be2ac0-adbe-403f-ab5a-6d1895be1574 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] Successfully created port: 2ddcf220-1678-4e7d-9792-15f3d090508d {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 686.779462] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6db122c2-ed19-4c4d-9baf-a27c45c69c1c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.792763] env[62066]: DEBUG nova.compute.provider_tree [None req-4d395966-f335-497d-b8a0-23e22d9b79f7 tempest-ServerActionsTestOtherA-1520864800 tempest-ServerActionsTestOtherA-1520864800-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 686.867958] env[62066]: DEBUG oslo_concurrency.lockutils [req-48b1c930-aa92-480a-962b-956c877bd190 req-9904d57d-2136-414c-b64a-a16c3c5c2eee service nova] Releasing lock "refresh_cache-a690ad15-2878-4fa7-a810-d155ce6b1a8c" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 686.868576] env[62066]: DEBUG nova.compute.manager [None req-19be2ac0-adbe-403f-ab5a-6d1895be1574 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 686.871037] env[62066]: DEBUG oslo_concurrency.lockutils [None req-60a20a80-792e-4a03-84da-c17f8b28425d tempest-InstanceActionsNegativeTestJSON-1895685779 tempest-InstanceActionsNegativeTestJSON-1895685779-project-member] Acquired lock "refresh_cache-a690ad15-2878-4fa7-a810-d155ce6b1a8c" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 686.871628] env[62066]: DEBUG nova.network.neutron [None req-60a20a80-792e-4a03-84da-c17f8b28425d tempest-InstanceActionsNegativeTestJSON-1895685779 tempest-InstanceActionsNegativeTestJSON-1895685779-project-member] [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 687.295736] env[62066]: DEBUG nova.scheduler.client.report [None req-4d395966-f335-497d-b8a0-23e22d9b79f7 tempest-ServerActionsTestOtherA-1520864800 tempest-ServerActionsTestOtherA-1520864800-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 687.395578] env[62066]: DEBUG nova.network.neutron [None req-60a20a80-792e-4a03-84da-c17f8b28425d tempest-InstanceActionsNegativeTestJSON-1895685779 tempest-InstanceActionsNegativeTestJSON-1895685779-project-member] [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 687.469119] env[62066]: DEBUG nova.network.neutron [None req-60a20a80-792e-4a03-84da-c17f8b28425d tempest-InstanceActionsNegativeTestJSON-1895685779 tempest-InstanceActionsNegativeTestJSON-1895685779-project-member] [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 687.777852] env[62066]: DEBUG nova.compute.manager [req-3c6c053a-b0f6-4b1d-846e-0c90190169ef req-dfc8ca02-d823-4c41-8766-ae3205ed9f7f service nova] [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] Received event network-vif-deleted-df0ac18e-7f9a-44a1-98ac-693de6d5fa5f {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 687.777852] env[62066]: DEBUG nova.compute.manager [req-3c6c053a-b0f6-4b1d-846e-0c90190169ef req-dfc8ca02-d823-4c41-8766-ae3205ed9f7f service nova] [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] Received event network-changed-2ddcf220-1678-4e7d-9792-15f3d090508d {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 687.777852] env[62066]: DEBUG nova.compute.manager [req-3c6c053a-b0f6-4b1d-846e-0c90190169ef req-dfc8ca02-d823-4c41-8766-ae3205ed9f7f service nova] [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] Refreshing instance network info cache due to event network-changed-2ddcf220-1678-4e7d-9792-15f3d090508d. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 687.777852] env[62066]: DEBUG oslo_concurrency.lockutils [req-3c6c053a-b0f6-4b1d-846e-0c90190169ef req-dfc8ca02-d823-4c41-8766-ae3205ed9f7f service nova] Acquiring lock "refresh_cache-5fad2258-f694-4993-bf64-bfa16abc09ba" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 687.778092] env[62066]: DEBUG oslo_concurrency.lockutils [req-3c6c053a-b0f6-4b1d-846e-0c90190169ef req-dfc8ca02-d823-4c41-8766-ae3205ed9f7f service nova] Acquired lock "refresh_cache-5fad2258-f694-4993-bf64-bfa16abc09ba" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 687.778453] env[62066]: DEBUG nova.network.neutron [req-3c6c053a-b0f6-4b1d-846e-0c90190169ef req-dfc8ca02-d823-4c41-8766-ae3205ed9f7f service nova] [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] Refreshing network info cache for port 2ddcf220-1678-4e7d-9792-15f3d090508d {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 687.805752] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4d395966-f335-497d-b8a0-23e22d9b79f7 tempest-ServerActionsTestOtherA-1520864800 tempest-ServerActionsTestOtherA-1520864800-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.950s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 687.806399] env[62066]: ERROR nova.compute.manager [None req-4d395966-f335-497d-b8a0-23e22d9b79f7 tempest-ServerActionsTestOtherA-1520864800 tempest-ServerActionsTestOtherA-1520864800-project-member] [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 22ad3b3e-efb9-4f6f-a5a5-e9bcbbbf5ec3, please check neutron logs for more information. [ 687.806399] env[62066]: ERROR nova.compute.manager [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] Traceback (most recent call last): [ 687.806399] env[62066]: ERROR nova.compute.manager [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 687.806399] env[62066]: ERROR nova.compute.manager [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] self.driver.spawn(context, instance, image_meta, [ 687.806399] env[62066]: ERROR nova.compute.manager [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 687.806399] env[62066]: ERROR nova.compute.manager [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] self._vmops.spawn(context, instance, image_meta, injected_files, [ 687.806399] env[62066]: ERROR nova.compute.manager [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 687.806399] env[62066]: ERROR nova.compute.manager [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] vm_ref = self.build_virtual_machine(instance, [ 687.806399] env[62066]: ERROR nova.compute.manager [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 687.806399] env[62066]: ERROR nova.compute.manager [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] vif_infos = vmwarevif.get_vif_info(self._session, [ 687.806399] env[62066]: ERROR nova.compute.manager [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 687.806791] env[62066]: ERROR nova.compute.manager [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] for vif in network_info: [ 687.806791] env[62066]: ERROR nova.compute.manager [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 687.806791] env[62066]: ERROR nova.compute.manager [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] return self._sync_wrapper(fn, *args, **kwargs) [ 687.806791] env[62066]: ERROR nova.compute.manager [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 687.806791] env[62066]: ERROR nova.compute.manager [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] self.wait() [ 687.806791] env[62066]: ERROR nova.compute.manager [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 687.806791] env[62066]: ERROR nova.compute.manager [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] self[:] = self._gt.wait() [ 687.806791] env[62066]: ERROR nova.compute.manager [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 687.806791] env[62066]: ERROR nova.compute.manager [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] return self._exit_event.wait() [ 687.806791] env[62066]: ERROR nova.compute.manager [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 687.806791] env[62066]: ERROR nova.compute.manager [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] result = hub.switch() [ 687.806791] env[62066]: ERROR nova.compute.manager [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 687.806791] env[62066]: ERROR nova.compute.manager [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] return self.greenlet.switch() [ 687.807214] env[62066]: ERROR nova.compute.manager [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 687.807214] env[62066]: ERROR nova.compute.manager [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] result = function(*args, **kwargs) [ 687.807214] env[62066]: ERROR nova.compute.manager [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 687.807214] env[62066]: ERROR nova.compute.manager [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] return func(*args, **kwargs) [ 687.807214] env[62066]: ERROR nova.compute.manager [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 687.807214] env[62066]: ERROR nova.compute.manager [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] raise e [ 687.807214] env[62066]: ERROR nova.compute.manager [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 687.807214] env[62066]: ERROR nova.compute.manager [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] nwinfo = self.network_api.allocate_for_instance( [ 687.807214] env[62066]: ERROR nova.compute.manager [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 687.807214] env[62066]: ERROR nova.compute.manager [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] created_port_ids = self._update_ports_for_instance( [ 687.807214] env[62066]: ERROR nova.compute.manager [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 687.807214] env[62066]: ERROR nova.compute.manager [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] with excutils.save_and_reraise_exception(): [ 687.807214] env[62066]: ERROR nova.compute.manager [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 687.807602] env[62066]: ERROR nova.compute.manager [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] self.force_reraise() [ 687.807602] env[62066]: ERROR nova.compute.manager [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 687.807602] env[62066]: ERROR nova.compute.manager [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] raise self.value [ 687.807602] env[62066]: ERROR nova.compute.manager [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 687.807602] env[62066]: ERROR nova.compute.manager [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] updated_port = self._update_port( [ 687.807602] env[62066]: ERROR nova.compute.manager [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 687.807602] env[62066]: ERROR nova.compute.manager [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] _ensure_no_port_binding_failure(port) [ 687.807602] env[62066]: ERROR nova.compute.manager [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 687.807602] env[62066]: ERROR nova.compute.manager [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] raise exception.PortBindingFailed(port_id=port['id']) [ 687.807602] env[62066]: ERROR nova.compute.manager [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] nova.exception.PortBindingFailed: Binding failed for port 22ad3b3e-efb9-4f6f-a5a5-e9bcbbbf5ec3, please check neutron logs for more information. [ 687.807602] env[62066]: ERROR nova.compute.manager [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] [ 687.807942] env[62066]: DEBUG nova.compute.utils [None req-4d395966-f335-497d-b8a0-23e22d9b79f7 tempest-ServerActionsTestOtherA-1520864800 tempest-ServerActionsTestOtherA-1520864800-project-member] [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] Binding failed for port 22ad3b3e-efb9-4f6f-a5a5-e9bcbbbf5ec3, please check neutron logs for more information. {{(pid=62066) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 687.808245] env[62066]: DEBUG oslo_concurrency.lockutils [None req-96f72479-6fed-4ff5-9d17-40fa20c1f416 tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.971s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 687.810098] env[62066]: INFO nova.compute.claims [None req-96f72479-6fed-4ff5-9d17-40fa20c1f416 tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 687.813084] env[62066]: DEBUG nova.compute.manager [None req-4d395966-f335-497d-b8a0-23e22d9b79f7 tempest-ServerActionsTestOtherA-1520864800 tempest-ServerActionsTestOtherA-1520864800-project-member] [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] Build of instance 56984993-a8b1-464e-b20c-79ffe67d2491 was re-scheduled: Binding failed for port 22ad3b3e-efb9-4f6f-a5a5-e9bcbbbf5ec3, please check neutron logs for more information. {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 687.813452] env[62066]: DEBUG nova.compute.manager [None req-4d395966-f335-497d-b8a0-23e22d9b79f7 tempest-ServerActionsTestOtherA-1520864800 tempest-ServerActionsTestOtherA-1520864800-project-member] [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] Unplugging VIFs for instance {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 687.813688] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4d395966-f335-497d-b8a0-23e22d9b79f7 tempest-ServerActionsTestOtherA-1520864800 tempest-ServerActionsTestOtherA-1520864800-project-member] Acquiring lock "refresh_cache-56984993-a8b1-464e-b20c-79ffe67d2491" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 687.814429] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4d395966-f335-497d-b8a0-23e22d9b79f7 tempest-ServerActionsTestOtherA-1520864800 tempest-ServerActionsTestOtherA-1520864800-project-member] Acquired lock "refresh_cache-56984993-a8b1-464e-b20c-79ffe67d2491" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 687.814429] env[62066]: DEBUG nova.network.neutron [None req-4d395966-f335-497d-b8a0-23e22d9b79f7 tempest-ServerActionsTestOtherA-1520864800 tempest-ServerActionsTestOtherA-1520864800-project-member] [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 687.880770] env[62066]: DEBUG nova.compute.manager [None req-19be2ac0-adbe-403f-ab5a-6d1895be1574 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 687.905851] env[62066]: ERROR nova.compute.manager [None req-19be2ac0-adbe-403f-ab5a-6d1895be1574 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 2ddcf220-1678-4e7d-9792-15f3d090508d, please check neutron logs for more information. [ 687.905851] env[62066]: ERROR nova.compute.manager Traceback (most recent call last): [ 687.905851] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 687.905851] env[62066]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 687.905851] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 687.905851] env[62066]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 687.905851] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 687.905851] env[62066]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 687.905851] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 687.905851] env[62066]: ERROR nova.compute.manager self.force_reraise() [ 687.905851] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 687.905851] env[62066]: ERROR nova.compute.manager raise self.value [ 687.905851] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 687.905851] env[62066]: ERROR nova.compute.manager updated_port = self._update_port( [ 687.905851] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 687.905851] env[62066]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 687.906389] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 687.906389] env[62066]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 687.906389] env[62066]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 2ddcf220-1678-4e7d-9792-15f3d090508d, please check neutron logs for more information. [ 687.906389] env[62066]: ERROR nova.compute.manager [ 687.906389] env[62066]: Traceback (most recent call last): [ 687.906389] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 687.906389] env[62066]: listener.cb(fileno) [ 687.906389] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 687.906389] env[62066]: result = function(*args, **kwargs) [ 687.906389] env[62066]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 687.906389] env[62066]: return func(*args, **kwargs) [ 687.906389] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 687.906389] env[62066]: raise e [ 687.906389] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 687.906389] env[62066]: nwinfo = self.network_api.allocate_for_instance( [ 687.906389] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 687.906389] env[62066]: created_port_ids = self._update_ports_for_instance( [ 687.906389] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 687.906389] env[62066]: with excutils.save_and_reraise_exception(): [ 687.906389] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 687.906389] env[62066]: self.force_reraise() [ 687.906389] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 687.906389] env[62066]: raise self.value [ 687.906389] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 687.906389] env[62066]: updated_port = self._update_port( [ 687.906389] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 687.906389] env[62066]: _ensure_no_port_binding_failure(port) [ 687.906389] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 687.906389] env[62066]: raise exception.PortBindingFailed(port_id=port['id']) [ 687.907148] env[62066]: nova.exception.PortBindingFailed: Binding failed for port 2ddcf220-1678-4e7d-9792-15f3d090508d, please check neutron logs for more information. [ 687.907148] env[62066]: Removing descriptor: 20 [ 687.912362] env[62066]: DEBUG nova.virt.hardware [None req-19be2ac0-adbe-403f-ab5a-6d1895be1574 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-23T13:43:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='3a4416c2-2af7-478e-b12e-2abbaac00281',id=33,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-499788702',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-23T13:40:23Z,direct_url=,disk_format='vmdk',id=50ff584c-3b50-4395-af07-3e66769bc9f7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='eb52f7069a374c61ae946f052007c6d9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-23T13:40:24Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 687.912589] env[62066]: DEBUG nova.virt.hardware [None req-19be2ac0-adbe-403f-ab5a-6d1895be1574 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 687.912746] env[62066]: DEBUG nova.virt.hardware [None req-19be2ac0-adbe-403f-ab5a-6d1895be1574 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 687.912922] env[62066]: DEBUG nova.virt.hardware [None req-19be2ac0-adbe-403f-ab5a-6d1895be1574 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 687.913079] env[62066]: DEBUG nova.virt.hardware [None req-19be2ac0-adbe-403f-ab5a-6d1895be1574 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 687.913230] env[62066]: DEBUG nova.virt.hardware [None req-19be2ac0-adbe-403f-ab5a-6d1895be1574 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 687.913441] env[62066]: DEBUG nova.virt.hardware [None req-19be2ac0-adbe-403f-ab5a-6d1895be1574 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 687.913598] env[62066]: DEBUG nova.virt.hardware [None req-19be2ac0-adbe-403f-ab5a-6d1895be1574 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 687.913762] env[62066]: DEBUG nova.virt.hardware [None req-19be2ac0-adbe-403f-ab5a-6d1895be1574 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 687.913921] env[62066]: DEBUG nova.virt.hardware [None req-19be2ac0-adbe-403f-ab5a-6d1895be1574 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 687.914105] env[62066]: DEBUG nova.virt.hardware [None req-19be2ac0-adbe-403f-ab5a-6d1895be1574 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 687.915233] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ebf26df-fd6d-4ab3-ab3b-c5bf31ec0ff9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.924731] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d429b6a-85ae-4afd-b513-b7ef7cfa150e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.938582] env[62066]: ERROR nova.compute.manager [None req-19be2ac0-adbe-403f-ab5a-6d1895be1574 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 2ddcf220-1678-4e7d-9792-15f3d090508d, please check neutron logs for more information. [ 687.938582] env[62066]: ERROR nova.compute.manager [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] Traceback (most recent call last): [ 687.938582] env[62066]: ERROR nova.compute.manager [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 687.938582] env[62066]: ERROR nova.compute.manager [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] yield resources [ 687.938582] env[62066]: ERROR nova.compute.manager [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 687.938582] env[62066]: ERROR nova.compute.manager [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] self.driver.spawn(context, instance, image_meta, [ 687.938582] env[62066]: ERROR nova.compute.manager [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 687.938582] env[62066]: ERROR nova.compute.manager [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] self._vmops.spawn(context, instance, image_meta, injected_files, [ 687.938582] env[62066]: ERROR nova.compute.manager [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 687.938582] env[62066]: ERROR nova.compute.manager [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] vm_ref = self.build_virtual_machine(instance, [ 687.938582] env[62066]: ERROR nova.compute.manager [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 687.939049] env[62066]: ERROR nova.compute.manager [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] vif_infos = vmwarevif.get_vif_info(self._session, [ 687.939049] env[62066]: ERROR nova.compute.manager [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 687.939049] env[62066]: ERROR nova.compute.manager [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] for vif in network_info: [ 687.939049] env[62066]: ERROR nova.compute.manager [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 687.939049] env[62066]: ERROR nova.compute.manager [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] return self._sync_wrapper(fn, *args, **kwargs) [ 687.939049] env[62066]: ERROR nova.compute.manager [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 687.939049] env[62066]: ERROR nova.compute.manager [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] self.wait() [ 687.939049] env[62066]: ERROR nova.compute.manager [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 687.939049] env[62066]: ERROR nova.compute.manager [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] self[:] = self._gt.wait() [ 687.939049] env[62066]: ERROR nova.compute.manager [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 687.939049] env[62066]: ERROR nova.compute.manager [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] return self._exit_event.wait() [ 687.939049] env[62066]: ERROR nova.compute.manager [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 687.939049] env[62066]: ERROR nova.compute.manager [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] current.throw(*self._exc) [ 687.939466] env[62066]: ERROR nova.compute.manager [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 687.939466] env[62066]: ERROR nova.compute.manager [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] result = function(*args, **kwargs) [ 687.939466] env[62066]: ERROR nova.compute.manager [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 687.939466] env[62066]: ERROR nova.compute.manager [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] return func(*args, **kwargs) [ 687.939466] env[62066]: ERROR nova.compute.manager [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 687.939466] env[62066]: ERROR nova.compute.manager [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] raise e [ 687.939466] env[62066]: ERROR nova.compute.manager [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 687.939466] env[62066]: ERROR nova.compute.manager [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] nwinfo = self.network_api.allocate_for_instance( [ 687.939466] env[62066]: ERROR nova.compute.manager [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 687.939466] env[62066]: ERROR nova.compute.manager [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] created_port_ids = self._update_ports_for_instance( [ 687.939466] env[62066]: ERROR nova.compute.manager [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 687.939466] env[62066]: ERROR nova.compute.manager [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] with excutils.save_and_reraise_exception(): [ 687.939466] env[62066]: ERROR nova.compute.manager [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 687.939877] env[62066]: ERROR nova.compute.manager [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] self.force_reraise() [ 687.939877] env[62066]: ERROR nova.compute.manager [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 687.939877] env[62066]: ERROR nova.compute.manager [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] raise self.value [ 687.939877] env[62066]: ERROR nova.compute.manager [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 687.939877] env[62066]: ERROR nova.compute.manager [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] updated_port = self._update_port( [ 687.939877] env[62066]: ERROR nova.compute.manager [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 687.939877] env[62066]: ERROR nova.compute.manager [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] _ensure_no_port_binding_failure(port) [ 687.939877] env[62066]: ERROR nova.compute.manager [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 687.939877] env[62066]: ERROR nova.compute.manager [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] raise exception.PortBindingFailed(port_id=port['id']) [ 687.939877] env[62066]: ERROR nova.compute.manager [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] nova.exception.PortBindingFailed: Binding failed for port 2ddcf220-1678-4e7d-9792-15f3d090508d, please check neutron logs for more information. [ 687.939877] env[62066]: ERROR nova.compute.manager [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] [ 687.939877] env[62066]: INFO nova.compute.manager [None req-19be2ac0-adbe-403f-ab5a-6d1895be1574 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] Terminating instance [ 687.940451] env[62066]: DEBUG oslo_concurrency.lockutils [None req-19be2ac0-adbe-403f-ab5a-6d1895be1574 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] Acquiring lock "refresh_cache-5fad2258-f694-4993-bf64-bfa16abc09ba" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 687.972116] env[62066]: DEBUG oslo_concurrency.lockutils [None req-60a20a80-792e-4a03-84da-c17f8b28425d tempest-InstanceActionsNegativeTestJSON-1895685779 tempest-InstanceActionsNegativeTestJSON-1895685779-project-member] Releasing lock "refresh_cache-a690ad15-2878-4fa7-a810-d155ce6b1a8c" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 687.972531] env[62066]: DEBUG nova.compute.manager [None req-60a20a80-792e-4a03-84da-c17f8b28425d tempest-InstanceActionsNegativeTestJSON-1895685779 tempest-InstanceActionsNegativeTestJSON-1895685779-project-member] [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 687.972716] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-60a20a80-792e-4a03-84da-c17f8b28425d tempest-InstanceActionsNegativeTestJSON-1895685779 tempest-InstanceActionsNegativeTestJSON-1895685779-project-member] [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 687.972991] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d1556de8-4ecb-4ba9-9fc2-663c73e4d1fd {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.981343] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70c129cb-c912-42b5-8886-ef7d9551899f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.002353] env[62066]: WARNING nova.virt.vmwareapi.vmops [None req-60a20a80-792e-4a03-84da-c17f8b28425d tempest-InstanceActionsNegativeTestJSON-1895685779 tempest-InstanceActionsNegativeTestJSON-1895685779-project-member] [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance a690ad15-2878-4fa7-a810-d155ce6b1a8c could not be found. [ 688.003078] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-60a20a80-792e-4a03-84da-c17f8b28425d tempest-InstanceActionsNegativeTestJSON-1895685779 tempest-InstanceActionsNegativeTestJSON-1895685779-project-member] [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 688.003188] env[62066]: INFO nova.compute.manager [None req-60a20a80-792e-4a03-84da-c17f8b28425d tempest-InstanceActionsNegativeTestJSON-1895685779 tempest-InstanceActionsNegativeTestJSON-1895685779-project-member] [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] Took 0.03 seconds to destroy the instance on the hypervisor. [ 688.003436] env[62066]: DEBUG oslo.service.loopingcall [None req-60a20a80-792e-4a03-84da-c17f8b28425d tempest-InstanceActionsNegativeTestJSON-1895685779 tempest-InstanceActionsNegativeTestJSON-1895685779-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 688.003650] env[62066]: DEBUG nova.compute.manager [-] [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 688.003745] env[62066]: DEBUG nova.network.neutron [-] [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 688.017376] env[62066]: DEBUG nova.network.neutron [-] [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 688.296211] env[62066]: DEBUG nova.network.neutron [req-3c6c053a-b0f6-4b1d-846e-0c90190169ef req-dfc8ca02-d823-4c41-8766-ae3205ed9f7f service nova] [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 688.333485] env[62066]: DEBUG nova.network.neutron [None req-4d395966-f335-497d-b8a0-23e22d9b79f7 tempest-ServerActionsTestOtherA-1520864800 tempest-ServerActionsTestOtherA-1520864800-project-member] [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 688.361485] env[62066]: DEBUG nova.network.neutron [req-3c6c053a-b0f6-4b1d-846e-0c90190169ef req-dfc8ca02-d823-4c41-8766-ae3205ed9f7f service nova] [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 688.407539] env[62066]: DEBUG nova.network.neutron [None req-4d395966-f335-497d-b8a0-23e22d9b79f7 tempest-ServerActionsTestOtherA-1520864800 tempest-ServerActionsTestOtherA-1520864800-project-member] [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 688.519599] env[62066]: DEBUG nova.network.neutron [-] [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 688.864398] env[62066]: DEBUG oslo_concurrency.lockutils [req-3c6c053a-b0f6-4b1d-846e-0c90190169ef req-dfc8ca02-d823-4c41-8766-ae3205ed9f7f service nova] Releasing lock "refresh_cache-5fad2258-f694-4993-bf64-bfa16abc09ba" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 688.864813] env[62066]: DEBUG oslo_concurrency.lockutils [None req-19be2ac0-adbe-403f-ab5a-6d1895be1574 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] Acquired lock "refresh_cache-5fad2258-f694-4993-bf64-bfa16abc09ba" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 688.864998] env[62066]: DEBUG nova.network.neutron [None req-19be2ac0-adbe-403f-ab5a-6d1895be1574 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 688.910320] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4d395966-f335-497d-b8a0-23e22d9b79f7 tempest-ServerActionsTestOtherA-1520864800 tempest-ServerActionsTestOtherA-1520864800-project-member] Releasing lock "refresh_cache-56984993-a8b1-464e-b20c-79ffe67d2491" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 688.910569] env[62066]: DEBUG nova.compute.manager [None req-4d395966-f335-497d-b8a0-23e22d9b79f7 tempest-ServerActionsTestOtherA-1520864800 tempest-ServerActionsTestOtherA-1520864800-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 688.910759] env[62066]: DEBUG nova.compute.manager [None req-4d395966-f335-497d-b8a0-23e22d9b79f7 tempest-ServerActionsTestOtherA-1520864800 tempest-ServerActionsTestOtherA-1520864800-project-member] [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 688.910922] env[62066]: DEBUG nova.network.neutron [None req-4d395966-f335-497d-b8a0-23e22d9b79f7 tempest-ServerActionsTestOtherA-1520864800 tempest-ServerActionsTestOtherA-1520864800-project-member] [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 688.925947] env[62066]: DEBUG nova.network.neutron [None req-4d395966-f335-497d-b8a0-23e22d9b79f7 tempest-ServerActionsTestOtherA-1520864800 tempest-ServerActionsTestOtherA-1520864800-project-member] [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 689.023030] env[62066]: INFO nova.compute.manager [-] [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] Took 1.02 seconds to deallocate network for instance. [ 689.025551] env[62066]: DEBUG nova.compute.claims [None req-60a20a80-792e-4a03-84da-c17f8b28425d tempest-InstanceActionsNegativeTestJSON-1895685779 tempest-InstanceActionsNegativeTestJSON-1895685779-project-member] [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] Aborting claim: {{(pid=62066) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 689.025551] env[62066]: DEBUG oslo_concurrency.lockutils [None req-60a20a80-792e-4a03-84da-c17f8b28425d tempest-InstanceActionsNegativeTestJSON-1895685779 tempest-InstanceActionsNegativeTestJSON-1895685779-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 689.176690] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25113973-9744-4e89-beb0-af273a827492 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.184418] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-954f660b-38e1-4624-a6c3-38e3fceedca6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.215526] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78e1da1c-eccb-4464-aaa7-c057dca45b3f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.223174] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b802795b-2dde-44c7-8ece-f442149febd6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.236877] env[62066]: DEBUG nova.compute.provider_tree [None req-96f72479-6fed-4ff5-9d17-40fa20c1f416 tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 689.385923] env[62066]: DEBUG nova.network.neutron [None req-19be2ac0-adbe-403f-ab5a-6d1895be1574 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 689.428594] env[62066]: DEBUG nova.network.neutron [None req-4d395966-f335-497d-b8a0-23e22d9b79f7 tempest-ServerActionsTestOtherA-1520864800 tempest-ServerActionsTestOtherA-1520864800-project-member] [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 689.447397] env[62066]: DEBUG nova.network.neutron [None req-19be2ac0-adbe-403f-ab5a-6d1895be1574 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 689.743046] env[62066]: DEBUG nova.scheduler.client.report [None req-96f72479-6fed-4ff5-9d17-40fa20c1f416 tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 689.840324] env[62066]: DEBUG nova.compute.manager [req-f4cdbac6-557a-4a27-961b-26593f94b7fa req-b3bcfb08-5d94-4cd4-a043-a0d8fe301824 service nova] [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] Received event network-vif-deleted-2ddcf220-1678-4e7d-9792-15f3d090508d {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 689.934069] env[62066]: INFO nova.compute.manager [None req-4d395966-f335-497d-b8a0-23e22d9b79f7 tempest-ServerActionsTestOtherA-1520864800 tempest-ServerActionsTestOtherA-1520864800-project-member] [instance: 56984993-a8b1-464e-b20c-79ffe67d2491] Took 1.02 seconds to deallocate network for instance. [ 689.950534] env[62066]: DEBUG oslo_concurrency.lockutils [None req-19be2ac0-adbe-403f-ab5a-6d1895be1574 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] Releasing lock "refresh_cache-5fad2258-f694-4993-bf64-bfa16abc09ba" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 689.951252] env[62066]: DEBUG nova.compute.manager [None req-19be2ac0-adbe-403f-ab5a-6d1895be1574 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 689.951786] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-19be2ac0-adbe-403f-ab5a-6d1895be1574 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 689.952228] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6e270f0e-f1a7-4a2b-b8d9-93ab8260ba0d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.963678] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2eea46a6-a032-46f3-9d69-0e7d6a353e66 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.989154] env[62066]: WARNING nova.virt.vmwareapi.vmops [None req-19be2ac0-adbe-403f-ab5a-6d1895be1574 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 5fad2258-f694-4993-bf64-bfa16abc09ba could not be found. [ 689.989154] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-19be2ac0-adbe-403f-ab5a-6d1895be1574 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 689.989154] env[62066]: INFO nova.compute.manager [None req-19be2ac0-adbe-403f-ab5a-6d1895be1574 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] Took 0.04 seconds to destroy the instance on the hypervisor. [ 689.989154] env[62066]: DEBUG oslo.service.loopingcall [None req-19be2ac0-adbe-403f-ab5a-6d1895be1574 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 689.989154] env[62066]: DEBUG nova.compute.manager [-] [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 689.989154] env[62066]: DEBUG nova.network.neutron [-] [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 690.005569] env[62066]: DEBUG nova.network.neutron [-] [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 690.248143] env[62066]: DEBUG oslo_concurrency.lockutils [None req-96f72479-6fed-4ff5-9d17-40fa20c1f416 tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.440s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 690.248838] env[62066]: DEBUG nova.compute.manager [None req-96f72479-6fed-4ff5-9d17-40fa20c1f416 tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 690.251533] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cf333fac-d104-4122-9317-eab87c4499f0 tempest-ServerDiagnosticsTest-1737887255 tempest-ServerDiagnosticsTest-1737887255-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.266s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 690.507446] env[62066]: DEBUG nova.network.neutron [-] [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 690.756094] env[62066]: DEBUG nova.compute.utils [None req-96f72479-6fed-4ff5-9d17-40fa20c1f416 tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 690.760556] env[62066]: DEBUG nova.compute.manager [None req-96f72479-6fed-4ff5-9d17-40fa20c1f416 tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 690.760789] env[62066]: DEBUG nova.network.neutron [None req-96f72479-6fed-4ff5-9d17-40fa20c1f416 tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 690.806594] env[62066]: DEBUG nova.policy [None req-96f72479-6fed-4ff5-9d17-40fa20c1f416 tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3fe859c5caeb4338a2cd8c6f8cd0653e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b69df8557ba0480893da58e1e8bf591d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 690.960287] env[62066]: INFO nova.scheduler.client.report [None req-4d395966-f335-497d-b8a0-23e22d9b79f7 tempest-ServerActionsTestOtherA-1520864800 tempest-ServerActionsTestOtherA-1520864800-project-member] Deleted allocations for instance 56984993-a8b1-464e-b20c-79ffe67d2491 [ 691.010756] env[62066]: INFO nova.compute.manager [-] [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] Took 1.02 seconds to deallocate network for instance. [ 691.015176] env[62066]: DEBUG nova.compute.claims [None req-19be2ac0-adbe-403f-ab5a-6d1895be1574 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] Aborting claim: {{(pid=62066) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 691.015420] env[62066]: DEBUG oslo_concurrency.lockutils [None req-19be2ac0-adbe-403f-ab5a-6d1895be1574 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 691.085992] env[62066]: DEBUG nova.network.neutron [None req-96f72479-6fed-4ff5-9d17-40fa20c1f416 tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] Successfully created port: 37035015-85cf-4d55-9ba0-068bae5b3dab {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 691.134165] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47784a4a-5ea3-40c2-8012-55447767459a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.142103] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b576ce9-2b17-4486-a4c4-b85865befad1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.170316] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7093240-9de8-43ab-a653-bcda861a6ff2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.177424] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ea56d87-5f01-4ae7-90ad-c7b497a028eb {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.190412] env[62066]: DEBUG nova.compute.provider_tree [None req-cf333fac-d104-4122-9317-eab87c4499f0 tempest-ServerDiagnosticsTest-1737887255 tempest-ServerDiagnosticsTest-1737887255-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 691.263850] env[62066]: DEBUG nova.compute.manager [None req-96f72479-6fed-4ff5-9d17-40fa20c1f416 tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 691.329906] env[62066]: DEBUG nova.network.neutron [None req-96f72479-6fed-4ff5-9d17-40fa20c1f416 tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] Successfully created port: c7d4a22a-bf6f-439d-91e1-98c478e261f0 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 691.475293] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4d395966-f335-497d-b8a0-23e22d9b79f7 tempest-ServerActionsTestOtherA-1520864800 tempest-ServerActionsTestOtherA-1520864800-project-member] Lock "56984993-a8b1-464e-b20c-79ffe67d2491" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 147.418s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 691.587942] env[62066]: DEBUG nova.network.neutron [None req-96f72479-6fed-4ff5-9d17-40fa20c1f416 tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] Successfully created port: e4dcdcfe-84b4-4c93-a17a-5b9230b2383a {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 691.693882] env[62066]: DEBUG nova.scheduler.client.report [None req-cf333fac-d104-4122-9317-eab87c4499f0 tempest-ServerDiagnosticsTest-1737887255 tempest-ServerDiagnosticsTest-1737887255-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 691.978567] env[62066]: DEBUG nova.compute.manager [None req-8d368cc7-6ac0-44cd-b71e-9136d8337cb4 tempest-AttachInterfacesUnderV243Test-1832696447 tempest-AttachInterfacesUnderV243Test-1832696447-project-member] [instance: f91a90f7-be73-424c-966c-1be6f37a0864] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 692.200305] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cf333fac-d104-4122-9317-eab87c4499f0 tempest-ServerDiagnosticsTest-1737887255 tempest-ServerDiagnosticsTest-1737887255-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.949s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 692.200960] env[62066]: ERROR nova.compute.manager [None req-cf333fac-d104-4122-9317-eab87c4499f0 tempest-ServerDiagnosticsTest-1737887255 tempest-ServerDiagnosticsTest-1737887255-project-member] [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 70d1fb9d-0a6e-46b4-a478-297d09fbe121, please check neutron logs for more information. [ 692.200960] env[62066]: ERROR nova.compute.manager [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] Traceback (most recent call last): [ 692.200960] env[62066]: ERROR nova.compute.manager [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 692.200960] env[62066]: ERROR nova.compute.manager [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] self.driver.spawn(context, instance, image_meta, [ 692.200960] env[62066]: ERROR nova.compute.manager [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 692.200960] env[62066]: ERROR nova.compute.manager [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 692.200960] env[62066]: ERROR nova.compute.manager [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 692.200960] env[62066]: ERROR nova.compute.manager [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] vm_ref = self.build_virtual_machine(instance, [ 692.200960] env[62066]: ERROR nova.compute.manager [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 692.200960] env[62066]: ERROR nova.compute.manager [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] vif_infos = vmwarevif.get_vif_info(self._session, [ 692.200960] env[62066]: ERROR nova.compute.manager [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 692.201351] env[62066]: ERROR nova.compute.manager [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] for vif in network_info: [ 692.201351] env[62066]: ERROR nova.compute.manager [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 692.201351] env[62066]: ERROR nova.compute.manager [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] return self._sync_wrapper(fn, *args, **kwargs) [ 692.201351] env[62066]: ERROR nova.compute.manager [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 692.201351] env[62066]: ERROR nova.compute.manager [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] self.wait() [ 692.201351] env[62066]: ERROR nova.compute.manager [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 692.201351] env[62066]: ERROR nova.compute.manager [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] self[:] = self._gt.wait() [ 692.201351] env[62066]: ERROR nova.compute.manager [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 692.201351] env[62066]: ERROR nova.compute.manager [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] return self._exit_event.wait() [ 692.201351] env[62066]: ERROR nova.compute.manager [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 692.201351] env[62066]: ERROR nova.compute.manager [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] result = hub.switch() [ 692.201351] env[62066]: ERROR nova.compute.manager [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 692.201351] env[62066]: ERROR nova.compute.manager [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] return self.greenlet.switch() [ 692.201878] env[62066]: ERROR nova.compute.manager [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 692.201878] env[62066]: ERROR nova.compute.manager [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] result = function(*args, **kwargs) [ 692.201878] env[62066]: ERROR nova.compute.manager [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 692.201878] env[62066]: ERROR nova.compute.manager [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] return func(*args, **kwargs) [ 692.201878] env[62066]: ERROR nova.compute.manager [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 692.201878] env[62066]: ERROR nova.compute.manager [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] raise e [ 692.201878] env[62066]: ERROR nova.compute.manager [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 692.201878] env[62066]: ERROR nova.compute.manager [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] nwinfo = self.network_api.allocate_for_instance( [ 692.201878] env[62066]: ERROR nova.compute.manager [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 692.201878] env[62066]: ERROR nova.compute.manager [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] created_port_ids = self._update_ports_for_instance( [ 692.201878] env[62066]: ERROR nova.compute.manager [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 692.201878] env[62066]: ERROR nova.compute.manager [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] with excutils.save_and_reraise_exception(): [ 692.201878] env[62066]: ERROR nova.compute.manager [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 692.202238] env[62066]: ERROR nova.compute.manager [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] self.force_reraise() [ 692.202238] env[62066]: ERROR nova.compute.manager [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 692.202238] env[62066]: ERROR nova.compute.manager [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] raise self.value [ 692.202238] env[62066]: ERROR nova.compute.manager [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 692.202238] env[62066]: ERROR nova.compute.manager [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] updated_port = self._update_port( [ 692.202238] env[62066]: ERROR nova.compute.manager [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 692.202238] env[62066]: ERROR nova.compute.manager [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] _ensure_no_port_binding_failure(port) [ 692.202238] env[62066]: ERROR nova.compute.manager [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 692.202238] env[62066]: ERROR nova.compute.manager [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] raise exception.PortBindingFailed(port_id=port['id']) [ 692.202238] env[62066]: ERROR nova.compute.manager [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] nova.exception.PortBindingFailed: Binding failed for port 70d1fb9d-0a6e-46b4-a478-297d09fbe121, please check neutron logs for more information. [ 692.202238] env[62066]: ERROR nova.compute.manager [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] [ 692.202616] env[62066]: DEBUG nova.compute.utils [None req-cf333fac-d104-4122-9317-eab87c4499f0 tempest-ServerDiagnosticsTest-1737887255 tempest-ServerDiagnosticsTest-1737887255-project-member] [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] Binding failed for port 70d1fb9d-0a6e-46b4-a478-297d09fbe121, please check neutron logs for more information. {{(pid=62066) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 692.202882] env[62066]: DEBUG oslo_concurrency.lockutils [None req-67a0ccea-4cde-4de9-9aa4-1bb67e83772e tempest-ServersNegativeTestJSON-789906455 tempest-ServersNegativeTestJSON-789906455-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 18.276s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 692.208674] env[62066]: DEBUG nova.compute.manager [None req-cf333fac-d104-4122-9317-eab87c4499f0 tempest-ServerDiagnosticsTest-1737887255 tempest-ServerDiagnosticsTest-1737887255-project-member] [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] Build of instance 2cf18a04-6bf9-4f47-a920-2c568207057d was re-scheduled: Binding failed for port 70d1fb9d-0a6e-46b4-a478-297d09fbe121, please check neutron logs for more information. {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 692.213201] env[62066]: DEBUG nova.compute.manager [None req-cf333fac-d104-4122-9317-eab87c4499f0 tempest-ServerDiagnosticsTest-1737887255 tempest-ServerDiagnosticsTest-1737887255-project-member] [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] Unplugging VIFs for instance {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 692.213439] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cf333fac-d104-4122-9317-eab87c4499f0 tempest-ServerDiagnosticsTest-1737887255 tempest-ServerDiagnosticsTest-1737887255-project-member] Acquiring lock "refresh_cache-2cf18a04-6bf9-4f47-a920-2c568207057d" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 692.213616] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cf333fac-d104-4122-9317-eab87c4499f0 tempest-ServerDiagnosticsTest-1737887255 tempest-ServerDiagnosticsTest-1737887255-project-member] Acquired lock "refresh_cache-2cf18a04-6bf9-4f47-a920-2c568207057d" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 692.213733] env[62066]: DEBUG nova.network.neutron [None req-cf333fac-d104-4122-9317-eab87c4499f0 tempest-ServerDiagnosticsTest-1737887255 tempest-ServerDiagnosticsTest-1737887255-project-member] [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 692.273433] env[62066]: DEBUG nova.compute.manager [None req-96f72479-6fed-4ff5-9d17-40fa20c1f416 tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 692.303134] env[62066]: DEBUG nova.virt.hardware [None req-96f72479-6fed-4ff5-9d17-40fa20c1f416 tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-23T13:40:41Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-23T13:40:23Z,direct_url=,disk_format='vmdk',id=50ff584c-3b50-4395-af07-3e66769bc9f7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='eb52f7069a374c61ae946f052007c6d9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-23T13:40:24Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 692.303134] env[62066]: DEBUG nova.virt.hardware [None req-96f72479-6fed-4ff5-9d17-40fa20c1f416 tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 692.303134] env[62066]: DEBUG nova.virt.hardware [None req-96f72479-6fed-4ff5-9d17-40fa20c1f416 tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 692.303134] env[62066]: DEBUG nova.virt.hardware [None req-96f72479-6fed-4ff5-9d17-40fa20c1f416 tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 692.303855] env[62066]: DEBUG nova.virt.hardware [None req-96f72479-6fed-4ff5-9d17-40fa20c1f416 tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 692.304357] env[62066]: DEBUG nova.virt.hardware [None req-96f72479-6fed-4ff5-9d17-40fa20c1f416 tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 692.305081] env[62066]: DEBUG nova.virt.hardware [None req-96f72479-6fed-4ff5-9d17-40fa20c1f416 tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 692.305659] env[62066]: DEBUG nova.virt.hardware [None req-96f72479-6fed-4ff5-9d17-40fa20c1f416 tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 692.305962] env[62066]: DEBUG nova.virt.hardware [None req-96f72479-6fed-4ff5-9d17-40fa20c1f416 tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 692.306285] env[62066]: DEBUG nova.virt.hardware [None req-96f72479-6fed-4ff5-9d17-40fa20c1f416 tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 692.306611] env[62066]: DEBUG nova.virt.hardware [None req-96f72479-6fed-4ff5-9d17-40fa20c1f416 tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 692.307658] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d68d9faa-3972-41fd-a1e2-24c54941c842 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.316793] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ded1fdd8-3172-4d42-b906-7ef0273d3344 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.442606] env[62066]: DEBUG nova.compute.manager [req-4d032fe9-f0d8-4463-a977-da21ac9916ab req-f820ce74-8705-418b-96f0-f0192b4d860c service nova] [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] Received event network-changed-37035015-85cf-4d55-9ba0-068bae5b3dab {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 692.442801] env[62066]: DEBUG nova.compute.manager [req-4d032fe9-f0d8-4463-a977-da21ac9916ab req-f820ce74-8705-418b-96f0-f0192b4d860c service nova] [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] Refreshing instance network info cache due to event network-changed-37035015-85cf-4d55-9ba0-068bae5b3dab. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 692.443021] env[62066]: DEBUG oslo_concurrency.lockutils [req-4d032fe9-f0d8-4463-a977-da21ac9916ab req-f820ce74-8705-418b-96f0-f0192b4d860c service nova] Acquiring lock "refresh_cache-45c32181-01a8-4e1e-9e0e-37035a298d55" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 692.443168] env[62066]: DEBUG oslo_concurrency.lockutils [req-4d032fe9-f0d8-4463-a977-da21ac9916ab req-f820ce74-8705-418b-96f0-f0192b4d860c service nova] Acquired lock "refresh_cache-45c32181-01a8-4e1e-9e0e-37035a298d55" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 692.443328] env[62066]: DEBUG nova.network.neutron [req-4d032fe9-f0d8-4463-a977-da21ac9916ab req-f820ce74-8705-418b-96f0-f0192b4d860c service nova] [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] Refreshing network info cache for port 37035015-85cf-4d55-9ba0-068bae5b3dab {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 692.511435] env[62066]: DEBUG oslo_concurrency.lockutils [None req-8d368cc7-6ac0-44cd-b71e-9136d8337cb4 tempest-AttachInterfacesUnderV243Test-1832696447 tempest-AttachInterfacesUnderV243Test-1832696447-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 692.703754] env[62066]: ERROR nova.compute.manager [None req-96f72479-6fed-4ff5-9d17-40fa20c1f416 tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 37035015-85cf-4d55-9ba0-068bae5b3dab, please check neutron logs for more information. [ 692.703754] env[62066]: ERROR nova.compute.manager Traceback (most recent call last): [ 692.703754] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 692.703754] env[62066]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 692.703754] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 692.703754] env[62066]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 692.703754] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 692.703754] env[62066]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 692.703754] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 692.703754] env[62066]: ERROR nova.compute.manager self.force_reraise() [ 692.703754] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 692.703754] env[62066]: ERROR nova.compute.manager raise self.value [ 692.703754] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 692.703754] env[62066]: ERROR nova.compute.manager updated_port = self._update_port( [ 692.703754] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 692.703754] env[62066]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 692.704518] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 692.704518] env[62066]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 692.704518] env[62066]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 37035015-85cf-4d55-9ba0-068bae5b3dab, please check neutron logs for more information. [ 692.704518] env[62066]: ERROR nova.compute.manager [ 692.704518] env[62066]: Traceback (most recent call last): [ 692.704518] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 692.704518] env[62066]: listener.cb(fileno) [ 692.704518] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 692.704518] env[62066]: result = function(*args, **kwargs) [ 692.704518] env[62066]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 692.704518] env[62066]: return func(*args, **kwargs) [ 692.704518] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 692.704518] env[62066]: raise e [ 692.704518] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 692.704518] env[62066]: nwinfo = self.network_api.allocate_for_instance( [ 692.704518] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 692.704518] env[62066]: created_port_ids = self._update_ports_for_instance( [ 692.704518] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 692.704518] env[62066]: with excutils.save_and_reraise_exception(): [ 692.704518] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 692.704518] env[62066]: self.force_reraise() [ 692.704518] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 692.704518] env[62066]: raise self.value [ 692.704518] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 692.704518] env[62066]: updated_port = self._update_port( [ 692.704518] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 692.704518] env[62066]: _ensure_no_port_binding_failure(port) [ 692.704518] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 692.704518] env[62066]: raise exception.PortBindingFailed(port_id=port['id']) [ 692.705286] env[62066]: nova.exception.PortBindingFailed: Binding failed for port 37035015-85cf-4d55-9ba0-068bae5b3dab, please check neutron logs for more information. [ 692.705286] env[62066]: Removing descriptor: 15 [ 692.705286] env[62066]: ERROR nova.compute.manager [None req-96f72479-6fed-4ff5-9d17-40fa20c1f416 tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 37035015-85cf-4d55-9ba0-068bae5b3dab, please check neutron logs for more information. [ 692.705286] env[62066]: ERROR nova.compute.manager [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] Traceback (most recent call last): [ 692.705286] env[62066]: ERROR nova.compute.manager [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 692.705286] env[62066]: ERROR nova.compute.manager [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] yield resources [ 692.705286] env[62066]: ERROR nova.compute.manager [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 692.705286] env[62066]: ERROR nova.compute.manager [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] self.driver.spawn(context, instance, image_meta, [ 692.705286] env[62066]: ERROR nova.compute.manager [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 692.705286] env[62066]: ERROR nova.compute.manager [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] self._vmops.spawn(context, instance, image_meta, injected_files, [ 692.705286] env[62066]: ERROR nova.compute.manager [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 692.705286] env[62066]: ERROR nova.compute.manager [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] vm_ref = self.build_virtual_machine(instance, [ 692.705645] env[62066]: ERROR nova.compute.manager [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 692.705645] env[62066]: ERROR nova.compute.manager [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] vif_infos = vmwarevif.get_vif_info(self._session, [ 692.705645] env[62066]: ERROR nova.compute.manager [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 692.705645] env[62066]: ERROR nova.compute.manager [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] for vif in network_info: [ 692.705645] env[62066]: ERROR nova.compute.manager [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 692.705645] env[62066]: ERROR nova.compute.manager [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] return self._sync_wrapper(fn, *args, **kwargs) [ 692.705645] env[62066]: ERROR nova.compute.manager [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 692.705645] env[62066]: ERROR nova.compute.manager [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] self.wait() [ 692.705645] env[62066]: ERROR nova.compute.manager [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 692.705645] env[62066]: ERROR nova.compute.manager [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] self[:] = self._gt.wait() [ 692.705645] env[62066]: ERROR nova.compute.manager [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 692.705645] env[62066]: ERROR nova.compute.manager [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] return self._exit_event.wait() [ 692.705645] env[62066]: ERROR nova.compute.manager [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 692.706033] env[62066]: ERROR nova.compute.manager [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] result = hub.switch() [ 692.706033] env[62066]: ERROR nova.compute.manager [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 692.706033] env[62066]: ERROR nova.compute.manager [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] return self.greenlet.switch() [ 692.706033] env[62066]: ERROR nova.compute.manager [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 692.706033] env[62066]: ERROR nova.compute.manager [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] result = function(*args, **kwargs) [ 692.706033] env[62066]: ERROR nova.compute.manager [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 692.706033] env[62066]: ERROR nova.compute.manager [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] return func(*args, **kwargs) [ 692.706033] env[62066]: ERROR nova.compute.manager [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 692.706033] env[62066]: ERROR nova.compute.manager [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] raise e [ 692.706033] env[62066]: ERROR nova.compute.manager [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 692.706033] env[62066]: ERROR nova.compute.manager [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] nwinfo = self.network_api.allocate_for_instance( [ 692.706033] env[62066]: ERROR nova.compute.manager [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 692.706033] env[62066]: ERROR nova.compute.manager [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] created_port_ids = self._update_ports_for_instance( [ 692.706420] env[62066]: ERROR nova.compute.manager [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 692.706420] env[62066]: ERROR nova.compute.manager [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] with excutils.save_and_reraise_exception(): [ 692.706420] env[62066]: ERROR nova.compute.manager [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 692.706420] env[62066]: ERROR nova.compute.manager [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] self.force_reraise() [ 692.706420] env[62066]: ERROR nova.compute.manager [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 692.706420] env[62066]: ERROR nova.compute.manager [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] raise self.value [ 692.706420] env[62066]: ERROR nova.compute.manager [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 692.706420] env[62066]: ERROR nova.compute.manager [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] updated_port = self._update_port( [ 692.706420] env[62066]: ERROR nova.compute.manager [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 692.706420] env[62066]: ERROR nova.compute.manager [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] _ensure_no_port_binding_failure(port) [ 692.706420] env[62066]: ERROR nova.compute.manager [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 692.706420] env[62066]: ERROR nova.compute.manager [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] raise exception.PortBindingFailed(port_id=port['id']) [ 692.706757] env[62066]: ERROR nova.compute.manager [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] nova.exception.PortBindingFailed: Binding failed for port 37035015-85cf-4d55-9ba0-068bae5b3dab, please check neutron logs for more information. [ 692.706757] env[62066]: ERROR nova.compute.manager [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] [ 692.706757] env[62066]: INFO nova.compute.manager [None req-96f72479-6fed-4ff5-9d17-40fa20c1f416 tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] Terminating instance [ 692.707614] env[62066]: DEBUG oslo_concurrency.lockutils [None req-96f72479-6fed-4ff5-9d17-40fa20c1f416 tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] Acquiring lock "refresh_cache-45c32181-01a8-4e1e-9e0e-37035a298d55" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 692.739817] env[62066]: DEBUG nova.network.neutron [None req-cf333fac-d104-4122-9317-eab87c4499f0 tempest-ServerDiagnosticsTest-1737887255 tempest-ServerDiagnosticsTest-1737887255-project-member] [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 692.839220] env[62066]: DEBUG nova.network.neutron [None req-cf333fac-d104-4122-9317-eab87c4499f0 tempest-ServerDiagnosticsTest-1737887255 tempest-ServerDiagnosticsTest-1737887255-project-member] [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 692.989740] env[62066]: DEBUG nova.network.neutron [req-4d032fe9-f0d8-4463-a977-da21ac9916ab req-f820ce74-8705-418b-96f0-f0192b4d860c service nova] [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 693.131419] env[62066]: DEBUG nova.network.neutron [req-4d032fe9-f0d8-4463-a977-da21ac9916ab req-f820ce74-8705-418b-96f0-f0192b4d860c service nova] [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 693.150431] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf961269-7198-40ab-98c3-1eb4d08bcfc1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.159202] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ea182df-cc42-4022-b865-2ee3cf1ed7a2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.190342] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67a3852b-3a93-4dbe-b974-4e3de4c44f41 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.197741] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ba7de32-2980-4185-8334-69d2d709d653 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.210627] env[62066]: DEBUG nova.compute.provider_tree [None req-67a0ccea-4cde-4de9-9aa4-1bb67e83772e tempest-ServersNegativeTestJSON-789906455 tempest-ServersNegativeTestJSON-789906455-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 693.343805] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cf333fac-d104-4122-9317-eab87c4499f0 tempest-ServerDiagnosticsTest-1737887255 tempest-ServerDiagnosticsTest-1737887255-project-member] Releasing lock "refresh_cache-2cf18a04-6bf9-4f47-a920-2c568207057d" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 693.343805] env[62066]: DEBUG nova.compute.manager [None req-cf333fac-d104-4122-9317-eab87c4499f0 tempest-ServerDiagnosticsTest-1737887255 tempest-ServerDiagnosticsTest-1737887255-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 693.343805] env[62066]: DEBUG nova.compute.manager [None req-cf333fac-d104-4122-9317-eab87c4499f0 tempest-ServerDiagnosticsTest-1737887255 tempest-ServerDiagnosticsTest-1737887255-project-member] [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 693.343805] env[62066]: DEBUG nova.network.neutron [None req-cf333fac-d104-4122-9317-eab87c4499f0 tempest-ServerDiagnosticsTest-1737887255 tempest-ServerDiagnosticsTest-1737887255-project-member] [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 693.370017] env[62066]: DEBUG nova.network.neutron [None req-cf333fac-d104-4122-9317-eab87c4499f0 tempest-ServerDiagnosticsTest-1737887255 tempest-ServerDiagnosticsTest-1737887255-project-member] [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 693.634429] env[62066]: DEBUG oslo_concurrency.lockutils [req-4d032fe9-f0d8-4463-a977-da21ac9916ab req-f820ce74-8705-418b-96f0-f0192b4d860c service nova] Releasing lock "refresh_cache-45c32181-01a8-4e1e-9e0e-37035a298d55" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 693.637999] env[62066]: DEBUG oslo_concurrency.lockutils [None req-96f72479-6fed-4ff5-9d17-40fa20c1f416 tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] Acquired lock "refresh_cache-45c32181-01a8-4e1e-9e0e-37035a298d55" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 693.637999] env[62066]: DEBUG nova.network.neutron [None req-96f72479-6fed-4ff5-9d17-40fa20c1f416 tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 693.715416] env[62066]: DEBUG nova.scheduler.client.report [None req-67a0ccea-4cde-4de9-9aa4-1bb67e83772e tempest-ServersNegativeTestJSON-789906455 tempest-ServersNegativeTestJSON-789906455-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 693.871829] env[62066]: DEBUG nova.network.neutron [None req-cf333fac-d104-4122-9317-eab87c4499f0 tempest-ServerDiagnosticsTest-1737887255 tempest-ServerDiagnosticsTest-1737887255-project-member] [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 694.170224] env[62066]: DEBUG nova.network.neutron [None req-96f72479-6fed-4ff5-9d17-40fa20c1f416 tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 694.224783] env[62066]: DEBUG oslo_concurrency.lockutils [None req-67a0ccea-4cde-4de9-9aa4-1bb67e83772e tempest-ServersNegativeTestJSON-789906455 tempest-ServersNegativeTestJSON-789906455-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.022s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 694.225458] env[62066]: ERROR nova.compute.manager [None req-67a0ccea-4cde-4de9-9aa4-1bb67e83772e tempest-ServersNegativeTestJSON-789906455 tempest-ServersNegativeTestJSON-789906455-project-member] [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port ea8fd014-c428-41e1-a6bb-7aba32299177, please check neutron logs for more information. [ 694.225458] env[62066]: ERROR nova.compute.manager [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] Traceback (most recent call last): [ 694.225458] env[62066]: ERROR nova.compute.manager [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 694.225458] env[62066]: ERROR nova.compute.manager [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] self.driver.spawn(context, instance, image_meta, [ 694.225458] env[62066]: ERROR nova.compute.manager [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 694.225458] env[62066]: ERROR nova.compute.manager [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] self._vmops.spawn(context, instance, image_meta, injected_files, [ 694.225458] env[62066]: ERROR nova.compute.manager [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 694.225458] env[62066]: ERROR nova.compute.manager [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] vm_ref = self.build_virtual_machine(instance, [ 694.225458] env[62066]: ERROR nova.compute.manager [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 694.225458] env[62066]: ERROR nova.compute.manager [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] vif_infos = vmwarevif.get_vif_info(self._session, [ 694.225458] env[62066]: ERROR nova.compute.manager [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 694.225817] env[62066]: ERROR nova.compute.manager [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] for vif in network_info: [ 694.225817] env[62066]: ERROR nova.compute.manager [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 694.225817] env[62066]: ERROR nova.compute.manager [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] return self._sync_wrapper(fn, *args, **kwargs) [ 694.225817] env[62066]: ERROR nova.compute.manager [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 694.225817] env[62066]: ERROR nova.compute.manager [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] self.wait() [ 694.225817] env[62066]: ERROR nova.compute.manager [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 694.225817] env[62066]: ERROR nova.compute.manager [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] self[:] = self._gt.wait() [ 694.225817] env[62066]: ERROR nova.compute.manager [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 694.225817] env[62066]: ERROR nova.compute.manager [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] return self._exit_event.wait() [ 694.225817] env[62066]: ERROR nova.compute.manager [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 694.225817] env[62066]: ERROR nova.compute.manager [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] result = hub.switch() [ 694.225817] env[62066]: ERROR nova.compute.manager [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 694.225817] env[62066]: ERROR nova.compute.manager [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] return self.greenlet.switch() [ 694.226344] env[62066]: ERROR nova.compute.manager [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 694.226344] env[62066]: ERROR nova.compute.manager [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] result = function(*args, **kwargs) [ 694.226344] env[62066]: ERROR nova.compute.manager [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 694.226344] env[62066]: ERROR nova.compute.manager [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] return func(*args, **kwargs) [ 694.226344] env[62066]: ERROR nova.compute.manager [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 694.226344] env[62066]: ERROR nova.compute.manager [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] raise e [ 694.226344] env[62066]: ERROR nova.compute.manager [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 694.226344] env[62066]: ERROR nova.compute.manager [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] nwinfo = self.network_api.allocate_for_instance( [ 694.226344] env[62066]: ERROR nova.compute.manager [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 694.226344] env[62066]: ERROR nova.compute.manager [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] created_port_ids = self._update_ports_for_instance( [ 694.226344] env[62066]: ERROR nova.compute.manager [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 694.226344] env[62066]: ERROR nova.compute.manager [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] with excutils.save_and_reraise_exception(): [ 694.226344] env[62066]: ERROR nova.compute.manager [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 694.226730] env[62066]: ERROR nova.compute.manager [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] self.force_reraise() [ 694.226730] env[62066]: ERROR nova.compute.manager [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 694.226730] env[62066]: ERROR nova.compute.manager [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] raise self.value [ 694.226730] env[62066]: ERROR nova.compute.manager [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 694.226730] env[62066]: ERROR nova.compute.manager [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] updated_port = self._update_port( [ 694.226730] env[62066]: ERROR nova.compute.manager [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 694.226730] env[62066]: ERROR nova.compute.manager [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] _ensure_no_port_binding_failure(port) [ 694.226730] env[62066]: ERROR nova.compute.manager [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 694.226730] env[62066]: ERROR nova.compute.manager [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] raise exception.PortBindingFailed(port_id=port['id']) [ 694.226730] env[62066]: ERROR nova.compute.manager [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] nova.exception.PortBindingFailed: Binding failed for port ea8fd014-c428-41e1-a6bb-7aba32299177, please check neutron logs for more information. [ 694.226730] env[62066]: ERROR nova.compute.manager [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] [ 694.227108] env[62066]: DEBUG nova.compute.utils [None req-67a0ccea-4cde-4de9-9aa4-1bb67e83772e tempest-ServersNegativeTestJSON-789906455 tempest-ServersNegativeTestJSON-789906455-project-member] [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] Binding failed for port ea8fd014-c428-41e1-a6bb-7aba32299177, please check neutron logs for more information. {{(pid=62066) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 694.227396] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b2b06eaa-4237-476b-9167-ed8be81ba3b9 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.846s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 694.228975] env[62066]: INFO nova.compute.claims [None req-b2b06eaa-4237-476b-9167-ed8be81ba3b9 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 694.232673] env[62066]: DEBUG nova.compute.manager [None req-67a0ccea-4cde-4de9-9aa4-1bb67e83772e tempest-ServersNegativeTestJSON-789906455 tempest-ServersNegativeTestJSON-789906455-project-member] [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] Build of instance 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393 was re-scheduled: Binding failed for port ea8fd014-c428-41e1-a6bb-7aba32299177, please check neutron logs for more information. {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 694.233464] env[62066]: DEBUG nova.compute.manager [None req-67a0ccea-4cde-4de9-9aa4-1bb67e83772e tempest-ServersNegativeTestJSON-789906455 tempest-ServersNegativeTestJSON-789906455-project-member] [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] Unplugging VIFs for instance {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 694.233464] env[62066]: DEBUG oslo_concurrency.lockutils [None req-67a0ccea-4cde-4de9-9aa4-1bb67e83772e tempest-ServersNegativeTestJSON-789906455 tempest-ServersNegativeTestJSON-789906455-project-member] Acquiring lock "refresh_cache-5bbdf138-0aaf-42d1-b5a1-4f7ebc650393" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 694.233618] env[62066]: DEBUG oslo_concurrency.lockutils [None req-67a0ccea-4cde-4de9-9aa4-1bb67e83772e tempest-ServersNegativeTestJSON-789906455 tempest-ServersNegativeTestJSON-789906455-project-member] Acquired lock "refresh_cache-5bbdf138-0aaf-42d1-b5a1-4f7ebc650393" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 694.233701] env[62066]: DEBUG nova.network.neutron [None req-67a0ccea-4cde-4de9-9aa4-1bb67e83772e tempest-ServersNegativeTestJSON-789906455 tempest-ServersNegativeTestJSON-789906455-project-member] [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 694.258623] env[62066]: DEBUG nova.network.neutron [None req-96f72479-6fed-4ff5-9d17-40fa20c1f416 tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 694.374592] env[62066]: INFO nova.compute.manager [None req-cf333fac-d104-4122-9317-eab87c4499f0 tempest-ServerDiagnosticsTest-1737887255 tempest-ServerDiagnosticsTest-1737887255-project-member] [instance: 2cf18a04-6bf9-4f47-a920-2c568207057d] Took 1.03 seconds to deallocate network for instance. [ 694.504502] env[62066]: DEBUG nova.compute.manager [req-2e5bf02a-4cbe-4015-98ab-e1a1ca7783b3 req-17010fa5-edd3-4df0-b319-15d804c8dc4d service nova] [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] Received event network-vif-deleted-37035015-85cf-4d55-9ba0-068bae5b3dab {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 694.759870] env[62066]: DEBUG nova.network.neutron [None req-67a0ccea-4cde-4de9-9aa4-1bb67e83772e tempest-ServersNegativeTestJSON-789906455 tempest-ServersNegativeTestJSON-789906455-project-member] [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 694.761862] env[62066]: DEBUG oslo_concurrency.lockutils [None req-96f72479-6fed-4ff5-9d17-40fa20c1f416 tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] Releasing lock "refresh_cache-45c32181-01a8-4e1e-9e0e-37035a298d55" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 694.762501] env[62066]: DEBUG nova.compute.manager [None req-96f72479-6fed-4ff5-9d17-40fa20c1f416 tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 694.762501] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-96f72479-6fed-4ff5-9d17-40fa20c1f416 tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 694.762977] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e7381183-4700-4632-a915-eecf8366c26b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.772575] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33de9b0f-feba-4f3f-b803-0edc71790026 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.795149] env[62066]: WARNING nova.virt.vmwareapi.vmops [None req-96f72479-6fed-4ff5-9d17-40fa20c1f416 tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 45c32181-01a8-4e1e-9e0e-37035a298d55 could not be found. [ 694.795362] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-96f72479-6fed-4ff5-9d17-40fa20c1f416 tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 694.795615] env[62066]: INFO nova.compute.manager [None req-96f72479-6fed-4ff5-9d17-40fa20c1f416 tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] Took 0.03 seconds to destroy the instance on the hypervisor. [ 694.795913] env[62066]: DEBUG oslo.service.loopingcall [None req-96f72479-6fed-4ff5-9d17-40fa20c1f416 tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 694.796156] env[62066]: DEBUG nova.compute.manager [-] [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 694.796242] env[62066]: DEBUG nova.network.neutron [-] [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 694.845190] env[62066]: DEBUG nova.network.neutron [-] [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 695.045637] env[62066]: DEBUG nova.network.neutron [None req-67a0ccea-4cde-4de9-9aa4-1bb67e83772e tempest-ServersNegativeTestJSON-789906455 tempest-ServersNegativeTestJSON-789906455-project-member] [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 695.408451] env[62066]: INFO nova.scheduler.client.report [None req-cf333fac-d104-4122-9317-eab87c4499f0 tempest-ServerDiagnosticsTest-1737887255 tempest-ServerDiagnosticsTest-1737887255-project-member] Deleted allocations for instance 2cf18a04-6bf9-4f47-a920-2c568207057d [ 695.548290] env[62066]: DEBUG oslo_concurrency.lockutils [None req-67a0ccea-4cde-4de9-9aa4-1bb67e83772e tempest-ServersNegativeTestJSON-789906455 tempest-ServersNegativeTestJSON-789906455-project-member] Releasing lock "refresh_cache-5bbdf138-0aaf-42d1-b5a1-4f7ebc650393" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 695.548542] env[62066]: DEBUG nova.compute.manager [None req-67a0ccea-4cde-4de9-9aa4-1bb67e83772e tempest-ServersNegativeTestJSON-789906455 tempest-ServersNegativeTestJSON-789906455-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 695.548724] env[62066]: DEBUG nova.compute.manager [None req-67a0ccea-4cde-4de9-9aa4-1bb67e83772e tempest-ServersNegativeTestJSON-789906455 tempest-ServersNegativeTestJSON-789906455-project-member] [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 695.548891] env[62066]: DEBUG nova.network.neutron [None req-67a0ccea-4cde-4de9-9aa4-1bb67e83772e tempest-ServersNegativeTestJSON-789906455 tempest-ServersNegativeTestJSON-789906455-project-member] [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 695.565974] env[62066]: DEBUG nova.network.neutron [None req-67a0ccea-4cde-4de9-9aa4-1bb67e83772e tempest-ServersNegativeTestJSON-789906455 tempest-ServersNegativeTestJSON-789906455-project-member] [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 695.696252] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87d24b76-2dc5-4333-8ab0-1fb63264f2ff {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.703974] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c42fc31d-b81a-4ee3-ac5f-5f316fbbb093 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.738525] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-374fc839-6a7e-4ba3-9764-cee5a80abc9f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.748011] env[62066]: DEBUG nova.network.neutron [-] [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 695.753281] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f8aad4a-2309-49d0-b4a8-8531e80131f5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.776177] env[62066]: DEBUG nova.compute.provider_tree [None req-b2b06eaa-4237-476b-9167-ed8be81ba3b9 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 695.921223] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cf333fac-d104-4122-9317-eab87c4499f0 tempest-ServerDiagnosticsTest-1737887255 tempest-ServerDiagnosticsTest-1737887255-project-member] Lock "2cf18a04-6bf9-4f47-a920-2c568207057d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 150.358s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 696.070345] env[62066]: DEBUG nova.network.neutron [None req-67a0ccea-4cde-4de9-9aa4-1bb67e83772e tempest-ServersNegativeTestJSON-789906455 tempest-ServersNegativeTestJSON-789906455-project-member] [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 696.259843] env[62066]: INFO nova.compute.manager [-] [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] Took 1.46 seconds to deallocate network for instance. [ 696.262300] env[62066]: DEBUG nova.compute.claims [None req-96f72479-6fed-4ff5-9d17-40fa20c1f416 tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] Aborting claim: {{(pid=62066) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 696.262497] env[62066]: DEBUG oslo_concurrency.lockutils [None req-96f72479-6fed-4ff5-9d17-40fa20c1f416 tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 696.279659] env[62066]: DEBUG nova.scheduler.client.report [None req-b2b06eaa-4237-476b-9167-ed8be81ba3b9 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 696.423795] env[62066]: DEBUG nova.compute.manager [None req-3d3acd2e-756a-442f-acf6-9a75ba7d933c tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] [instance: 0ff2f9f4-6d85-4465-8e0a-fb39f2a25c9b] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 696.574166] env[62066]: INFO nova.compute.manager [None req-67a0ccea-4cde-4de9-9aa4-1bb67e83772e tempest-ServersNegativeTestJSON-789906455 tempest-ServersNegativeTestJSON-789906455-project-member] [instance: 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393] Took 1.02 seconds to deallocate network for instance. [ 696.787834] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b2b06eaa-4237-476b-9167-ed8be81ba3b9 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.559s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 696.787834] env[62066]: DEBUG nova.compute.manager [None req-b2b06eaa-4237-476b-9167-ed8be81ba3b9 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 696.791904] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b96d7111-6e60-4fd2-a38a-3863242412fb tempest-ServerActionsTestOtherB-426172427 tempest-ServerActionsTestOtherB-426172427-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.404s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 696.952891] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3d3acd2e-756a-442f-acf6-9a75ba7d933c tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 697.301498] env[62066]: DEBUG nova.compute.utils [None req-b2b06eaa-4237-476b-9167-ed8be81ba3b9 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 697.311029] env[62066]: DEBUG nova.compute.manager [None req-b2b06eaa-4237-476b-9167-ed8be81ba3b9 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 697.311029] env[62066]: DEBUG nova.network.neutron [None req-b2b06eaa-4237-476b-9167-ed8be81ba3b9 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 697.380653] env[62066]: DEBUG nova.policy [None req-b2b06eaa-4237-476b-9167-ed8be81ba3b9 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8a9b164f679f4f4a9fa28f1362ceb50d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1f05d8a66d0c4a479e49a947e4b4bbff', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 697.627238] env[62066]: INFO nova.scheduler.client.report [None req-67a0ccea-4cde-4de9-9aa4-1bb67e83772e tempest-ServersNegativeTestJSON-789906455 tempest-ServersNegativeTestJSON-789906455-project-member] Deleted allocations for instance 5bbdf138-0aaf-42d1-b5a1-4f7ebc650393 [ 697.800341] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be45a218-e730-4c7d-9527-7d99974e988f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.808612] env[62066]: DEBUG nova.compute.manager [None req-b2b06eaa-4237-476b-9167-ed8be81ba3b9 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 697.812271] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0f5975e-fdb0-4e65-a261-bd22997ace4e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.846011] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcb41a23-49e7-4df0-85ff-880d4d889714 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.855782] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a6ed2b3-58ac-4886-8d60-5d5809d7c7fc {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.871961] env[62066]: DEBUG nova.compute.provider_tree [None req-b96d7111-6e60-4fd2-a38a-3863242412fb tempest-ServerActionsTestOtherB-426172427 tempest-ServerActionsTestOtherB-426172427-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 697.945609] env[62066]: DEBUG nova.network.neutron [None req-b2b06eaa-4237-476b-9167-ed8be81ba3b9 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] Successfully created port: eb8ff994-60f5-461b-b984-ff082644aebc {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 698.143034] env[62066]: DEBUG oslo_concurrency.lockutils [None req-67a0ccea-4cde-4de9-9aa4-1bb67e83772e tempest-ServersNegativeTestJSON-789906455 tempest-ServersNegativeTestJSON-789906455-project-member] Lock "5bbdf138-0aaf-42d1-b5a1-4f7ebc650393" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 151.844s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 698.375785] env[62066]: DEBUG nova.scheduler.client.report [None req-b96d7111-6e60-4fd2-a38a-3863242412fb tempest-ServerActionsTestOtherB-426172427 tempest-ServerActionsTestOtherB-426172427-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 698.645348] env[62066]: DEBUG nova.compute.manager [None req-5b5d7867-70eb-472d-8d1a-c2bb035ec6db tempest-InstanceActionsV221TestJSON-1098427857 tempest-InstanceActionsV221TestJSON-1098427857-project-member] [instance: d7de90c0-95e9-4899-92a0-81658831713a] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 698.728623] env[62066]: DEBUG oslo_concurrency.lockutils [None req-8034bd7f-2294-4f4c-b924-f35bec7ebaac tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] Acquiring lock "738ebb2f-7cc6-4d0f-871c-1428c9c21384" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 698.728725] env[62066]: DEBUG oslo_concurrency.lockutils [None req-8034bd7f-2294-4f4c-b924-f35bec7ebaac tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] Lock "738ebb2f-7cc6-4d0f-871c-1428c9c21384" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 698.823044] env[62066]: DEBUG nova.compute.manager [None req-b2b06eaa-4237-476b-9167-ed8be81ba3b9 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 698.856506] env[62066]: DEBUG nova.virt.hardware [None req-b2b06eaa-4237-476b-9167-ed8be81ba3b9 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-23T13:40:41Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-23T13:40:23Z,direct_url=,disk_format='vmdk',id=50ff584c-3b50-4395-af07-3e66769bc9f7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='eb52f7069a374c61ae946f052007c6d9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-23T13:40:24Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 698.856761] env[62066]: DEBUG nova.virt.hardware [None req-b2b06eaa-4237-476b-9167-ed8be81ba3b9 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 698.856918] env[62066]: DEBUG nova.virt.hardware [None req-b2b06eaa-4237-476b-9167-ed8be81ba3b9 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 698.857114] env[62066]: DEBUG nova.virt.hardware [None req-b2b06eaa-4237-476b-9167-ed8be81ba3b9 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 698.857263] env[62066]: DEBUG nova.virt.hardware [None req-b2b06eaa-4237-476b-9167-ed8be81ba3b9 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 698.857411] env[62066]: DEBUG nova.virt.hardware [None req-b2b06eaa-4237-476b-9167-ed8be81ba3b9 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 698.857616] env[62066]: DEBUG nova.virt.hardware [None req-b2b06eaa-4237-476b-9167-ed8be81ba3b9 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 698.857771] env[62066]: DEBUG nova.virt.hardware [None req-b2b06eaa-4237-476b-9167-ed8be81ba3b9 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 698.857938] env[62066]: DEBUG nova.virt.hardware [None req-b2b06eaa-4237-476b-9167-ed8be81ba3b9 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 698.858348] env[62066]: DEBUG nova.virt.hardware [None req-b2b06eaa-4237-476b-9167-ed8be81ba3b9 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 698.858773] env[62066]: DEBUG nova.virt.hardware [None req-b2b06eaa-4237-476b-9167-ed8be81ba3b9 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 698.859649] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f64acd6-23bf-4899-af42-9b87f1bd1f2b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.869582] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc639c53-0e42-4419-ae05-6f28ff51234a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.886690] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b96d7111-6e60-4fd2-a38a-3863242412fb tempest-ServerActionsTestOtherB-426172427 tempest-ServerActionsTestOtherB-426172427-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.094s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 698.887268] env[62066]: ERROR nova.compute.manager [None req-b96d7111-6e60-4fd2-a38a-3863242412fb tempest-ServerActionsTestOtherB-426172427 tempest-ServerActionsTestOtherB-426172427-project-member] [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 9380aec8-fb60-43d1-b0f9-2068fbbc6c28, please check neutron logs for more information. [ 698.887268] env[62066]: ERROR nova.compute.manager [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] Traceback (most recent call last): [ 698.887268] env[62066]: ERROR nova.compute.manager [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 698.887268] env[62066]: ERROR nova.compute.manager [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] self.driver.spawn(context, instance, image_meta, [ 698.887268] env[62066]: ERROR nova.compute.manager [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 698.887268] env[62066]: ERROR nova.compute.manager [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] self._vmops.spawn(context, instance, image_meta, injected_files, [ 698.887268] env[62066]: ERROR nova.compute.manager [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 698.887268] env[62066]: ERROR nova.compute.manager [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] vm_ref = self.build_virtual_machine(instance, [ 698.887268] env[62066]: ERROR nova.compute.manager [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 698.887268] env[62066]: ERROR nova.compute.manager [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] vif_infos = vmwarevif.get_vif_info(self._session, [ 698.887268] env[62066]: ERROR nova.compute.manager [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 698.887639] env[62066]: ERROR nova.compute.manager [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] for vif in network_info: [ 698.887639] env[62066]: ERROR nova.compute.manager [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 698.887639] env[62066]: ERROR nova.compute.manager [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] return self._sync_wrapper(fn, *args, **kwargs) [ 698.887639] env[62066]: ERROR nova.compute.manager [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 698.887639] env[62066]: ERROR nova.compute.manager [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] self.wait() [ 698.887639] env[62066]: ERROR nova.compute.manager [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 698.887639] env[62066]: ERROR nova.compute.manager [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] self[:] = self._gt.wait() [ 698.887639] env[62066]: ERROR nova.compute.manager [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 698.887639] env[62066]: ERROR nova.compute.manager [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] return self._exit_event.wait() [ 698.887639] env[62066]: ERROR nova.compute.manager [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 698.887639] env[62066]: ERROR nova.compute.manager [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] current.throw(*self._exc) [ 698.887639] env[62066]: ERROR nova.compute.manager [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 698.887639] env[62066]: ERROR nova.compute.manager [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] result = function(*args, **kwargs) [ 698.888130] env[62066]: ERROR nova.compute.manager [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 698.888130] env[62066]: ERROR nova.compute.manager [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] return func(*args, **kwargs) [ 698.888130] env[62066]: ERROR nova.compute.manager [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 698.888130] env[62066]: ERROR nova.compute.manager [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] raise e [ 698.888130] env[62066]: ERROR nova.compute.manager [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 698.888130] env[62066]: ERROR nova.compute.manager [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] nwinfo = self.network_api.allocate_for_instance( [ 698.888130] env[62066]: ERROR nova.compute.manager [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 698.888130] env[62066]: ERROR nova.compute.manager [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] created_port_ids = self._update_ports_for_instance( [ 698.888130] env[62066]: ERROR nova.compute.manager [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 698.888130] env[62066]: ERROR nova.compute.manager [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] with excutils.save_and_reraise_exception(): [ 698.888130] env[62066]: ERROR nova.compute.manager [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 698.888130] env[62066]: ERROR nova.compute.manager [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] self.force_reraise() [ 698.888130] env[62066]: ERROR nova.compute.manager [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 698.888531] env[62066]: ERROR nova.compute.manager [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] raise self.value [ 698.888531] env[62066]: ERROR nova.compute.manager [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 698.888531] env[62066]: ERROR nova.compute.manager [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] updated_port = self._update_port( [ 698.888531] env[62066]: ERROR nova.compute.manager [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 698.888531] env[62066]: ERROR nova.compute.manager [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] _ensure_no_port_binding_failure(port) [ 698.888531] env[62066]: ERROR nova.compute.manager [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 698.888531] env[62066]: ERROR nova.compute.manager [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] raise exception.PortBindingFailed(port_id=port['id']) [ 698.888531] env[62066]: ERROR nova.compute.manager [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] nova.exception.PortBindingFailed: Binding failed for port 9380aec8-fb60-43d1-b0f9-2068fbbc6c28, please check neutron logs for more information. [ 698.888531] env[62066]: ERROR nova.compute.manager [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] [ 698.892334] env[62066]: DEBUG nova.compute.utils [None req-b96d7111-6e60-4fd2-a38a-3863242412fb tempest-ServerActionsTestOtherB-426172427 tempest-ServerActionsTestOtherB-426172427-project-member] [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] Binding failed for port 9380aec8-fb60-43d1-b0f9-2068fbbc6c28, please check neutron logs for more information. {{(pid=62066) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 698.893758] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3515d98a-2869-41bb-bf8e-e38d995a4d22 tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.637s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 698.896598] env[62066]: DEBUG nova.compute.manager [None req-b96d7111-6e60-4fd2-a38a-3863242412fb tempest-ServerActionsTestOtherB-426172427 tempest-ServerActionsTestOtherB-426172427-project-member] [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] Build of instance e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8 was re-scheduled: Binding failed for port 9380aec8-fb60-43d1-b0f9-2068fbbc6c28, please check neutron logs for more information. {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 698.897042] env[62066]: DEBUG nova.compute.manager [None req-b96d7111-6e60-4fd2-a38a-3863242412fb tempest-ServerActionsTestOtherB-426172427 tempest-ServerActionsTestOtherB-426172427-project-member] [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] Unplugging VIFs for instance {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 698.897269] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b96d7111-6e60-4fd2-a38a-3863242412fb tempest-ServerActionsTestOtherB-426172427 tempest-ServerActionsTestOtherB-426172427-project-member] Acquiring lock "refresh_cache-e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 698.897415] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b96d7111-6e60-4fd2-a38a-3863242412fb tempest-ServerActionsTestOtherB-426172427 tempest-ServerActionsTestOtherB-426172427-project-member] Acquired lock "refresh_cache-e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 698.897571] env[62066]: DEBUG nova.network.neutron [None req-b96d7111-6e60-4fd2-a38a-3863242412fb tempest-ServerActionsTestOtherB-426172427 tempest-ServerActionsTestOtherB-426172427-project-member] [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 699.169062] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5b5d7867-70eb-472d-8d1a-c2bb035ec6db tempest-InstanceActionsV221TestJSON-1098427857 tempest-InstanceActionsV221TestJSON-1098427857-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 699.423899] env[62066]: DEBUG nova.network.neutron [None req-b96d7111-6e60-4fd2-a38a-3863242412fb tempest-ServerActionsTestOtherB-426172427 tempest-ServerActionsTestOtherB-426172427-project-member] [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 699.503537] env[62066]: DEBUG nova.network.neutron [None req-b96d7111-6e60-4fd2-a38a-3863242412fb tempest-ServerActionsTestOtherB-426172427 tempest-ServerActionsTestOtherB-426172427-project-member] [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 699.580613] env[62066]: DEBUG nova.compute.manager [req-a6db693b-555a-4b5c-a71d-c3c3d26b2bfb req-f760474b-e6de-4dd1-9188-58a92163c5a4 service nova] [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] Received event network-changed-eb8ff994-60f5-461b-b984-ff082644aebc {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 699.580798] env[62066]: DEBUG nova.compute.manager [req-a6db693b-555a-4b5c-a71d-c3c3d26b2bfb req-f760474b-e6de-4dd1-9188-58a92163c5a4 service nova] [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] Refreshing instance network info cache due to event network-changed-eb8ff994-60f5-461b-b984-ff082644aebc. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 699.581020] env[62066]: DEBUG oslo_concurrency.lockutils [req-a6db693b-555a-4b5c-a71d-c3c3d26b2bfb req-f760474b-e6de-4dd1-9188-58a92163c5a4 service nova] Acquiring lock "refresh_cache-f9a8e863-6e7f-4f00-b54a-78802659cd59" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 699.581164] env[62066]: DEBUG oslo_concurrency.lockutils [req-a6db693b-555a-4b5c-a71d-c3c3d26b2bfb req-f760474b-e6de-4dd1-9188-58a92163c5a4 service nova] Acquired lock "refresh_cache-f9a8e863-6e7f-4f00-b54a-78802659cd59" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 699.582380] env[62066]: DEBUG nova.network.neutron [req-a6db693b-555a-4b5c-a71d-c3c3d26b2bfb req-f760474b-e6de-4dd1-9188-58a92163c5a4 service nova] [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] Refreshing network info cache for port eb8ff994-60f5-461b-b984-ff082644aebc {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 699.796735] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af2f80a4-81db-422b-b3a9-86093c7cba86 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.807801] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed9ff660-651c-4251-b579-1162d2dafb2f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.838535] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b18bd598-2bf7-4ed3-90e0-cf97ec69edc0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.850366] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7b475c8-6f19-47f6-a50a-1fcc292d7ee9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.866545] env[62066]: DEBUG nova.compute.provider_tree [None req-3515d98a-2869-41bb-bf8e-e38d995a4d22 tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 699.898783] env[62066]: ERROR nova.compute.manager [None req-b2b06eaa-4237-476b-9167-ed8be81ba3b9 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port eb8ff994-60f5-461b-b984-ff082644aebc, please check neutron logs for more information. [ 699.898783] env[62066]: ERROR nova.compute.manager Traceback (most recent call last): [ 699.898783] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 699.898783] env[62066]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 699.898783] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 699.898783] env[62066]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 699.898783] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 699.898783] env[62066]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 699.898783] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 699.898783] env[62066]: ERROR nova.compute.manager self.force_reraise() [ 699.898783] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 699.898783] env[62066]: ERROR nova.compute.manager raise self.value [ 699.898783] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 699.898783] env[62066]: ERROR nova.compute.manager updated_port = self._update_port( [ 699.898783] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 699.898783] env[62066]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 699.899350] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 699.899350] env[62066]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 699.899350] env[62066]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port eb8ff994-60f5-461b-b984-ff082644aebc, please check neutron logs for more information. [ 699.899350] env[62066]: ERROR nova.compute.manager [ 699.899350] env[62066]: Traceback (most recent call last): [ 699.899350] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 699.899350] env[62066]: listener.cb(fileno) [ 699.899350] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 699.899350] env[62066]: result = function(*args, **kwargs) [ 699.899350] env[62066]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 699.899350] env[62066]: return func(*args, **kwargs) [ 699.899350] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 699.899350] env[62066]: raise e [ 699.899350] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 699.899350] env[62066]: nwinfo = self.network_api.allocate_for_instance( [ 699.899350] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 699.899350] env[62066]: created_port_ids = self._update_ports_for_instance( [ 699.899350] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 699.899350] env[62066]: with excutils.save_and_reraise_exception(): [ 699.899350] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 699.899350] env[62066]: self.force_reraise() [ 699.899350] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 699.899350] env[62066]: raise self.value [ 699.899350] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 699.899350] env[62066]: updated_port = self._update_port( [ 699.899350] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 699.899350] env[62066]: _ensure_no_port_binding_failure(port) [ 699.899350] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 699.899350] env[62066]: raise exception.PortBindingFailed(port_id=port['id']) [ 699.900273] env[62066]: nova.exception.PortBindingFailed: Binding failed for port eb8ff994-60f5-461b-b984-ff082644aebc, please check neutron logs for more information. [ 699.900273] env[62066]: Removing descriptor: 20 [ 699.900273] env[62066]: ERROR nova.compute.manager [None req-b2b06eaa-4237-476b-9167-ed8be81ba3b9 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port eb8ff994-60f5-461b-b984-ff082644aebc, please check neutron logs for more information. [ 699.900273] env[62066]: ERROR nova.compute.manager [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] Traceback (most recent call last): [ 699.900273] env[62066]: ERROR nova.compute.manager [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 699.900273] env[62066]: ERROR nova.compute.manager [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] yield resources [ 699.900273] env[62066]: ERROR nova.compute.manager [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 699.900273] env[62066]: ERROR nova.compute.manager [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] self.driver.spawn(context, instance, image_meta, [ 699.900273] env[62066]: ERROR nova.compute.manager [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 699.900273] env[62066]: ERROR nova.compute.manager [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] self._vmops.spawn(context, instance, image_meta, injected_files, [ 699.900273] env[62066]: ERROR nova.compute.manager [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 699.900273] env[62066]: ERROR nova.compute.manager [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] vm_ref = self.build_virtual_machine(instance, [ 699.900736] env[62066]: ERROR nova.compute.manager [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 699.900736] env[62066]: ERROR nova.compute.manager [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] vif_infos = vmwarevif.get_vif_info(self._session, [ 699.900736] env[62066]: ERROR nova.compute.manager [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 699.900736] env[62066]: ERROR nova.compute.manager [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] for vif in network_info: [ 699.900736] env[62066]: ERROR nova.compute.manager [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 699.900736] env[62066]: ERROR nova.compute.manager [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] return self._sync_wrapper(fn, *args, **kwargs) [ 699.900736] env[62066]: ERROR nova.compute.manager [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 699.900736] env[62066]: ERROR nova.compute.manager [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] self.wait() [ 699.900736] env[62066]: ERROR nova.compute.manager [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 699.900736] env[62066]: ERROR nova.compute.manager [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] self[:] = self._gt.wait() [ 699.900736] env[62066]: ERROR nova.compute.manager [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 699.900736] env[62066]: ERROR nova.compute.manager [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] return self._exit_event.wait() [ 699.900736] env[62066]: ERROR nova.compute.manager [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 699.901170] env[62066]: ERROR nova.compute.manager [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] result = hub.switch() [ 699.901170] env[62066]: ERROR nova.compute.manager [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 699.901170] env[62066]: ERROR nova.compute.manager [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] return self.greenlet.switch() [ 699.901170] env[62066]: ERROR nova.compute.manager [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 699.901170] env[62066]: ERROR nova.compute.manager [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] result = function(*args, **kwargs) [ 699.901170] env[62066]: ERROR nova.compute.manager [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 699.901170] env[62066]: ERROR nova.compute.manager [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] return func(*args, **kwargs) [ 699.901170] env[62066]: ERROR nova.compute.manager [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 699.901170] env[62066]: ERROR nova.compute.manager [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] raise e [ 699.901170] env[62066]: ERROR nova.compute.manager [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 699.901170] env[62066]: ERROR nova.compute.manager [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] nwinfo = self.network_api.allocate_for_instance( [ 699.901170] env[62066]: ERROR nova.compute.manager [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 699.901170] env[62066]: ERROR nova.compute.manager [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] created_port_ids = self._update_ports_for_instance( [ 699.901584] env[62066]: ERROR nova.compute.manager [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 699.901584] env[62066]: ERROR nova.compute.manager [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] with excutils.save_and_reraise_exception(): [ 699.901584] env[62066]: ERROR nova.compute.manager [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 699.901584] env[62066]: ERROR nova.compute.manager [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] self.force_reraise() [ 699.901584] env[62066]: ERROR nova.compute.manager [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 699.901584] env[62066]: ERROR nova.compute.manager [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] raise self.value [ 699.901584] env[62066]: ERROR nova.compute.manager [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 699.901584] env[62066]: ERROR nova.compute.manager [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] updated_port = self._update_port( [ 699.901584] env[62066]: ERROR nova.compute.manager [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 699.901584] env[62066]: ERROR nova.compute.manager [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] _ensure_no_port_binding_failure(port) [ 699.901584] env[62066]: ERROR nova.compute.manager [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 699.901584] env[62066]: ERROR nova.compute.manager [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] raise exception.PortBindingFailed(port_id=port['id']) [ 699.901975] env[62066]: ERROR nova.compute.manager [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] nova.exception.PortBindingFailed: Binding failed for port eb8ff994-60f5-461b-b984-ff082644aebc, please check neutron logs for more information. [ 699.901975] env[62066]: ERROR nova.compute.manager [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] [ 699.901975] env[62066]: INFO nova.compute.manager [None req-b2b06eaa-4237-476b-9167-ed8be81ba3b9 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] Terminating instance [ 699.903186] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b2b06eaa-4237-476b-9167-ed8be81ba3b9 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Acquiring lock "refresh_cache-f9a8e863-6e7f-4f00-b54a-78802659cd59" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 700.009022] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b96d7111-6e60-4fd2-a38a-3863242412fb tempest-ServerActionsTestOtherB-426172427 tempest-ServerActionsTestOtherB-426172427-project-member] Releasing lock "refresh_cache-e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 700.009022] env[62066]: DEBUG nova.compute.manager [None req-b96d7111-6e60-4fd2-a38a-3863242412fb tempest-ServerActionsTestOtherB-426172427 tempest-ServerActionsTestOtherB-426172427-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 700.009022] env[62066]: DEBUG nova.compute.manager [None req-b96d7111-6e60-4fd2-a38a-3863242412fb tempest-ServerActionsTestOtherB-426172427 tempest-ServerActionsTestOtherB-426172427-project-member] [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 700.009022] env[62066]: DEBUG nova.network.neutron [None req-b96d7111-6e60-4fd2-a38a-3863242412fb tempest-ServerActionsTestOtherB-426172427 tempest-ServerActionsTestOtherB-426172427-project-member] [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 700.037752] env[62066]: DEBUG nova.network.neutron [None req-b96d7111-6e60-4fd2-a38a-3863242412fb tempest-ServerActionsTestOtherB-426172427 tempest-ServerActionsTestOtherB-426172427-project-member] [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 700.119467] env[62066]: DEBUG nova.network.neutron [req-a6db693b-555a-4b5c-a71d-c3c3d26b2bfb req-f760474b-e6de-4dd1-9188-58a92163c5a4 service nova] [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 700.304619] env[62066]: DEBUG nova.network.neutron [req-a6db693b-555a-4b5c-a71d-c3c3d26b2bfb req-f760474b-e6de-4dd1-9188-58a92163c5a4 service nova] [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 700.372139] env[62066]: DEBUG nova.scheduler.client.report [None req-3515d98a-2869-41bb-bf8e-e38d995a4d22 tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 700.539880] env[62066]: DEBUG nova.network.neutron [None req-b96d7111-6e60-4fd2-a38a-3863242412fb tempest-ServerActionsTestOtherB-426172427 tempest-ServerActionsTestOtherB-426172427-project-member] [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 700.806089] env[62066]: DEBUG oslo_concurrency.lockutils [req-a6db693b-555a-4b5c-a71d-c3c3d26b2bfb req-f760474b-e6de-4dd1-9188-58a92163c5a4 service nova] Releasing lock "refresh_cache-f9a8e863-6e7f-4f00-b54a-78802659cd59" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 700.806445] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b2b06eaa-4237-476b-9167-ed8be81ba3b9 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Acquired lock "refresh_cache-f9a8e863-6e7f-4f00-b54a-78802659cd59" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 700.806725] env[62066]: DEBUG nova.network.neutron [None req-b2b06eaa-4237-476b-9167-ed8be81ba3b9 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 700.875878] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3515d98a-2869-41bb-bf8e-e38d995a4d22 tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.982s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 700.876534] env[62066]: ERROR nova.compute.manager [None req-3515d98a-2869-41bb-bf8e-e38d995a4d22 tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port e43c5325-d9ab-45b3-9e9e-e4736fb5735f, please check neutron logs for more information. [ 700.876534] env[62066]: ERROR nova.compute.manager [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] Traceback (most recent call last): [ 700.876534] env[62066]: ERROR nova.compute.manager [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 700.876534] env[62066]: ERROR nova.compute.manager [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] self.driver.spawn(context, instance, image_meta, [ 700.876534] env[62066]: ERROR nova.compute.manager [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 700.876534] env[62066]: ERROR nova.compute.manager [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] self._vmops.spawn(context, instance, image_meta, injected_files, [ 700.876534] env[62066]: ERROR nova.compute.manager [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 700.876534] env[62066]: ERROR nova.compute.manager [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] vm_ref = self.build_virtual_machine(instance, [ 700.876534] env[62066]: ERROR nova.compute.manager [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 700.876534] env[62066]: ERROR nova.compute.manager [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] vif_infos = vmwarevif.get_vif_info(self._session, [ 700.876534] env[62066]: ERROR nova.compute.manager [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 700.876930] env[62066]: ERROR nova.compute.manager [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] for vif in network_info: [ 700.876930] env[62066]: ERROR nova.compute.manager [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 700.876930] env[62066]: ERROR nova.compute.manager [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] return self._sync_wrapper(fn, *args, **kwargs) [ 700.876930] env[62066]: ERROR nova.compute.manager [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 700.876930] env[62066]: ERROR nova.compute.manager [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] self.wait() [ 700.876930] env[62066]: ERROR nova.compute.manager [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 700.876930] env[62066]: ERROR nova.compute.manager [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] self[:] = self._gt.wait() [ 700.876930] env[62066]: ERROR nova.compute.manager [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 700.876930] env[62066]: ERROR nova.compute.manager [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] return self._exit_event.wait() [ 700.876930] env[62066]: ERROR nova.compute.manager [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 700.876930] env[62066]: ERROR nova.compute.manager [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] result = hub.switch() [ 700.876930] env[62066]: ERROR nova.compute.manager [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 700.876930] env[62066]: ERROR nova.compute.manager [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] return self.greenlet.switch() [ 700.877318] env[62066]: ERROR nova.compute.manager [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 700.877318] env[62066]: ERROR nova.compute.manager [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] result = function(*args, **kwargs) [ 700.877318] env[62066]: ERROR nova.compute.manager [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 700.877318] env[62066]: ERROR nova.compute.manager [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] return func(*args, **kwargs) [ 700.877318] env[62066]: ERROR nova.compute.manager [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 700.877318] env[62066]: ERROR nova.compute.manager [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] raise e [ 700.877318] env[62066]: ERROR nova.compute.manager [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 700.877318] env[62066]: ERROR nova.compute.manager [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] nwinfo = self.network_api.allocate_for_instance( [ 700.877318] env[62066]: ERROR nova.compute.manager [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 700.877318] env[62066]: ERROR nova.compute.manager [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] created_port_ids = self._update_ports_for_instance( [ 700.877318] env[62066]: ERROR nova.compute.manager [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 700.877318] env[62066]: ERROR nova.compute.manager [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] with excutils.save_and_reraise_exception(): [ 700.877318] env[62066]: ERROR nova.compute.manager [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 700.877675] env[62066]: ERROR nova.compute.manager [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] self.force_reraise() [ 700.877675] env[62066]: ERROR nova.compute.manager [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 700.877675] env[62066]: ERROR nova.compute.manager [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] raise self.value [ 700.877675] env[62066]: ERROR nova.compute.manager [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 700.877675] env[62066]: ERROR nova.compute.manager [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] updated_port = self._update_port( [ 700.877675] env[62066]: ERROR nova.compute.manager [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 700.877675] env[62066]: ERROR nova.compute.manager [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] _ensure_no_port_binding_failure(port) [ 700.877675] env[62066]: ERROR nova.compute.manager [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 700.877675] env[62066]: ERROR nova.compute.manager [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] raise exception.PortBindingFailed(port_id=port['id']) [ 700.877675] env[62066]: ERROR nova.compute.manager [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] nova.exception.PortBindingFailed: Binding failed for port e43c5325-d9ab-45b3-9e9e-e4736fb5735f, please check neutron logs for more information. [ 700.877675] env[62066]: ERROR nova.compute.manager [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] [ 700.877995] env[62066]: DEBUG nova.compute.utils [None req-3515d98a-2869-41bb-bf8e-e38d995a4d22 tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] Binding failed for port e43c5325-d9ab-45b3-9e9e-e4736fb5735f, please check neutron logs for more information. {{(pid=62066) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 700.878704] env[62066]: DEBUG oslo_concurrency.lockutils [None req-84147b21-d606-4daf-b0b9-f581072be1cb tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.312s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 700.881290] env[62066]: INFO nova.compute.claims [None req-84147b21-d606-4daf-b0b9-f581072be1cb tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 700.882768] env[62066]: DEBUG nova.compute.manager [None req-3515d98a-2869-41bb-bf8e-e38d995a4d22 tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] Build of instance e795e53e-af33-4d0c-8dbd-9c59f236acd7 was re-scheduled: Binding failed for port e43c5325-d9ab-45b3-9e9e-e4736fb5735f, please check neutron logs for more information. {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 700.883227] env[62066]: DEBUG nova.compute.manager [None req-3515d98a-2869-41bb-bf8e-e38d995a4d22 tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] Unplugging VIFs for instance {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 700.885599] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3515d98a-2869-41bb-bf8e-e38d995a4d22 tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Acquiring lock "refresh_cache-e795e53e-af33-4d0c-8dbd-9c59f236acd7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 700.886630] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3515d98a-2869-41bb-bf8e-e38d995a4d22 tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Acquired lock "refresh_cache-e795e53e-af33-4d0c-8dbd-9c59f236acd7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 700.886820] env[62066]: DEBUG nova.network.neutron [None req-3515d98a-2869-41bb-bf8e-e38d995a4d22 tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 701.049329] env[62066]: INFO nova.compute.manager [None req-b96d7111-6e60-4fd2-a38a-3863242412fb tempest-ServerActionsTestOtherB-426172427 tempest-ServerActionsTestOtherB-426172427-project-member] [instance: e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8] Took 1.04 seconds to deallocate network for instance. [ 701.339343] env[62066]: DEBUG nova.network.neutron [None req-b2b06eaa-4237-476b-9167-ed8be81ba3b9 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 701.406494] env[62066]: DEBUG nova.network.neutron [None req-3515d98a-2869-41bb-bf8e-e38d995a4d22 tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 701.434657] env[62066]: DEBUG nova.network.neutron [None req-b2b06eaa-4237-476b-9167-ed8be81ba3b9 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 701.476789] env[62066]: DEBUG nova.network.neutron [None req-3515d98a-2869-41bb-bf8e-e38d995a4d22 tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 701.619809] env[62066]: DEBUG nova.compute.manager [req-a822921b-8f69-4e10-bde6-70a9142cefc8 req-e8771d70-76e8-44dd-8110-fc0eab2ef149 service nova] [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] Received event network-vif-deleted-eb8ff994-60f5-461b-b984-ff082644aebc {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 701.938025] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b2b06eaa-4237-476b-9167-ed8be81ba3b9 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Releasing lock "refresh_cache-f9a8e863-6e7f-4f00-b54a-78802659cd59" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 701.938518] env[62066]: DEBUG nova.compute.manager [None req-b2b06eaa-4237-476b-9167-ed8be81ba3b9 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 701.938715] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-b2b06eaa-4237-476b-9167-ed8be81ba3b9 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 701.941184] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ddc6ebd3-7fb3-4119-8d48-a4199e317088 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.949938] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a5becd9-b9b2-4f5b-b3f9-d9a33f0a882a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.973308] env[62066]: WARNING nova.virt.vmwareapi.vmops [None req-b2b06eaa-4237-476b-9167-ed8be81ba3b9 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance f9a8e863-6e7f-4f00-b54a-78802659cd59 could not be found. [ 701.973537] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-b2b06eaa-4237-476b-9167-ed8be81ba3b9 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 701.973806] env[62066]: INFO nova.compute.manager [None req-b2b06eaa-4237-476b-9167-ed8be81ba3b9 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] Took 0.04 seconds to destroy the instance on the hypervisor. [ 701.974144] env[62066]: DEBUG oslo.service.loopingcall [None req-b2b06eaa-4237-476b-9167-ed8be81ba3b9 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 701.976512] env[62066]: DEBUG nova.compute.manager [-] [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 701.976606] env[62066]: DEBUG nova.network.neutron [-] [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 701.979630] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3515d98a-2869-41bb-bf8e-e38d995a4d22 tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Releasing lock "refresh_cache-e795e53e-af33-4d0c-8dbd-9c59f236acd7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 701.979829] env[62066]: DEBUG nova.compute.manager [None req-3515d98a-2869-41bb-bf8e-e38d995a4d22 tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 701.979998] env[62066]: DEBUG nova.compute.manager [None req-3515d98a-2869-41bb-bf8e-e38d995a4d22 tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 701.980201] env[62066]: DEBUG nova.network.neutron [None req-3515d98a-2869-41bb-bf8e-e38d995a4d22 tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 701.992346] env[62066]: DEBUG nova.network.neutron [-] [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 702.003394] env[62066]: DEBUG nova.network.neutron [None req-3515d98a-2869-41bb-bf8e-e38d995a4d22 tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 702.078565] env[62066]: INFO nova.scheduler.client.report [None req-b96d7111-6e60-4fd2-a38a-3863242412fb tempest-ServerActionsTestOtherB-426172427 tempest-ServerActionsTestOtherB-426172427-project-member] Deleted allocations for instance e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8 [ 702.302735] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9b483e8-f497-40af-879d-ad4150e23f6f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.311558] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8445a2af-0300-485e-aac2-2e10397ee5db {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.342799] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c3081b4-b190-4715-885d-d3a3b83aa4be {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.350827] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ff0c670-06d1-43c4-bb5f-f860fe882a22 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.364265] env[62066]: DEBUG nova.compute.provider_tree [None req-84147b21-d606-4daf-b0b9-f581072be1cb tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 702.494695] env[62066]: DEBUG nova.network.neutron [-] [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 702.507488] env[62066]: DEBUG nova.network.neutron [None req-3515d98a-2869-41bb-bf8e-e38d995a4d22 tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 702.586736] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b96d7111-6e60-4fd2-a38a-3863242412fb tempest-ServerActionsTestOtherB-426172427 tempest-ServerActionsTestOtherB-426172427-project-member] Lock "e5f6387c-b941-4b3e-8c9a-8a643f7dd4b8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 155.047s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 702.867879] env[62066]: DEBUG nova.scheduler.client.report [None req-84147b21-d606-4daf-b0b9-f581072be1cb tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 702.997224] env[62066]: INFO nova.compute.manager [-] [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] Took 1.02 seconds to deallocate network for instance. [ 702.999584] env[62066]: DEBUG nova.compute.claims [None req-b2b06eaa-4237-476b-9167-ed8be81ba3b9 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] Aborting claim: {{(pid=62066) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 702.999763] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b2b06eaa-4237-476b-9167-ed8be81ba3b9 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 703.009392] env[62066]: INFO nova.compute.manager [None req-3515d98a-2869-41bb-bf8e-e38d995a4d22 tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] [instance: e795e53e-af33-4d0c-8dbd-9c59f236acd7] Took 1.03 seconds to deallocate network for instance. [ 703.090670] env[62066]: DEBUG nova.compute.manager [None req-5bf8bcd1-f041-4648-9b3c-4febff568fa9 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] [instance: 3624ca43-9ee3-418a-9205-debb699295df] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 703.374360] env[62066]: DEBUG oslo_concurrency.lockutils [None req-84147b21-d606-4daf-b0b9-f581072be1cb tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.495s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 703.374360] env[62066]: DEBUG nova.compute.manager [None req-84147b21-d606-4daf-b0b9-f581072be1cb tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 703.377562] env[62066]: DEBUG oslo_concurrency.lockutils [None req-60a20a80-792e-4a03-84da-c17f8b28425d tempest-InstanceActionsNegativeTestJSON-1895685779 tempest-InstanceActionsNegativeTestJSON-1895685779-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.352s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 703.616256] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5bf8bcd1-f041-4648-9b3c-4febff568fa9 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 703.880357] env[62066]: DEBUG nova.compute.utils [None req-84147b21-d606-4daf-b0b9-f581072be1cb tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 703.881730] env[62066]: DEBUG nova.compute.manager [None req-84147b21-d606-4daf-b0b9-f581072be1cb tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 703.882099] env[62066]: DEBUG nova.network.neutron [None req-84147b21-d606-4daf-b0b9-f581072be1cb tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 703.942686] env[62066]: DEBUG nova.policy [None req-84147b21-d606-4daf-b0b9-f581072be1cb tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a00e7a5187d949e9a8924e9fda469dff', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'dd07c5b9369f4803900c3974049ecf33', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 704.044358] env[62066]: INFO nova.scheduler.client.report [None req-3515d98a-2869-41bb-bf8e-e38d995a4d22 tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Deleted allocations for instance e795e53e-af33-4d0c-8dbd-9c59f236acd7 [ 704.250968] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b60af39-4690-45ff-9e5b-b5d6858da075 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.257918] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-778ce0da-1bdc-4c9a-a7ac-ed3e6959834a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.290296] env[62066]: DEBUG nova.network.neutron [None req-84147b21-d606-4daf-b0b9-f581072be1cb tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] Successfully created port: 11527383-8bd1-4760-9702-997440e384c8 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 704.293025] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c31797c-b11b-472b-becc-63bb3d43bae0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.300124] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73f554d1-3b83-407d-9175-a191cf48341e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.313267] env[62066]: DEBUG nova.compute.provider_tree [None req-60a20a80-792e-4a03-84da-c17f8b28425d tempest-InstanceActionsNegativeTestJSON-1895685779 tempest-InstanceActionsNegativeTestJSON-1895685779-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 704.390456] env[62066]: DEBUG nova.compute.manager [None req-84147b21-d606-4daf-b0b9-f581072be1cb tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 704.555391] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3515d98a-2869-41bb-bf8e-e38d995a4d22 tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Lock "e795e53e-af33-4d0c-8dbd-9c59f236acd7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 151.605s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 704.817194] env[62066]: DEBUG nova.scheduler.client.report [None req-60a20a80-792e-4a03-84da-c17f8b28425d tempest-InstanceActionsNegativeTestJSON-1895685779 tempest-InstanceActionsNegativeTestJSON-1895685779-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 705.058946] env[62066]: DEBUG nova.compute.manager [None req-7d1d419b-6e61-4f85-9698-fe0c6e4f3988 tempest-AttachInterfacesV270Test-236309384 tempest-AttachInterfacesV270Test-236309384-project-member] [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 705.323697] env[62066]: DEBUG oslo_concurrency.lockutils [None req-60a20a80-792e-4a03-84da-c17f8b28425d tempest-InstanceActionsNegativeTestJSON-1895685779 tempest-InstanceActionsNegativeTestJSON-1895685779-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.945s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 705.324442] env[62066]: ERROR nova.compute.manager [None req-60a20a80-792e-4a03-84da-c17f8b28425d tempest-InstanceActionsNegativeTestJSON-1895685779 tempest-InstanceActionsNegativeTestJSON-1895685779-project-member] [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port df0ac18e-7f9a-44a1-98ac-693de6d5fa5f, please check neutron logs for more information. [ 705.324442] env[62066]: ERROR nova.compute.manager [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] Traceback (most recent call last): [ 705.324442] env[62066]: ERROR nova.compute.manager [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 705.324442] env[62066]: ERROR nova.compute.manager [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] self.driver.spawn(context, instance, image_meta, [ 705.324442] env[62066]: ERROR nova.compute.manager [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 705.324442] env[62066]: ERROR nova.compute.manager [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 705.324442] env[62066]: ERROR nova.compute.manager [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 705.324442] env[62066]: ERROR nova.compute.manager [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] vm_ref = self.build_virtual_machine(instance, [ 705.324442] env[62066]: ERROR nova.compute.manager [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 705.324442] env[62066]: ERROR nova.compute.manager [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] vif_infos = vmwarevif.get_vif_info(self._session, [ 705.324442] env[62066]: ERROR nova.compute.manager [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 705.324906] env[62066]: ERROR nova.compute.manager [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] for vif in network_info: [ 705.324906] env[62066]: ERROR nova.compute.manager [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 705.324906] env[62066]: ERROR nova.compute.manager [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] return self._sync_wrapper(fn, *args, **kwargs) [ 705.324906] env[62066]: ERROR nova.compute.manager [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 705.324906] env[62066]: ERROR nova.compute.manager [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] self.wait() [ 705.324906] env[62066]: ERROR nova.compute.manager [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 705.324906] env[62066]: ERROR nova.compute.manager [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] self[:] = self._gt.wait() [ 705.324906] env[62066]: ERROR nova.compute.manager [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 705.324906] env[62066]: ERROR nova.compute.manager [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] return self._exit_event.wait() [ 705.324906] env[62066]: ERROR nova.compute.manager [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 705.324906] env[62066]: ERROR nova.compute.manager [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] result = hub.switch() [ 705.324906] env[62066]: ERROR nova.compute.manager [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 705.324906] env[62066]: ERROR nova.compute.manager [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] return self.greenlet.switch() [ 705.325411] env[62066]: ERROR nova.compute.manager [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 705.325411] env[62066]: ERROR nova.compute.manager [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] result = function(*args, **kwargs) [ 705.325411] env[62066]: ERROR nova.compute.manager [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 705.325411] env[62066]: ERROR nova.compute.manager [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] return func(*args, **kwargs) [ 705.325411] env[62066]: ERROR nova.compute.manager [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 705.325411] env[62066]: ERROR nova.compute.manager [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] raise e [ 705.325411] env[62066]: ERROR nova.compute.manager [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 705.325411] env[62066]: ERROR nova.compute.manager [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] nwinfo = self.network_api.allocate_for_instance( [ 705.325411] env[62066]: ERROR nova.compute.manager [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 705.325411] env[62066]: ERROR nova.compute.manager [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] created_port_ids = self._update_ports_for_instance( [ 705.325411] env[62066]: ERROR nova.compute.manager [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 705.325411] env[62066]: ERROR nova.compute.manager [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] with excutils.save_and_reraise_exception(): [ 705.325411] env[62066]: ERROR nova.compute.manager [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 705.325835] env[62066]: ERROR nova.compute.manager [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] self.force_reraise() [ 705.325835] env[62066]: ERROR nova.compute.manager [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 705.325835] env[62066]: ERROR nova.compute.manager [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] raise self.value [ 705.325835] env[62066]: ERROR nova.compute.manager [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 705.325835] env[62066]: ERROR nova.compute.manager [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] updated_port = self._update_port( [ 705.325835] env[62066]: ERROR nova.compute.manager [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 705.325835] env[62066]: ERROR nova.compute.manager [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] _ensure_no_port_binding_failure(port) [ 705.325835] env[62066]: ERROR nova.compute.manager [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 705.325835] env[62066]: ERROR nova.compute.manager [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] raise exception.PortBindingFailed(port_id=port['id']) [ 705.325835] env[62066]: ERROR nova.compute.manager [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] nova.exception.PortBindingFailed: Binding failed for port df0ac18e-7f9a-44a1-98ac-693de6d5fa5f, please check neutron logs for more information. [ 705.325835] env[62066]: ERROR nova.compute.manager [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] [ 705.326817] env[62066]: DEBUG nova.compute.utils [None req-60a20a80-792e-4a03-84da-c17f8b28425d tempest-InstanceActionsNegativeTestJSON-1895685779 tempest-InstanceActionsNegativeTestJSON-1895685779-project-member] [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] Binding failed for port df0ac18e-7f9a-44a1-98ac-693de6d5fa5f, please check neutron logs for more information. {{(pid=62066) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 705.328169] env[62066]: DEBUG oslo_concurrency.lockutils [None req-19be2ac0-adbe-403f-ab5a-6d1895be1574 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.313s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 705.332629] env[62066]: DEBUG nova.compute.manager [None req-60a20a80-792e-4a03-84da-c17f8b28425d tempest-InstanceActionsNegativeTestJSON-1895685779 tempest-InstanceActionsNegativeTestJSON-1895685779-project-member] [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] Build of instance a690ad15-2878-4fa7-a810-d155ce6b1a8c was re-scheduled: Binding failed for port df0ac18e-7f9a-44a1-98ac-693de6d5fa5f, please check neutron logs for more information. {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 705.333269] env[62066]: DEBUG nova.compute.manager [None req-60a20a80-792e-4a03-84da-c17f8b28425d tempest-InstanceActionsNegativeTestJSON-1895685779 tempest-InstanceActionsNegativeTestJSON-1895685779-project-member] [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] Unplugging VIFs for instance {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 705.334690] env[62066]: DEBUG oslo_concurrency.lockutils [None req-60a20a80-792e-4a03-84da-c17f8b28425d tempest-InstanceActionsNegativeTestJSON-1895685779 tempest-InstanceActionsNegativeTestJSON-1895685779-project-member] Acquiring lock "refresh_cache-a690ad15-2878-4fa7-a810-d155ce6b1a8c" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 705.334690] env[62066]: DEBUG oslo_concurrency.lockutils [None req-60a20a80-792e-4a03-84da-c17f8b28425d tempest-InstanceActionsNegativeTestJSON-1895685779 tempest-InstanceActionsNegativeTestJSON-1895685779-project-member] Acquired lock "refresh_cache-a690ad15-2878-4fa7-a810-d155ce6b1a8c" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 705.334690] env[62066]: DEBUG nova.network.neutron [None req-60a20a80-792e-4a03-84da-c17f8b28425d tempest-InstanceActionsNegativeTestJSON-1895685779 tempest-InstanceActionsNegativeTestJSON-1895685779-project-member] [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 705.400704] env[62066]: DEBUG nova.compute.manager [None req-84147b21-d606-4daf-b0b9-f581072be1cb tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 705.405662] env[62066]: DEBUG nova.compute.manager [req-3cd71441-7b02-495b-9b29-b0fae90d6554 req-233cb060-a372-4250-9b55-a1bd2c4b9b1b service nova] [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] Received event network-changed-11527383-8bd1-4760-9702-997440e384c8 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 705.406232] env[62066]: DEBUG nova.compute.manager [req-3cd71441-7b02-495b-9b29-b0fae90d6554 req-233cb060-a372-4250-9b55-a1bd2c4b9b1b service nova] [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] Refreshing instance network info cache due to event network-changed-11527383-8bd1-4760-9702-997440e384c8. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 705.407663] env[62066]: DEBUG oslo_concurrency.lockutils [req-3cd71441-7b02-495b-9b29-b0fae90d6554 req-233cb060-a372-4250-9b55-a1bd2c4b9b1b service nova] Acquiring lock "refresh_cache-2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 705.407663] env[62066]: DEBUG oslo_concurrency.lockutils [req-3cd71441-7b02-495b-9b29-b0fae90d6554 req-233cb060-a372-4250-9b55-a1bd2c4b9b1b service nova] Acquired lock "refresh_cache-2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 705.407663] env[62066]: DEBUG nova.network.neutron [req-3cd71441-7b02-495b-9b29-b0fae90d6554 req-233cb060-a372-4250-9b55-a1bd2c4b9b1b service nova] [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] Refreshing network info cache for port 11527383-8bd1-4760-9702-997440e384c8 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 705.429193] env[62066]: DEBUG nova.virt.hardware [None req-84147b21-d606-4daf-b0b9-f581072be1cb tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-23T13:40:41Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-23T13:40:23Z,direct_url=,disk_format='vmdk',id=50ff584c-3b50-4395-af07-3e66769bc9f7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='eb52f7069a374c61ae946f052007c6d9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-23T13:40:24Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 705.429455] env[62066]: DEBUG nova.virt.hardware [None req-84147b21-d606-4daf-b0b9-f581072be1cb tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 705.429617] env[62066]: DEBUG nova.virt.hardware [None req-84147b21-d606-4daf-b0b9-f581072be1cb tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 705.429793] env[62066]: DEBUG nova.virt.hardware [None req-84147b21-d606-4daf-b0b9-f581072be1cb tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 705.430045] env[62066]: DEBUG nova.virt.hardware [None req-84147b21-d606-4daf-b0b9-f581072be1cb tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 705.430167] env[62066]: DEBUG nova.virt.hardware [None req-84147b21-d606-4daf-b0b9-f581072be1cb tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 705.430371] env[62066]: DEBUG nova.virt.hardware [None req-84147b21-d606-4daf-b0b9-f581072be1cb tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 705.430524] env[62066]: DEBUG nova.virt.hardware [None req-84147b21-d606-4daf-b0b9-f581072be1cb tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 705.430683] env[62066]: DEBUG nova.virt.hardware [None req-84147b21-d606-4daf-b0b9-f581072be1cb tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 705.430838] env[62066]: DEBUG nova.virt.hardware [None req-84147b21-d606-4daf-b0b9-f581072be1cb tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 705.431010] env[62066]: DEBUG nova.virt.hardware [None req-84147b21-d606-4daf-b0b9-f581072be1cb tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 705.432189] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41e7fb77-9171-45de-96fa-fb7817b902b7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.440385] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0ce35ec-91fc-4290-ae58-1c962e459242 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.578930] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7d1d419b-6e61-4f85-9698-fe0c6e4f3988 tempest-AttachInterfacesV270Test-236309384 tempest-AttachInterfacesV270Test-236309384-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 705.652413] env[62066]: ERROR nova.compute.manager [None req-84147b21-d606-4daf-b0b9-f581072be1cb tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 11527383-8bd1-4760-9702-997440e384c8, please check neutron logs for more information. [ 705.652413] env[62066]: ERROR nova.compute.manager Traceback (most recent call last): [ 705.652413] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 705.652413] env[62066]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 705.652413] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 705.652413] env[62066]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 705.652413] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 705.652413] env[62066]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 705.652413] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 705.652413] env[62066]: ERROR nova.compute.manager self.force_reraise() [ 705.652413] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 705.652413] env[62066]: ERROR nova.compute.manager raise self.value [ 705.652413] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 705.652413] env[62066]: ERROR nova.compute.manager updated_port = self._update_port( [ 705.652413] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 705.652413] env[62066]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 705.653184] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 705.653184] env[62066]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 705.653184] env[62066]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 11527383-8bd1-4760-9702-997440e384c8, please check neutron logs for more information. [ 705.653184] env[62066]: ERROR nova.compute.manager [ 705.653184] env[62066]: Traceback (most recent call last): [ 705.653184] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 705.653184] env[62066]: listener.cb(fileno) [ 705.653184] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 705.653184] env[62066]: result = function(*args, **kwargs) [ 705.653184] env[62066]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 705.653184] env[62066]: return func(*args, **kwargs) [ 705.653184] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 705.653184] env[62066]: raise e [ 705.653184] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 705.653184] env[62066]: nwinfo = self.network_api.allocate_for_instance( [ 705.653184] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 705.653184] env[62066]: created_port_ids = self._update_ports_for_instance( [ 705.653184] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 705.653184] env[62066]: with excutils.save_and_reraise_exception(): [ 705.653184] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 705.653184] env[62066]: self.force_reraise() [ 705.653184] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 705.653184] env[62066]: raise self.value [ 705.653184] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 705.653184] env[62066]: updated_port = self._update_port( [ 705.653184] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 705.653184] env[62066]: _ensure_no_port_binding_failure(port) [ 705.653184] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 705.653184] env[62066]: raise exception.PortBindingFailed(port_id=port['id']) [ 705.654855] env[62066]: nova.exception.PortBindingFailed: Binding failed for port 11527383-8bd1-4760-9702-997440e384c8, please check neutron logs for more information. [ 705.654855] env[62066]: Removing descriptor: 20 [ 705.654855] env[62066]: ERROR nova.compute.manager [None req-84147b21-d606-4daf-b0b9-f581072be1cb tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 11527383-8bd1-4760-9702-997440e384c8, please check neutron logs for more information. [ 705.654855] env[62066]: ERROR nova.compute.manager [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] Traceback (most recent call last): [ 705.654855] env[62066]: ERROR nova.compute.manager [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 705.654855] env[62066]: ERROR nova.compute.manager [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] yield resources [ 705.654855] env[62066]: ERROR nova.compute.manager [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 705.654855] env[62066]: ERROR nova.compute.manager [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] self.driver.spawn(context, instance, image_meta, [ 705.654855] env[62066]: ERROR nova.compute.manager [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 705.654855] env[62066]: ERROR nova.compute.manager [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] self._vmops.spawn(context, instance, image_meta, injected_files, [ 705.654855] env[62066]: ERROR nova.compute.manager [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 705.654855] env[62066]: ERROR nova.compute.manager [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] vm_ref = self.build_virtual_machine(instance, [ 705.655418] env[62066]: ERROR nova.compute.manager [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 705.655418] env[62066]: ERROR nova.compute.manager [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] vif_infos = vmwarevif.get_vif_info(self._session, [ 705.655418] env[62066]: ERROR nova.compute.manager [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 705.655418] env[62066]: ERROR nova.compute.manager [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] for vif in network_info: [ 705.655418] env[62066]: ERROR nova.compute.manager [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 705.655418] env[62066]: ERROR nova.compute.manager [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] return self._sync_wrapper(fn, *args, **kwargs) [ 705.655418] env[62066]: ERROR nova.compute.manager [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 705.655418] env[62066]: ERROR nova.compute.manager [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] self.wait() [ 705.655418] env[62066]: ERROR nova.compute.manager [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 705.655418] env[62066]: ERROR nova.compute.manager [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] self[:] = self._gt.wait() [ 705.655418] env[62066]: ERROR nova.compute.manager [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 705.655418] env[62066]: ERROR nova.compute.manager [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] return self._exit_event.wait() [ 705.655418] env[62066]: ERROR nova.compute.manager [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 705.656017] env[62066]: ERROR nova.compute.manager [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] result = hub.switch() [ 705.656017] env[62066]: ERROR nova.compute.manager [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 705.656017] env[62066]: ERROR nova.compute.manager [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] return self.greenlet.switch() [ 705.656017] env[62066]: ERROR nova.compute.manager [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 705.656017] env[62066]: ERROR nova.compute.manager [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] result = function(*args, **kwargs) [ 705.656017] env[62066]: ERROR nova.compute.manager [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 705.656017] env[62066]: ERROR nova.compute.manager [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] return func(*args, **kwargs) [ 705.656017] env[62066]: ERROR nova.compute.manager [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 705.656017] env[62066]: ERROR nova.compute.manager [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] raise e [ 705.656017] env[62066]: ERROR nova.compute.manager [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 705.656017] env[62066]: ERROR nova.compute.manager [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] nwinfo = self.network_api.allocate_for_instance( [ 705.656017] env[62066]: ERROR nova.compute.manager [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 705.656017] env[62066]: ERROR nova.compute.manager [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] created_port_ids = self._update_ports_for_instance( [ 705.656635] env[62066]: ERROR nova.compute.manager [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 705.656635] env[62066]: ERROR nova.compute.manager [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] with excutils.save_and_reraise_exception(): [ 705.656635] env[62066]: ERROR nova.compute.manager [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 705.656635] env[62066]: ERROR nova.compute.manager [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] self.force_reraise() [ 705.656635] env[62066]: ERROR nova.compute.manager [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 705.656635] env[62066]: ERROR nova.compute.manager [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] raise self.value [ 705.656635] env[62066]: ERROR nova.compute.manager [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 705.656635] env[62066]: ERROR nova.compute.manager [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] updated_port = self._update_port( [ 705.656635] env[62066]: ERROR nova.compute.manager [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 705.656635] env[62066]: ERROR nova.compute.manager [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] _ensure_no_port_binding_failure(port) [ 705.656635] env[62066]: ERROR nova.compute.manager [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 705.656635] env[62066]: ERROR nova.compute.manager [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] raise exception.PortBindingFailed(port_id=port['id']) [ 705.657176] env[62066]: ERROR nova.compute.manager [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] nova.exception.PortBindingFailed: Binding failed for port 11527383-8bd1-4760-9702-997440e384c8, please check neutron logs for more information. [ 705.657176] env[62066]: ERROR nova.compute.manager [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] [ 705.657176] env[62066]: INFO nova.compute.manager [None req-84147b21-d606-4daf-b0b9-f581072be1cb tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] Terminating instance [ 705.657176] env[62066]: DEBUG oslo_concurrency.lockutils [None req-84147b21-d606-4daf-b0b9-f581072be1cb tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] Acquiring lock "refresh_cache-2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 705.860955] env[62066]: DEBUG nova.network.neutron [None req-60a20a80-792e-4a03-84da-c17f8b28425d tempest-InstanceActionsNegativeTestJSON-1895685779 tempest-InstanceActionsNegativeTestJSON-1895685779-project-member] [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 705.930923] env[62066]: DEBUG nova.network.neutron [req-3cd71441-7b02-495b-9b29-b0fae90d6554 req-233cb060-a372-4250-9b55-a1bd2c4b9b1b service nova] [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 706.007507] env[62066]: DEBUG nova.network.neutron [req-3cd71441-7b02-495b-9b29-b0fae90d6554 req-233cb060-a372-4250-9b55-a1bd2c4b9b1b service nova] [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 706.020141] env[62066]: DEBUG nova.network.neutron [None req-60a20a80-792e-4a03-84da-c17f8b28425d tempest-InstanceActionsNegativeTestJSON-1895685779 tempest-InstanceActionsNegativeTestJSON-1895685779-project-member] [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 706.189652] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a503e74-0801-4b80-9cc5-b2a179819820 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.196246] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4490ad3c-a573-4c11-a982-c095c3cc155c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.225471] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3db56d72-b727-48f6-8fdb-0891f91d4348 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.232517] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5f8db69-590a-4716-81c4-becbdb2268da {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.247267] env[62066]: DEBUG nova.compute.provider_tree [None req-19be2ac0-adbe-403f-ab5a-6d1895be1574 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 706.510028] env[62066]: DEBUG oslo_concurrency.lockutils [req-3cd71441-7b02-495b-9b29-b0fae90d6554 req-233cb060-a372-4250-9b55-a1bd2c4b9b1b service nova] Releasing lock "refresh_cache-2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 706.510478] env[62066]: DEBUG oslo_concurrency.lockutils [None req-84147b21-d606-4daf-b0b9-f581072be1cb tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] Acquired lock "refresh_cache-2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 706.510681] env[62066]: DEBUG nova.network.neutron [None req-84147b21-d606-4daf-b0b9-f581072be1cb tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 706.522914] env[62066]: DEBUG oslo_concurrency.lockutils [None req-60a20a80-792e-4a03-84da-c17f8b28425d tempest-InstanceActionsNegativeTestJSON-1895685779 tempest-InstanceActionsNegativeTestJSON-1895685779-project-member] Releasing lock "refresh_cache-a690ad15-2878-4fa7-a810-d155ce6b1a8c" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 706.522914] env[62066]: DEBUG nova.compute.manager [None req-60a20a80-792e-4a03-84da-c17f8b28425d tempest-InstanceActionsNegativeTestJSON-1895685779 tempest-InstanceActionsNegativeTestJSON-1895685779-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 706.523124] env[62066]: DEBUG nova.compute.manager [None req-60a20a80-792e-4a03-84da-c17f8b28425d tempest-InstanceActionsNegativeTestJSON-1895685779 tempest-InstanceActionsNegativeTestJSON-1895685779-project-member] [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 706.523277] env[62066]: DEBUG nova.network.neutron [None req-60a20a80-792e-4a03-84da-c17f8b28425d tempest-InstanceActionsNegativeTestJSON-1895685779 tempest-InstanceActionsNegativeTestJSON-1895685779-project-member] [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 706.539651] env[62066]: DEBUG nova.network.neutron [None req-60a20a80-792e-4a03-84da-c17f8b28425d tempest-InstanceActionsNegativeTestJSON-1895685779 tempest-InstanceActionsNegativeTestJSON-1895685779-project-member] [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 706.753428] env[62066]: DEBUG nova.scheduler.client.report [None req-19be2ac0-adbe-403f-ab5a-6d1895be1574 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 707.032563] env[62066]: DEBUG nova.network.neutron [None req-84147b21-d606-4daf-b0b9-f581072be1cb tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 707.042150] env[62066]: DEBUG nova.network.neutron [None req-60a20a80-792e-4a03-84da-c17f8b28425d tempest-InstanceActionsNegativeTestJSON-1895685779 tempest-InstanceActionsNegativeTestJSON-1895685779-project-member] [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 707.166333] env[62066]: DEBUG nova.network.neutron [None req-84147b21-d606-4daf-b0b9-f581072be1cb tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 707.257352] env[62066]: DEBUG oslo_concurrency.lockutils [None req-19be2ac0-adbe-403f-ab5a-6d1895be1574 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.929s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 707.257948] env[62066]: ERROR nova.compute.manager [None req-19be2ac0-adbe-403f-ab5a-6d1895be1574 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 2ddcf220-1678-4e7d-9792-15f3d090508d, please check neutron logs for more information. [ 707.257948] env[62066]: ERROR nova.compute.manager [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] Traceback (most recent call last): [ 707.257948] env[62066]: ERROR nova.compute.manager [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 707.257948] env[62066]: ERROR nova.compute.manager [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] self.driver.spawn(context, instance, image_meta, [ 707.257948] env[62066]: ERROR nova.compute.manager [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 707.257948] env[62066]: ERROR nova.compute.manager [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] self._vmops.spawn(context, instance, image_meta, injected_files, [ 707.257948] env[62066]: ERROR nova.compute.manager [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 707.257948] env[62066]: ERROR nova.compute.manager [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] vm_ref = self.build_virtual_machine(instance, [ 707.257948] env[62066]: ERROR nova.compute.manager [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 707.257948] env[62066]: ERROR nova.compute.manager [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] vif_infos = vmwarevif.get_vif_info(self._session, [ 707.257948] env[62066]: ERROR nova.compute.manager [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 707.258543] env[62066]: ERROR nova.compute.manager [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] for vif in network_info: [ 707.258543] env[62066]: ERROR nova.compute.manager [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 707.258543] env[62066]: ERROR nova.compute.manager [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] return self._sync_wrapper(fn, *args, **kwargs) [ 707.258543] env[62066]: ERROR nova.compute.manager [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 707.258543] env[62066]: ERROR nova.compute.manager [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] self.wait() [ 707.258543] env[62066]: ERROR nova.compute.manager [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 707.258543] env[62066]: ERROR nova.compute.manager [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] self[:] = self._gt.wait() [ 707.258543] env[62066]: ERROR nova.compute.manager [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 707.258543] env[62066]: ERROR nova.compute.manager [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] return self._exit_event.wait() [ 707.258543] env[62066]: ERROR nova.compute.manager [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 707.258543] env[62066]: ERROR nova.compute.manager [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] current.throw(*self._exc) [ 707.258543] env[62066]: ERROR nova.compute.manager [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 707.258543] env[62066]: ERROR nova.compute.manager [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] result = function(*args, **kwargs) [ 707.259119] env[62066]: ERROR nova.compute.manager [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 707.259119] env[62066]: ERROR nova.compute.manager [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] return func(*args, **kwargs) [ 707.259119] env[62066]: ERROR nova.compute.manager [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 707.259119] env[62066]: ERROR nova.compute.manager [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] raise e [ 707.259119] env[62066]: ERROR nova.compute.manager [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 707.259119] env[62066]: ERROR nova.compute.manager [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] nwinfo = self.network_api.allocate_for_instance( [ 707.259119] env[62066]: ERROR nova.compute.manager [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 707.259119] env[62066]: ERROR nova.compute.manager [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] created_port_ids = self._update_ports_for_instance( [ 707.259119] env[62066]: ERROR nova.compute.manager [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 707.259119] env[62066]: ERROR nova.compute.manager [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] with excutils.save_and_reraise_exception(): [ 707.259119] env[62066]: ERROR nova.compute.manager [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 707.259119] env[62066]: ERROR nova.compute.manager [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] self.force_reraise() [ 707.259119] env[62066]: ERROR nova.compute.manager [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 707.260331] env[62066]: ERROR nova.compute.manager [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] raise self.value [ 707.260331] env[62066]: ERROR nova.compute.manager [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 707.260331] env[62066]: ERROR nova.compute.manager [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] updated_port = self._update_port( [ 707.260331] env[62066]: ERROR nova.compute.manager [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 707.260331] env[62066]: ERROR nova.compute.manager [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] _ensure_no_port_binding_failure(port) [ 707.260331] env[62066]: ERROR nova.compute.manager [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 707.260331] env[62066]: ERROR nova.compute.manager [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] raise exception.PortBindingFailed(port_id=port['id']) [ 707.260331] env[62066]: ERROR nova.compute.manager [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] nova.exception.PortBindingFailed: Binding failed for port 2ddcf220-1678-4e7d-9792-15f3d090508d, please check neutron logs for more information. [ 707.260331] env[62066]: ERROR nova.compute.manager [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] [ 707.260331] env[62066]: DEBUG nova.compute.utils [None req-19be2ac0-adbe-403f-ab5a-6d1895be1574 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] Binding failed for port 2ddcf220-1678-4e7d-9792-15f3d090508d, please check neutron logs for more information. {{(pid=62066) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 707.260906] env[62066]: DEBUG oslo_concurrency.lockutils [None req-8d368cc7-6ac0-44cd-b71e-9136d8337cb4 tempest-AttachInterfacesUnderV243Test-1832696447 tempest-AttachInterfacesUnderV243Test-1832696447-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.749s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 707.261377] env[62066]: INFO nova.compute.claims [None req-8d368cc7-6ac0-44cd-b71e-9136d8337cb4 tempest-AttachInterfacesUnderV243Test-1832696447 tempest-AttachInterfacesUnderV243Test-1832696447-project-member] [instance: f91a90f7-be73-424c-966c-1be6f37a0864] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 707.264025] env[62066]: DEBUG nova.compute.manager [None req-19be2ac0-adbe-403f-ab5a-6d1895be1574 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] Build of instance 5fad2258-f694-4993-bf64-bfa16abc09ba was re-scheduled: Binding failed for port 2ddcf220-1678-4e7d-9792-15f3d090508d, please check neutron logs for more information. {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 707.264442] env[62066]: DEBUG nova.compute.manager [None req-19be2ac0-adbe-403f-ab5a-6d1895be1574 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] Unplugging VIFs for instance {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 707.264665] env[62066]: DEBUG oslo_concurrency.lockutils [None req-19be2ac0-adbe-403f-ab5a-6d1895be1574 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] Acquiring lock "refresh_cache-5fad2258-f694-4993-bf64-bfa16abc09ba" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 707.264808] env[62066]: DEBUG oslo_concurrency.lockutils [None req-19be2ac0-adbe-403f-ab5a-6d1895be1574 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] Acquired lock "refresh_cache-5fad2258-f694-4993-bf64-bfa16abc09ba" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 707.264965] env[62066]: DEBUG nova.network.neutron [None req-19be2ac0-adbe-403f-ab5a-6d1895be1574 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 707.474862] env[62066]: DEBUG nova.compute.manager [req-e98530e1-b403-4f8a-82a2-6e0936b0d3b6 req-852b5bad-65d3-4a88-823c-6128c1613f05 service nova] [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] Received event network-vif-deleted-11527383-8bd1-4760-9702-997440e384c8 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 707.550038] env[62066]: INFO nova.compute.manager [None req-60a20a80-792e-4a03-84da-c17f8b28425d tempest-InstanceActionsNegativeTestJSON-1895685779 tempest-InstanceActionsNegativeTestJSON-1895685779-project-member] [instance: a690ad15-2878-4fa7-a810-d155ce6b1a8c] Took 1.03 seconds to deallocate network for instance. [ 707.670238] env[62066]: DEBUG oslo_concurrency.lockutils [None req-84147b21-d606-4daf-b0b9-f581072be1cb tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] Releasing lock "refresh_cache-2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 707.670673] env[62066]: DEBUG nova.compute.manager [None req-84147b21-d606-4daf-b0b9-f581072be1cb tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 707.670869] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-84147b21-d606-4daf-b0b9-f581072be1cb tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 707.671183] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-98bff0ba-4e3a-4872-b9ce-f3a01d21b1c2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.683521] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b48a9852-e528-40b1-8591-9d2a9c58bb0f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.709177] env[62066]: WARNING nova.virt.vmwareapi.vmops [None req-84147b21-d606-4daf-b0b9-f581072be1cb tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3 could not be found. [ 707.709432] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-84147b21-d606-4daf-b0b9-f581072be1cb tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 707.709627] env[62066]: INFO nova.compute.manager [None req-84147b21-d606-4daf-b0b9-f581072be1cb tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] Took 0.04 seconds to destroy the instance on the hypervisor. [ 707.709871] env[62066]: DEBUG oslo.service.loopingcall [None req-84147b21-d606-4daf-b0b9-f581072be1cb tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 707.710113] env[62066]: DEBUG nova.compute.manager [-] [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 707.710216] env[62066]: DEBUG nova.network.neutron [-] [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 707.741290] env[62066]: DEBUG nova.network.neutron [-] [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 707.791942] env[62066]: DEBUG nova.network.neutron [None req-19be2ac0-adbe-403f-ab5a-6d1895be1574 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 707.864350] env[62066]: DEBUG nova.network.neutron [None req-19be2ac0-adbe-403f-ab5a-6d1895be1574 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 708.242208] env[62066]: DEBUG nova.network.neutron [-] [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 708.296488] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f6cc809d-278e-4fb4-906b-f468fa41739d tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Acquiring lock "5f645eaa-6edc-4362-82ad-38c4c57b2be2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 708.296488] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f6cc809d-278e-4fb4-906b-f468fa41739d tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Lock "5f645eaa-6edc-4362-82ad-38c4c57b2be2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 708.366903] env[62066]: DEBUG oslo_concurrency.lockutils [None req-19be2ac0-adbe-403f-ab5a-6d1895be1574 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] Releasing lock "refresh_cache-5fad2258-f694-4993-bf64-bfa16abc09ba" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 708.367189] env[62066]: DEBUG nova.compute.manager [None req-19be2ac0-adbe-403f-ab5a-6d1895be1574 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 708.367328] env[62066]: DEBUG nova.compute.manager [None req-19be2ac0-adbe-403f-ab5a-6d1895be1574 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 708.367492] env[62066]: DEBUG nova.network.neutron [None req-19be2ac0-adbe-403f-ab5a-6d1895be1574 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 708.381682] env[62066]: DEBUG nova.network.neutron [None req-19be2ac0-adbe-403f-ab5a-6d1895be1574 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 708.574335] env[62066]: INFO nova.scheduler.client.report [None req-60a20a80-792e-4a03-84da-c17f8b28425d tempest-InstanceActionsNegativeTestJSON-1895685779 tempest-InstanceActionsNegativeTestJSON-1895685779-project-member] Deleted allocations for instance a690ad15-2878-4fa7-a810-d155ce6b1a8c [ 708.591185] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff0c646b-fdb1-47c3-941f-a6e7ef371b1b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.599582] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9404f5f-1ea2-4426-b70d-57ed55879dd7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.628714] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b2072d9-03d0-4808-88cc-9961c5a8d838 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.635675] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7cf18d16-56f5-4b84-b3a6-e295f35bd6a7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.649813] env[62066]: DEBUG nova.compute.provider_tree [None req-8d368cc7-6ac0-44cd-b71e-9136d8337cb4 tempest-AttachInterfacesUnderV243Test-1832696447 tempest-AttachInterfacesUnderV243Test-1832696447-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 708.745460] env[62066]: INFO nova.compute.manager [-] [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] Took 1.04 seconds to deallocate network for instance. [ 708.748129] env[62066]: DEBUG nova.compute.claims [None req-84147b21-d606-4daf-b0b9-f581072be1cb tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] Aborting claim: {{(pid=62066) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 708.748129] env[62066]: DEBUG oslo_concurrency.lockutils [None req-84147b21-d606-4daf-b0b9-f581072be1cb tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 708.884662] env[62066]: DEBUG nova.network.neutron [None req-19be2ac0-adbe-403f-ab5a-6d1895be1574 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 709.087693] env[62066]: DEBUG oslo_concurrency.lockutils [None req-60a20a80-792e-4a03-84da-c17f8b28425d tempest-InstanceActionsNegativeTestJSON-1895685779 tempest-InstanceActionsNegativeTestJSON-1895685779-project-member] Lock "a690ad15-2878-4fa7-a810-d155ce6b1a8c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 155.897s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 709.152849] env[62066]: DEBUG nova.scheduler.client.report [None req-8d368cc7-6ac0-44cd-b71e-9136d8337cb4 tempest-AttachInterfacesUnderV243Test-1832696447 tempest-AttachInterfacesUnderV243Test-1832696447-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 709.387550] env[62066]: INFO nova.compute.manager [None req-19be2ac0-adbe-403f-ab5a-6d1895be1574 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] [instance: 5fad2258-f694-4993-bf64-bfa16abc09ba] Took 1.02 seconds to deallocate network for instance. [ 709.590538] env[62066]: DEBUG nova.compute.manager [None req-a4d93091-69b6-4b35-887c-e7b927c2300a tempest-ServerMetadataTestJSON-1590103242 tempest-ServerMetadataTestJSON-1590103242-project-member] [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 709.657816] env[62066]: DEBUG oslo_concurrency.lockutils [None req-8d368cc7-6ac0-44cd-b71e-9136d8337cb4 tempest-AttachInterfacesUnderV243Test-1832696447 tempest-AttachInterfacesUnderV243Test-1832696447-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.398s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 709.658403] env[62066]: DEBUG nova.compute.manager [None req-8d368cc7-6ac0-44cd-b71e-9136d8337cb4 tempest-AttachInterfacesUnderV243Test-1832696447 tempest-AttachInterfacesUnderV243Test-1832696447-project-member] [instance: f91a90f7-be73-424c-966c-1be6f37a0864] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 709.661161] env[62066]: DEBUG oslo_concurrency.lockutils [None req-96f72479-6fed-4ff5-9d17-40fa20c1f416 tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.399s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 710.110772] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a4d93091-69b6-4b35-887c-e7b927c2300a tempest-ServerMetadataTestJSON-1590103242 tempest-ServerMetadataTestJSON-1590103242-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 710.166684] env[62066]: DEBUG nova.compute.utils [None req-8d368cc7-6ac0-44cd-b71e-9136d8337cb4 tempest-AttachInterfacesUnderV243Test-1832696447 tempest-AttachInterfacesUnderV243Test-1832696447-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 710.175540] env[62066]: DEBUG nova.compute.manager [None req-8d368cc7-6ac0-44cd-b71e-9136d8337cb4 tempest-AttachInterfacesUnderV243Test-1832696447 tempest-AttachInterfacesUnderV243Test-1832696447-project-member] [instance: f91a90f7-be73-424c-966c-1be6f37a0864] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 710.176197] env[62066]: DEBUG nova.network.neutron [None req-8d368cc7-6ac0-44cd-b71e-9136d8337cb4 tempest-AttachInterfacesUnderV243Test-1832696447 tempest-AttachInterfacesUnderV243Test-1832696447-project-member] [instance: f91a90f7-be73-424c-966c-1be6f37a0864] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 710.267135] env[62066]: DEBUG nova.policy [None req-8d368cc7-6ac0-44cd-b71e-9136d8337cb4 tempest-AttachInterfacesUnderV243Test-1832696447 tempest-AttachInterfacesUnderV243Test-1832696447-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '244be47a4aba42c396ee9b9919e87cd0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '80f992c26c7d461e8d6fa68c9356824f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 710.422673] env[62066]: INFO nova.scheduler.client.report [None req-19be2ac0-adbe-403f-ab5a-6d1895be1574 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] Deleted allocations for instance 5fad2258-f694-4993-bf64-bfa16abc09ba [ 710.590017] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-083dfec0-9939-482d-8c23-8ea77e2d09b4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.600737] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f780f64-1acd-4477-a541-b2c63192088d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.635567] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-375f7e5e-6554-49aa-83b1-d25f1a03903b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.643011] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39f169a5-ea16-472d-8e27-df3cb42676be {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.661126] env[62066]: DEBUG nova.compute.provider_tree [None req-96f72479-6fed-4ff5-9d17-40fa20c1f416 tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 710.676925] env[62066]: DEBUG nova.compute.manager [None req-8d368cc7-6ac0-44cd-b71e-9136d8337cb4 tempest-AttachInterfacesUnderV243Test-1832696447 tempest-AttachInterfacesUnderV243Test-1832696447-project-member] [instance: f91a90f7-be73-424c-966c-1be6f37a0864] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 710.741762] env[62066]: DEBUG nova.network.neutron [None req-8d368cc7-6ac0-44cd-b71e-9136d8337cb4 tempest-AttachInterfacesUnderV243Test-1832696447 tempest-AttachInterfacesUnderV243Test-1832696447-project-member] [instance: f91a90f7-be73-424c-966c-1be6f37a0864] Successfully created port: 01c976c4-59c2-4855-942f-61915dbd9fe3 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 710.940019] env[62066]: DEBUG oslo_concurrency.lockutils [None req-19be2ac0-adbe-403f-ab5a-6d1895be1574 tempest-MigrationsAdminTest-999999169 tempest-MigrationsAdminTest-999999169-project-member] Lock "5fad2258-f694-4993-bf64-bfa16abc09ba" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 156.587s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 711.167580] env[62066]: DEBUG nova.scheduler.client.report [None req-96f72479-6fed-4ff5-9d17-40fa20c1f416 tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 711.337831] env[62066]: DEBUG oslo_service.periodic_task [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 711.338072] env[62066]: DEBUG oslo_service.periodic_task [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 711.442181] env[62066]: DEBUG nova.compute.manager [None req-e9793203-98b3-4ca0-8a49-79767b1f52f7 tempest-ServersTestBootFromVolume-1535558002 tempest-ServersTestBootFromVolume-1535558002-project-member] [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 711.672409] env[62066]: DEBUG oslo_concurrency.lockutils [None req-96f72479-6fed-4ff5-9d17-40fa20c1f416 tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.011s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 711.673058] env[62066]: ERROR nova.compute.manager [None req-96f72479-6fed-4ff5-9d17-40fa20c1f416 tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 37035015-85cf-4d55-9ba0-068bae5b3dab, please check neutron logs for more information. [ 711.673058] env[62066]: ERROR nova.compute.manager [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] Traceback (most recent call last): [ 711.673058] env[62066]: ERROR nova.compute.manager [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 711.673058] env[62066]: ERROR nova.compute.manager [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] self.driver.spawn(context, instance, image_meta, [ 711.673058] env[62066]: ERROR nova.compute.manager [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 711.673058] env[62066]: ERROR nova.compute.manager [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] self._vmops.spawn(context, instance, image_meta, injected_files, [ 711.673058] env[62066]: ERROR nova.compute.manager [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 711.673058] env[62066]: ERROR nova.compute.manager [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] vm_ref = self.build_virtual_machine(instance, [ 711.673058] env[62066]: ERROR nova.compute.manager [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 711.673058] env[62066]: ERROR nova.compute.manager [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] vif_infos = vmwarevif.get_vif_info(self._session, [ 711.673058] env[62066]: ERROR nova.compute.manager [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 711.673392] env[62066]: ERROR nova.compute.manager [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] for vif in network_info: [ 711.673392] env[62066]: ERROR nova.compute.manager [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 711.673392] env[62066]: ERROR nova.compute.manager [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] return self._sync_wrapper(fn, *args, **kwargs) [ 711.673392] env[62066]: ERROR nova.compute.manager [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 711.673392] env[62066]: ERROR nova.compute.manager [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] self.wait() [ 711.673392] env[62066]: ERROR nova.compute.manager [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 711.673392] env[62066]: ERROR nova.compute.manager [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] self[:] = self._gt.wait() [ 711.673392] env[62066]: ERROR nova.compute.manager [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 711.673392] env[62066]: ERROR nova.compute.manager [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] return self._exit_event.wait() [ 711.673392] env[62066]: ERROR nova.compute.manager [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 711.673392] env[62066]: ERROR nova.compute.manager [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] result = hub.switch() [ 711.673392] env[62066]: ERROR nova.compute.manager [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 711.673392] env[62066]: ERROR nova.compute.manager [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] return self.greenlet.switch() [ 711.673754] env[62066]: ERROR nova.compute.manager [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 711.673754] env[62066]: ERROR nova.compute.manager [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] result = function(*args, **kwargs) [ 711.673754] env[62066]: ERROR nova.compute.manager [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 711.673754] env[62066]: ERROR nova.compute.manager [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] return func(*args, **kwargs) [ 711.673754] env[62066]: ERROR nova.compute.manager [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 711.673754] env[62066]: ERROR nova.compute.manager [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] raise e [ 711.673754] env[62066]: ERROR nova.compute.manager [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 711.673754] env[62066]: ERROR nova.compute.manager [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] nwinfo = self.network_api.allocate_for_instance( [ 711.673754] env[62066]: ERROR nova.compute.manager [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 711.673754] env[62066]: ERROR nova.compute.manager [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] created_port_ids = self._update_ports_for_instance( [ 711.673754] env[62066]: ERROR nova.compute.manager [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 711.673754] env[62066]: ERROR nova.compute.manager [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] with excutils.save_and_reraise_exception(): [ 711.673754] env[62066]: ERROR nova.compute.manager [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 711.674132] env[62066]: ERROR nova.compute.manager [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] self.force_reraise() [ 711.674132] env[62066]: ERROR nova.compute.manager [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 711.674132] env[62066]: ERROR nova.compute.manager [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] raise self.value [ 711.674132] env[62066]: ERROR nova.compute.manager [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 711.674132] env[62066]: ERROR nova.compute.manager [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] updated_port = self._update_port( [ 711.674132] env[62066]: ERROR nova.compute.manager [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 711.674132] env[62066]: ERROR nova.compute.manager [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] _ensure_no_port_binding_failure(port) [ 711.674132] env[62066]: ERROR nova.compute.manager [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 711.674132] env[62066]: ERROR nova.compute.manager [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] raise exception.PortBindingFailed(port_id=port['id']) [ 711.674132] env[62066]: ERROR nova.compute.manager [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] nova.exception.PortBindingFailed: Binding failed for port 37035015-85cf-4d55-9ba0-068bae5b3dab, please check neutron logs for more information. [ 711.674132] env[62066]: ERROR nova.compute.manager [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] [ 711.674538] env[62066]: DEBUG nova.compute.utils [None req-96f72479-6fed-4ff5-9d17-40fa20c1f416 tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] Binding failed for port 37035015-85cf-4d55-9ba0-068bae5b3dab, please check neutron logs for more information. {{(pid=62066) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 711.674976] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3d3acd2e-756a-442f-acf6-9a75ba7d933c tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.722s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 711.676494] env[62066]: INFO nova.compute.claims [None req-3d3acd2e-756a-442f-acf6-9a75ba7d933c tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] [instance: 0ff2f9f4-6d85-4465-8e0a-fb39f2a25c9b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 711.679081] env[62066]: DEBUG nova.compute.manager [None req-96f72479-6fed-4ff5-9d17-40fa20c1f416 tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] Build of instance 45c32181-01a8-4e1e-9e0e-37035a298d55 was re-scheduled: Binding failed for port 37035015-85cf-4d55-9ba0-068bae5b3dab, please check neutron logs for more information. {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 711.679528] env[62066]: DEBUG nova.compute.manager [None req-96f72479-6fed-4ff5-9d17-40fa20c1f416 tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] Unplugging VIFs for instance {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 711.679751] env[62066]: DEBUG oslo_concurrency.lockutils [None req-96f72479-6fed-4ff5-9d17-40fa20c1f416 tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] Acquiring lock "refresh_cache-45c32181-01a8-4e1e-9e0e-37035a298d55" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 711.680052] env[62066]: DEBUG oslo_concurrency.lockutils [None req-96f72479-6fed-4ff5-9d17-40fa20c1f416 tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] Acquired lock "refresh_cache-45c32181-01a8-4e1e-9e0e-37035a298d55" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 711.680052] env[62066]: DEBUG nova.network.neutron [None req-96f72479-6fed-4ff5-9d17-40fa20c1f416 tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 711.689654] env[62066]: DEBUG nova.compute.manager [None req-8d368cc7-6ac0-44cd-b71e-9136d8337cb4 tempest-AttachInterfacesUnderV243Test-1832696447 tempest-AttachInterfacesUnderV243Test-1832696447-project-member] [instance: f91a90f7-be73-424c-966c-1be6f37a0864] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 711.721632] env[62066]: DEBUG nova.virt.hardware [None req-8d368cc7-6ac0-44cd-b71e-9136d8337cb4 tempest-AttachInterfacesUnderV243Test-1832696447 tempest-AttachInterfacesUnderV243Test-1832696447-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-23T13:40:41Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-23T13:40:23Z,direct_url=,disk_format='vmdk',id=50ff584c-3b50-4395-af07-3e66769bc9f7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='eb52f7069a374c61ae946f052007c6d9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-23T13:40:24Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 711.721779] env[62066]: DEBUG nova.virt.hardware [None req-8d368cc7-6ac0-44cd-b71e-9136d8337cb4 tempest-AttachInterfacesUnderV243Test-1832696447 tempest-AttachInterfacesUnderV243Test-1832696447-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 711.721891] env[62066]: DEBUG nova.virt.hardware [None req-8d368cc7-6ac0-44cd-b71e-9136d8337cb4 tempest-AttachInterfacesUnderV243Test-1832696447 tempest-AttachInterfacesUnderV243Test-1832696447-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 711.722709] env[62066]: DEBUG nova.virt.hardware [None req-8d368cc7-6ac0-44cd-b71e-9136d8337cb4 tempest-AttachInterfacesUnderV243Test-1832696447 tempest-AttachInterfacesUnderV243Test-1832696447-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 711.722709] env[62066]: DEBUG nova.virt.hardware [None req-8d368cc7-6ac0-44cd-b71e-9136d8337cb4 tempest-AttachInterfacesUnderV243Test-1832696447 tempest-AttachInterfacesUnderV243Test-1832696447-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 711.722709] env[62066]: DEBUG nova.virt.hardware [None req-8d368cc7-6ac0-44cd-b71e-9136d8337cb4 tempest-AttachInterfacesUnderV243Test-1832696447 tempest-AttachInterfacesUnderV243Test-1832696447-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 711.722871] env[62066]: DEBUG nova.virt.hardware [None req-8d368cc7-6ac0-44cd-b71e-9136d8337cb4 tempest-AttachInterfacesUnderV243Test-1832696447 tempest-AttachInterfacesUnderV243Test-1832696447-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 711.722992] env[62066]: DEBUG nova.virt.hardware [None req-8d368cc7-6ac0-44cd-b71e-9136d8337cb4 tempest-AttachInterfacesUnderV243Test-1832696447 tempest-AttachInterfacesUnderV243Test-1832696447-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 711.723193] env[62066]: DEBUG nova.virt.hardware [None req-8d368cc7-6ac0-44cd-b71e-9136d8337cb4 tempest-AttachInterfacesUnderV243Test-1832696447 tempest-AttachInterfacesUnderV243Test-1832696447-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 711.723490] env[62066]: DEBUG nova.virt.hardware [None req-8d368cc7-6ac0-44cd-b71e-9136d8337cb4 tempest-AttachInterfacesUnderV243Test-1832696447 tempest-AttachInterfacesUnderV243Test-1832696447-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 711.723490] env[62066]: DEBUG nova.virt.hardware [None req-8d368cc7-6ac0-44cd-b71e-9136d8337cb4 tempest-AttachInterfacesUnderV243Test-1832696447 tempest-AttachInterfacesUnderV243Test-1832696447-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 711.724356] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2dbff6d9-3b1c-455a-a875-a11e3969bce3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.733771] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eea5615c-66fe-49c7-8ab5-098c62be5c9a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.785890] env[62066]: DEBUG nova.compute.manager [req-9cef8423-b7b1-4a16-b944-18ccf242be97 req-5dd631d3-f88d-4b7e-8b49-dc8a9a350d92 service nova] [instance: f91a90f7-be73-424c-966c-1be6f37a0864] Received event network-changed-01c976c4-59c2-4855-942f-61915dbd9fe3 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 711.785890] env[62066]: DEBUG nova.compute.manager [req-9cef8423-b7b1-4a16-b944-18ccf242be97 req-5dd631d3-f88d-4b7e-8b49-dc8a9a350d92 service nova] [instance: f91a90f7-be73-424c-966c-1be6f37a0864] Refreshing instance network info cache due to event network-changed-01c976c4-59c2-4855-942f-61915dbd9fe3. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 711.786417] env[62066]: DEBUG oslo_concurrency.lockutils [req-9cef8423-b7b1-4a16-b944-18ccf242be97 req-5dd631d3-f88d-4b7e-8b49-dc8a9a350d92 service nova] Acquiring lock "refresh_cache-f91a90f7-be73-424c-966c-1be6f37a0864" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 711.786417] env[62066]: DEBUG oslo_concurrency.lockutils [req-9cef8423-b7b1-4a16-b944-18ccf242be97 req-5dd631d3-f88d-4b7e-8b49-dc8a9a350d92 service nova] Acquired lock "refresh_cache-f91a90f7-be73-424c-966c-1be6f37a0864" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 711.786532] env[62066]: DEBUG nova.network.neutron [req-9cef8423-b7b1-4a16-b944-18ccf242be97 req-5dd631d3-f88d-4b7e-8b49-dc8a9a350d92 service nova] [instance: f91a90f7-be73-424c-966c-1be6f37a0864] Refreshing network info cache for port 01c976c4-59c2-4855-942f-61915dbd9fe3 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 711.843817] env[62066]: DEBUG oslo_service.periodic_task [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 711.843997] env[62066]: DEBUG nova.compute.manager [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Starting heal instance info cache {{(pid=62066) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 711.844129] env[62066]: DEBUG nova.compute.manager [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Rebuilding the list of instances to heal {{(pid=62066) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 711.969809] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e9793203-98b3-4ca0-8a49-79767b1f52f7 tempest-ServersTestBootFromVolume-1535558002 tempest-ServersTestBootFromVolume-1535558002-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 712.066331] env[62066]: ERROR nova.compute.manager [None req-8d368cc7-6ac0-44cd-b71e-9136d8337cb4 tempest-AttachInterfacesUnderV243Test-1832696447 tempest-AttachInterfacesUnderV243Test-1832696447-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 01c976c4-59c2-4855-942f-61915dbd9fe3, please check neutron logs for more information. [ 712.066331] env[62066]: ERROR nova.compute.manager Traceback (most recent call last): [ 712.066331] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 712.066331] env[62066]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 712.066331] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 712.066331] env[62066]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 712.066331] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 712.066331] env[62066]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 712.066331] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 712.066331] env[62066]: ERROR nova.compute.manager self.force_reraise() [ 712.066331] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 712.066331] env[62066]: ERROR nova.compute.manager raise self.value [ 712.066331] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 712.066331] env[62066]: ERROR nova.compute.manager updated_port = self._update_port( [ 712.066331] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 712.066331] env[62066]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 712.066950] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 712.066950] env[62066]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 712.066950] env[62066]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 01c976c4-59c2-4855-942f-61915dbd9fe3, please check neutron logs for more information. [ 712.066950] env[62066]: ERROR nova.compute.manager [ 712.066950] env[62066]: Traceback (most recent call last): [ 712.066950] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 712.066950] env[62066]: listener.cb(fileno) [ 712.066950] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 712.066950] env[62066]: result = function(*args, **kwargs) [ 712.066950] env[62066]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 712.066950] env[62066]: return func(*args, **kwargs) [ 712.066950] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 712.066950] env[62066]: raise e [ 712.066950] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 712.066950] env[62066]: nwinfo = self.network_api.allocate_for_instance( [ 712.066950] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 712.066950] env[62066]: created_port_ids = self._update_ports_for_instance( [ 712.066950] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 712.066950] env[62066]: with excutils.save_and_reraise_exception(): [ 712.066950] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 712.066950] env[62066]: self.force_reraise() [ 712.066950] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 712.066950] env[62066]: raise self.value [ 712.066950] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 712.066950] env[62066]: updated_port = self._update_port( [ 712.066950] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 712.066950] env[62066]: _ensure_no_port_binding_failure(port) [ 712.066950] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 712.066950] env[62066]: raise exception.PortBindingFailed(port_id=port['id']) [ 712.067840] env[62066]: nova.exception.PortBindingFailed: Binding failed for port 01c976c4-59c2-4855-942f-61915dbd9fe3, please check neutron logs for more information. [ 712.067840] env[62066]: Removing descriptor: 20 [ 712.067840] env[62066]: ERROR nova.compute.manager [None req-8d368cc7-6ac0-44cd-b71e-9136d8337cb4 tempest-AttachInterfacesUnderV243Test-1832696447 tempest-AttachInterfacesUnderV243Test-1832696447-project-member] [instance: f91a90f7-be73-424c-966c-1be6f37a0864] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 01c976c4-59c2-4855-942f-61915dbd9fe3, please check neutron logs for more information. [ 712.067840] env[62066]: ERROR nova.compute.manager [instance: f91a90f7-be73-424c-966c-1be6f37a0864] Traceback (most recent call last): [ 712.067840] env[62066]: ERROR nova.compute.manager [instance: f91a90f7-be73-424c-966c-1be6f37a0864] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 712.067840] env[62066]: ERROR nova.compute.manager [instance: f91a90f7-be73-424c-966c-1be6f37a0864] yield resources [ 712.067840] env[62066]: ERROR nova.compute.manager [instance: f91a90f7-be73-424c-966c-1be6f37a0864] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 712.067840] env[62066]: ERROR nova.compute.manager [instance: f91a90f7-be73-424c-966c-1be6f37a0864] self.driver.spawn(context, instance, image_meta, [ 712.067840] env[62066]: ERROR nova.compute.manager [instance: f91a90f7-be73-424c-966c-1be6f37a0864] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 712.067840] env[62066]: ERROR nova.compute.manager [instance: f91a90f7-be73-424c-966c-1be6f37a0864] self._vmops.spawn(context, instance, image_meta, injected_files, [ 712.067840] env[62066]: ERROR nova.compute.manager [instance: f91a90f7-be73-424c-966c-1be6f37a0864] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 712.067840] env[62066]: ERROR nova.compute.manager [instance: f91a90f7-be73-424c-966c-1be6f37a0864] vm_ref = self.build_virtual_machine(instance, [ 712.068221] env[62066]: ERROR nova.compute.manager [instance: f91a90f7-be73-424c-966c-1be6f37a0864] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 712.068221] env[62066]: ERROR nova.compute.manager [instance: f91a90f7-be73-424c-966c-1be6f37a0864] vif_infos = vmwarevif.get_vif_info(self._session, [ 712.068221] env[62066]: ERROR nova.compute.manager [instance: f91a90f7-be73-424c-966c-1be6f37a0864] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 712.068221] env[62066]: ERROR nova.compute.manager [instance: f91a90f7-be73-424c-966c-1be6f37a0864] for vif in network_info: [ 712.068221] env[62066]: ERROR nova.compute.manager [instance: f91a90f7-be73-424c-966c-1be6f37a0864] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 712.068221] env[62066]: ERROR nova.compute.manager [instance: f91a90f7-be73-424c-966c-1be6f37a0864] return self._sync_wrapper(fn, *args, **kwargs) [ 712.068221] env[62066]: ERROR nova.compute.manager [instance: f91a90f7-be73-424c-966c-1be6f37a0864] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 712.068221] env[62066]: ERROR nova.compute.manager [instance: f91a90f7-be73-424c-966c-1be6f37a0864] self.wait() [ 712.068221] env[62066]: ERROR nova.compute.manager [instance: f91a90f7-be73-424c-966c-1be6f37a0864] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 712.068221] env[62066]: ERROR nova.compute.manager [instance: f91a90f7-be73-424c-966c-1be6f37a0864] self[:] = self._gt.wait() [ 712.068221] env[62066]: ERROR nova.compute.manager [instance: f91a90f7-be73-424c-966c-1be6f37a0864] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 712.068221] env[62066]: ERROR nova.compute.manager [instance: f91a90f7-be73-424c-966c-1be6f37a0864] return self._exit_event.wait() [ 712.068221] env[62066]: ERROR nova.compute.manager [instance: f91a90f7-be73-424c-966c-1be6f37a0864] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 712.068635] env[62066]: ERROR nova.compute.manager [instance: f91a90f7-be73-424c-966c-1be6f37a0864] result = hub.switch() [ 712.068635] env[62066]: ERROR nova.compute.manager [instance: f91a90f7-be73-424c-966c-1be6f37a0864] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 712.068635] env[62066]: ERROR nova.compute.manager [instance: f91a90f7-be73-424c-966c-1be6f37a0864] return self.greenlet.switch() [ 712.068635] env[62066]: ERROR nova.compute.manager [instance: f91a90f7-be73-424c-966c-1be6f37a0864] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 712.068635] env[62066]: ERROR nova.compute.manager [instance: f91a90f7-be73-424c-966c-1be6f37a0864] result = function(*args, **kwargs) [ 712.068635] env[62066]: ERROR nova.compute.manager [instance: f91a90f7-be73-424c-966c-1be6f37a0864] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 712.068635] env[62066]: ERROR nova.compute.manager [instance: f91a90f7-be73-424c-966c-1be6f37a0864] return func(*args, **kwargs) [ 712.068635] env[62066]: ERROR nova.compute.manager [instance: f91a90f7-be73-424c-966c-1be6f37a0864] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 712.068635] env[62066]: ERROR nova.compute.manager [instance: f91a90f7-be73-424c-966c-1be6f37a0864] raise e [ 712.068635] env[62066]: ERROR nova.compute.manager [instance: f91a90f7-be73-424c-966c-1be6f37a0864] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 712.068635] env[62066]: ERROR nova.compute.manager [instance: f91a90f7-be73-424c-966c-1be6f37a0864] nwinfo = self.network_api.allocate_for_instance( [ 712.068635] env[62066]: ERROR nova.compute.manager [instance: f91a90f7-be73-424c-966c-1be6f37a0864] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 712.068635] env[62066]: ERROR nova.compute.manager [instance: f91a90f7-be73-424c-966c-1be6f37a0864] created_port_ids = self._update_ports_for_instance( [ 712.069042] env[62066]: ERROR nova.compute.manager [instance: f91a90f7-be73-424c-966c-1be6f37a0864] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 712.069042] env[62066]: ERROR nova.compute.manager [instance: f91a90f7-be73-424c-966c-1be6f37a0864] with excutils.save_and_reraise_exception(): [ 712.069042] env[62066]: ERROR nova.compute.manager [instance: f91a90f7-be73-424c-966c-1be6f37a0864] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 712.069042] env[62066]: ERROR nova.compute.manager [instance: f91a90f7-be73-424c-966c-1be6f37a0864] self.force_reraise() [ 712.069042] env[62066]: ERROR nova.compute.manager [instance: f91a90f7-be73-424c-966c-1be6f37a0864] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 712.069042] env[62066]: ERROR nova.compute.manager [instance: f91a90f7-be73-424c-966c-1be6f37a0864] raise self.value [ 712.069042] env[62066]: ERROR nova.compute.manager [instance: f91a90f7-be73-424c-966c-1be6f37a0864] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 712.069042] env[62066]: ERROR nova.compute.manager [instance: f91a90f7-be73-424c-966c-1be6f37a0864] updated_port = self._update_port( [ 712.069042] env[62066]: ERROR nova.compute.manager [instance: f91a90f7-be73-424c-966c-1be6f37a0864] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 712.069042] env[62066]: ERROR nova.compute.manager [instance: f91a90f7-be73-424c-966c-1be6f37a0864] _ensure_no_port_binding_failure(port) [ 712.069042] env[62066]: ERROR nova.compute.manager [instance: f91a90f7-be73-424c-966c-1be6f37a0864] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 712.069042] env[62066]: ERROR nova.compute.manager [instance: f91a90f7-be73-424c-966c-1be6f37a0864] raise exception.PortBindingFailed(port_id=port['id']) [ 712.069418] env[62066]: ERROR nova.compute.manager [instance: f91a90f7-be73-424c-966c-1be6f37a0864] nova.exception.PortBindingFailed: Binding failed for port 01c976c4-59c2-4855-942f-61915dbd9fe3, please check neutron logs for more information. [ 712.069418] env[62066]: ERROR nova.compute.manager [instance: f91a90f7-be73-424c-966c-1be6f37a0864] [ 712.069418] env[62066]: INFO nova.compute.manager [None req-8d368cc7-6ac0-44cd-b71e-9136d8337cb4 tempest-AttachInterfacesUnderV243Test-1832696447 tempest-AttachInterfacesUnderV243Test-1832696447-project-member] [instance: f91a90f7-be73-424c-966c-1be6f37a0864] Terminating instance [ 712.069937] env[62066]: DEBUG oslo_concurrency.lockutils [None req-8d368cc7-6ac0-44cd-b71e-9136d8337cb4 tempest-AttachInterfacesUnderV243Test-1832696447 tempest-AttachInterfacesUnderV243Test-1832696447-project-member] Acquiring lock "refresh_cache-f91a90f7-be73-424c-966c-1be6f37a0864" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 712.202471] env[62066]: DEBUG nova.network.neutron [None req-96f72479-6fed-4ff5-9d17-40fa20c1f416 tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 712.287397] env[62066]: DEBUG nova.network.neutron [None req-96f72479-6fed-4ff5-9d17-40fa20c1f416 tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 712.302769] env[62066]: DEBUG nova.network.neutron [req-9cef8423-b7b1-4a16-b944-18ccf242be97 req-5dd631d3-f88d-4b7e-8b49-dc8a9a350d92 service nova] [instance: f91a90f7-be73-424c-966c-1be6f37a0864] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 712.353403] env[62066]: DEBUG nova.compute.manager [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] Skipping network cache update for instance because it is Building. {{(pid=62066) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 712.353578] env[62066]: DEBUG nova.compute.manager [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] Skipping network cache update for instance because it is Building. {{(pid=62066) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 712.353706] env[62066]: DEBUG nova.compute.manager [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] [instance: f91a90f7-be73-424c-966c-1be6f37a0864] Skipping network cache update for instance because it is Building. {{(pid=62066) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 712.353820] env[62066]: DEBUG nova.compute.manager [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] [instance: 0ff2f9f4-6d85-4465-8e0a-fb39f2a25c9b] Skipping network cache update for instance because it is Building. {{(pid=62066) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 712.353953] env[62066]: DEBUG nova.compute.manager [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Didn't find any instances for network info cache update. {{(pid=62066) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 712.354159] env[62066]: DEBUG oslo_service.periodic_task [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 712.354314] env[62066]: DEBUG oslo_service.periodic_task [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 712.354459] env[62066]: DEBUG oslo_service.periodic_task [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 712.354607] env[62066]: DEBUG oslo_service.periodic_task [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 712.354793] env[62066]: DEBUG oslo_service.periodic_task [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 712.354975] env[62066]: DEBUG oslo_service.periodic_task [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 712.355118] env[62066]: DEBUG nova.compute.manager [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62066) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 712.355261] env[62066]: DEBUG oslo_service.periodic_task [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Running periodic task ComputeManager.update_available_resource {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 712.397498] env[62066]: DEBUG nova.network.neutron [req-9cef8423-b7b1-4a16-b944-18ccf242be97 req-5dd631d3-f88d-4b7e-8b49-dc8a9a350d92 service nova] [instance: f91a90f7-be73-424c-966c-1be6f37a0864] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 712.792216] env[62066]: DEBUG oslo_concurrency.lockutils [None req-96f72479-6fed-4ff5-9d17-40fa20c1f416 tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] Releasing lock "refresh_cache-45c32181-01a8-4e1e-9e0e-37035a298d55" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 712.792505] env[62066]: DEBUG nova.compute.manager [None req-96f72479-6fed-4ff5-9d17-40fa20c1f416 tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 712.792713] env[62066]: DEBUG nova.compute.manager [None req-96f72479-6fed-4ff5-9d17-40fa20c1f416 tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 712.792880] env[62066]: DEBUG nova.network.neutron [None req-96f72479-6fed-4ff5-9d17-40fa20c1f416 tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 712.825735] env[62066]: DEBUG nova.network.neutron [None req-96f72479-6fed-4ff5-9d17-40fa20c1f416 tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 712.858693] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 712.902143] env[62066]: DEBUG oslo_concurrency.lockutils [req-9cef8423-b7b1-4a16-b944-18ccf242be97 req-5dd631d3-f88d-4b7e-8b49-dc8a9a350d92 service nova] Releasing lock "refresh_cache-f91a90f7-be73-424c-966c-1be6f37a0864" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 712.903365] env[62066]: DEBUG oslo_concurrency.lockutils [None req-8d368cc7-6ac0-44cd-b71e-9136d8337cb4 tempest-AttachInterfacesUnderV243Test-1832696447 tempest-AttachInterfacesUnderV243Test-1832696447-project-member] Acquired lock "refresh_cache-f91a90f7-be73-424c-966c-1be6f37a0864" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 712.903365] env[62066]: DEBUG nova.network.neutron [None req-8d368cc7-6ac0-44cd-b71e-9136d8337cb4 tempest-AttachInterfacesUnderV243Test-1832696447 tempest-AttachInterfacesUnderV243Test-1832696447-project-member] [instance: f91a90f7-be73-424c-966c-1be6f37a0864] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 713.026965] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b60ebce-034d-4b21-8a54-2e172b1e8c55 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.035119] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b6f5899-d34f-4a97-b1a0-14f345f1dc95 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.065360] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cac62c72-9115-461d-97bd-49491036fd35 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.072735] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63a05dcb-3b2c-48d1-807d-97591a2222d8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.085547] env[62066]: DEBUG nova.compute.provider_tree [None req-3d3acd2e-756a-442f-acf6-9a75ba7d933c tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 713.328311] env[62066]: DEBUG nova.network.neutron [None req-96f72479-6fed-4ff5-9d17-40fa20c1f416 tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 713.421700] env[62066]: DEBUG nova.network.neutron [None req-8d368cc7-6ac0-44cd-b71e-9136d8337cb4 tempest-AttachInterfacesUnderV243Test-1832696447 tempest-AttachInterfacesUnderV243Test-1832696447-project-member] [instance: f91a90f7-be73-424c-966c-1be6f37a0864] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 713.540016] env[62066]: DEBUG nova.network.neutron [None req-8d368cc7-6ac0-44cd-b71e-9136d8337cb4 tempest-AttachInterfacesUnderV243Test-1832696447 tempest-AttachInterfacesUnderV243Test-1832696447-project-member] [instance: f91a90f7-be73-424c-966c-1be6f37a0864] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 713.589300] env[62066]: DEBUG nova.scheduler.client.report [None req-3d3acd2e-756a-442f-acf6-9a75ba7d933c tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 713.831635] env[62066]: INFO nova.compute.manager [None req-96f72479-6fed-4ff5-9d17-40fa20c1f416 tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] [instance: 45c32181-01a8-4e1e-9e0e-37035a298d55] Took 1.04 seconds to deallocate network for instance. [ 713.861337] env[62066]: DEBUG nova.compute.manager [req-b472e82c-134f-4320-9b14-100b8aab857c req-4bf6a506-fe35-4aa7-a70c-171775266207 service nova] [instance: f91a90f7-be73-424c-966c-1be6f37a0864] Received event network-vif-deleted-01c976c4-59c2-4855-942f-61915dbd9fe3 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 714.044067] env[62066]: DEBUG oslo_concurrency.lockutils [None req-8d368cc7-6ac0-44cd-b71e-9136d8337cb4 tempest-AttachInterfacesUnderV243Test-1832696447 tempest-AttachInterfacesUnderV243Test-1832696447-project-member] Releasing lock "refresh_cache-f91a90f7-be73-424c-966c-1be6f37a0864" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 714.044537] env[62066]: DEBUG nova.compute.manager [None req-8d368cc7-6ac0-44cd-b71e-9136d8337cb4 tempest-AttachInterfacesUnderV243Test-1832696447 tempest-AttachInterfacesUnderV243Test-1832696447-project-member] [instance: f91a90f7-be73-424c-966c-1be6f37a0864] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 714.044736] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-8d368cc7-6ac0-44cd-b71e-9136d8337cb4 tempest-AttachInterfacesUnderV243Test-1832696447 tempest-AttachInterfacesUnderV243Test-1832696447-project-member] [instance: f91a90f7-be73-424c-966c-1be6f37a0864] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 714.045064] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b591d83e-cd08-4f92-b86f-7845227a53bb {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.054879] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6ae7b03-423b-4380-ab7e-ca0e5f91f4ea {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.076143] env[62066]: WARNING nova.virt.vmwareapi.vmops [None req-8d368cc7-6ac0-44cd-b71e-9136d8337cb4 tempest-AttachInterfacesUnderV243Test-1832696447 tempest-AttachInterfacesUnderV243Test-1832696447-project-member] [instance: f91a90f7-be73-424c-966c-1be6f37a0864] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance f91a90f7-be73-424c-966c-1be6f37a0864 could not be found. [ 714.076491] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-8d368cc7-6ac0-44cd-b71e-9136d8337cb4 tempest-AttachInterfacesUnderV243Test-1832696447 tempest-AttachInterfacesUnderV243Test-1832696447-project-member] [instance: f91a90f7-be73-424c-966c-1be6f37a0864] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 714.076734] env[62066]: INFO nova.compute.manager [None req-8d368cc7-6ac0-44cd-b71e-9136d8337cb4 tempest-AttachInterfacesUnderV243Test-1832696447 tempest-AttachInterfacesUnderV243Test-1832696447-project-member] [instance: f91a90f7-be73-424c-966c-1be6f37a0864] Took 0.03 seconds to destroy the instance on the hypervisor. [ 714.076972] env[62066]: DEBUG oslo.service.loopingcall [None req-8d368cc7-6ac0-44cd-b71e-9136d8337cb4 tempest-AttachInterfacesUnderV243Test-1832696447 tempest-AttachInterfacesUnderV243Test-1832696447-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 714.077201] env[62066]: DEBUG nova.compute.manager [-] [instance: f91a90f7-be73-424c-966c-1be6f37a0864] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 714.077294] env[62066]: DEBUG nova.network.neutron [-] [instance: f91a90f7-be73-424c-966c-1be6f37a0864] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 714.092948] env[62066]: DEBUG nova.network.neutron [-] [instance: f91a90f7-be73-424c-966c-1be6f37a0864] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 714.094689] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3d3acd2e-756a-442f-acf6-9a75ba7d933c tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.420s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 714.095182] env[62066]: DEBUG nova.compute.manager [None req-3d3acd2e-756a-442f-acf6-9a75ba7d933c tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] [instance: 0ff2f9f4-6d85-4465-8e0a-fb39f2a25c9b] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 714.098540] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5b5d7867-70eb-472d-8d1a-c2bb035ec6db tempest-InstanceActionsV221TestJSON-1098427857 tempest-InstanceActionsV221TestJSON-1098427857-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.930s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 714.099942] env[62066]: INFO nova.compute.claims [None req-5b5d7867-70eb-472d-8d1a-c2bb035ec6db tempest-InstanceActionsV221TestJSON-1098427857 tempest-InstanceActionsV221TestJSON-1098427857-project-member] [instance: d7de90c0-95e9-4899-92a0-81658831713a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 714.599570] env[62066]: DEBUG nova.network.neutron [-] [instance: f91a90f7-be73-424c-966c-1be6f37a0864] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 714.604113] env[62066]: DEBUG nova.compute.utils [None req-3d3acd2e-756a-442f-acf6-9a75ba7d933c tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 714.607380] env[62066]: DEBUG nova.compute.manager [None req-3d3acd2e-756a-442f-acf6-9a75ba7d933c tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] [instance: 0ff2f9f4-6d85-4465-8e0a-fb39f2a25c9b] Not allocating networking since 'none' was specified. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 714.863273] env[62066]: INFO nova.scheduler.client.report [None req-96f72479-6fed-4ff5-9d17-40fa20c1f416 tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] Deleted allocations for instance 45c32181-01a8-4e1e-9e0e-37035a298d55 [ 715.103025] env[62066]: INFO nova.compute.manager [-] [instance: f91a90f7-be73-424c-966c-1be6f37a0864] Took 1.02 seconds to deallocate network for instance. [ 715.104452] env[62066]: DEBUG nova.compute.claims [None req-8d368cc7-6ac0-44cd-b71e-9136d8337cb4 tempest-AttachInterfacesUnderV243Test-1832696447 tempest-AttachInterfacesUnderV243Test-1832696447-project-member] [instance: f91a90f7-be73-424c-966c-1be6f37a0864] Aborting claim: {{(pid=62066) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 715.104632] env[62066]: DEBUG oslo_concurrency.lockutils [None req-8d368cc7-6ac0-44cd-b71e-9136d8337cb4 tempest-AttachInterfacesUnderV243Test-1832696447 tempest-AttachInterfacesUnderV243Test-1832696447-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 715.109277] env[62066]: DEBUG nova.compute.manager [None req-3d3acd2e-756a-442f-acf6-9a75ba7d933c tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] [instance: 0ff2f9f4-6d85-4465-8e0a-fb39f2a25c9b] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 715.372805] env[62066]: DEBUG oslo_concurrency.lockutils [None req-96f72479-6fed-4ff5-9d17-40fa20c1f416 tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] Lock "45c32181-01a8-4e1e-9e0e-37035a298d55" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 156.999s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 715.391497] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3dba141-f57f-4410-afa8-362068673091 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.400363] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f174b23-9a33-4e6a-a58f-0c6701c078f9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.433576] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d420c19-0691-491c-91b2-d630e25f32f2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.441064] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6317817-6f20-4bbd-b1de-60051d2857d7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.454113] env[62066]: DEBUG nova.compute.provider_tree [None req-5b5d7867-70eb-472d-8d1a-c2bb035ec6db tempest-InstanceActionsV221TestJSON-1098427857 tempest-InstanceActionsV221TestJSON-1098427857-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 715.875473] env[62066]: DEBUG nova.compute.manager [None req-478d4566-e74c-4f71-b94d-419dc0a31446 tempest-ServerPasswordTestJSON-849130511 tempest-ServerPasswordTestJSON-849130511-project-member] [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 715.960849] env[62066]: DEBUG nova.scheduler.client.report [None req-5b5d7867-70eb-472d-8d1a-c2bb035ec6db tempest-InstanceActionsV221TestJSON-1098427857 tempest-InstanceActionsV221TestJSON-1098427857-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 716.121860] env[62066]: DEBUG nova.compute.manager [None req-3d3acd2e-756a-442f-acf6-9a75ba7d933c tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] [instance: 0ff2f9f4-6d85-4465-8e0a-fb39f2a25c9b] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 716.157612] env[62066]: DEBUG nova.virt.hardware [None req-3d3acd2e-756a-442f-acf6-9a75ba7d933c tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-23T13:40:41Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-23T13:40:23Z,direct_url=,disk_format='vmdk',id=50ff584c-3b50-4395-af07-3e66769bc9f7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='eb52f7069a374c61ae946f052007c6d9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-23T13:40:24Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 716.157872] env[62066]: DEBUG nova.virt.hardware [None req-3d3acd2e-756a-442f-acf6-9a75ba7d933c tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 716.158043] env[62066]: DEBUG nova.virt.hardware [None req-3d3acd2e-756a-442f-acf6-9a75ba7d933c tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 716.161345] env[62066]: DEBUG nova.virt.hardware [None req-3d3acd2e-756a-442f-acf6-9a75ba7d933c tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 716.161345] env[62066]: DEBUG nova.virt.hardware [None req-3d3acd2e-756a-442f-acf6-9a75ba7d933c tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 716.161345] env[62066]: DEBUG nova.virt.hardware [None req-3d3acd2e-756a-442f-acf6-9a75ba7d933c tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 716.161345] env[62066]: DEBUG nova.virt.hardware [None req-3d3acd2e-756a-442f-acf6-9a75ba7d933c tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 716.161345] env[62066]: DEBUG nova.virt.hardware [None req-3d3acd2e-756a-442f-acf6-9a75ba7d933c tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 716.161345] env[62066]: DEBUG nova.virt.hardware [None req-3d3acd2e-756a-442f-acf6-9a75ba7d933c tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 716.161872] env[62066]: DEBUG nova.virt.hardware [None req-3d3acd2e-756a-442f-acf6-9a75ba7d933c tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 716.161872] env[62066]: DEBUG nova.virt.hardware [None req-3d3acd2e-756a-442f-acf6-9a75ba7d933c tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 716.161872] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae9811df-8b0b-4bce-81ac-9d620ca0af6c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.168684] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e06f40bc-2152-4d6b-b9eb-e75af42b450f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.185024] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-3d3acd2e-756a-442f-acf6-9a75ba7d933c tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] [instance: 0ff2f9f4-6d85-4465-8e0a-fb39f2a25c9b] Instance VIF info [] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 716.187856] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d3acd2e-756a-442f-acf6-9a75ba7d933c tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Creating folder: Project (87c481391d924a80aed4123708157411). Parent ref: group-v285980. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 716.189143] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ff82b394-6217-4277-9ae7-e79640a35851 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.200089] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-3d3acd2e-756a-442f-acf6-9a75ba7d933c tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Created folder: Project (87c481391d924a80aed4123708157411) in parent group-v285980. [ 716.200237] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d3acd2e-756a-442f-acf6-9a75ba7d933c tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Creating folder: Instances. Parent ref: group-v286000. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 716.200460] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-bca9d7ff-0469-41af-8d57-d369df93a2df {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.209541] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-3d3acd2e-756a-442f-acf6-9a75ba7d933c tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Created folder: Instances in parent group-v286000. [ 716.209781] env[62066]: DEBUG oslo.service.loopingcall [None req-3d3acd2e-756a-442f-acf6-9a75ba7d933c tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 716.210240] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0ff2f9f4-6d85-4465-8e0a-fb39f2a25c9b] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 716.210440] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e54c2d4d-5508-42a6-9d57-ad1339dc0678 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.228767] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 716.228767] env[62066]: value = "task-1340677" [ 716.228767] env[62066]: _type = "Task" [ 716.228767] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 716.236236] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1340677, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 716.407155] env[62066]: DEBUG oslo_concurrency.lockutils [None req-478d4566-e74c-4f71-b94d-419dc0a31446 tempest-ServerPasswordTestJSON-849130511 tempest-ServerPasswordTestJSON-849130511-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 716.467462] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5b5d7867-70eb-472d-8d1a-c2bb035ec6db tempest-InstanceActionsV221TestJSON-1098427857 tempest-InstanceActionsV221TestJSON-1098427857-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.369s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 716.468128] env[62066]: DEBUG nova.compute.manager [None req-5b5d7867-70eb-472d-8d1a-c2bb035ec6db tempest-InstanceActionsV221TestJSON-1098427857 tempest-InstanceActionsV221TestJSON-1098427857-project-member] [instance: d7de90c0-95e9-4899-92a0-81658831713a] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 716.471762] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b2b06eaa-4237-476b-9167-ed8be81ba3b9 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.472s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 716.743295] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1340677, 'name': CreateVM_Task, 'duration_secs': 0.29081} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 716.743295] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0ff2f9f4-6d85-4465-8e0a-fb39f2a25c9b] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 716.743295] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3d3acd2e-756a-442f-acf6-9a75ba7d933c tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 716.743295] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3d3acd2e-756a-442f-acf6-9a75ba7d933c tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Acquired lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 716.743295] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3d3acd2e-756a-442f-acf6-9a75ba7d933c tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 716.743600] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-60fcf636-d3ba-4205-b0ff-6e7bbe4b30cf {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.749956] env[62066]: DEBUG oslo_vmware.api [None req-3d3acd2e-756a-442f-acf6-9a75ba7d933c tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Waiting for the task: (returnval){ [ 716.749956] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]5286db20-2b2f-1b0a-c3dc-356a0841fcad" [ 716.749956] env[62066]: _type = "Task" [ 716.749956] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 716.761271] env[62066]: DEBUG oslo_vmware.api [None req-3d3acd2e-756a-442f-acf6-9a75ba7d933c tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]5286db20-2b2f-1b0a-c3dc-356a0841fcad, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 716.977129] env[62066]: DEBUG nova.compute.utils [None req-5b5d7867-70eb-472d-8d1a-c2bb035ec6db tempest-InstanceActionsV221TestJSON-1098427857 tempest-InstanceActionsV221TestJSON-1098427857-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 716.981490] env[62066]: DEBUG nova.compute.manager [None req-5b5d7867-70eb-472d-8d1a-c2bb035ec6db tempest-InstanceActionsV221TestJSON-1098427857 tempest-InstanceActionsV221TestJSON-1098427857-project-member] [instance: d7de90c0-95e9-4899-92a0-81658831713a] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 716.981610] env[62066]: DEBUG nova.network.neutron [None req-5b5d7867-70eb-472d-8d1a-c2bb035ec6db tempest-InstanceActionsV221TestJSON-1098427857 tempest-InstanceActionsV221TestJSON-1098427857-project-member] [instance: d7de90c0-95e9-4899-92a0-81658831713a] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 717.021868] env[62066]: DEBUG nova.policy [None req-5b5d7867-70eb-472d-8d1a-c2bb035ec6db tempest-InstanceActionsV221TestJSON-1098427857 tempest-InstanceActionsV221TestJSON-1098427857-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b8b56e7a70b643ad87673a1cfb3306a2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e9a28b1832aa4a16a82e6f22c94a1543', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 717.261182] env[62066]: DEBUG oslo_vmware.api [None req-3d3acd2e-756a-442f-acf6-9a75ba7d933c tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]5286db20-2b2f-1b0a-c3dc-356a0841fcad, 'name': SearchDatastore_Task, 'duration_secs': 0.014276} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 717.263176] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3d3acd2e-756a-442f-acf6-9a75ba7d933c tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Releasing lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 717.263716] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-3d3acd2e-756a-442f-acf6-9a75ba7d933c tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] [instance: 0ff2f9f4-6d85-4465-8e0a-fb39f2a25c9b] Processing image 50ff584c-3b50-4395-af07-3e66769bc9f7 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 717.263804] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3d3acd2e-756a-442f-acf6-9a75ba7d933c tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 717.263937] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3d3acd2e-756a-442f-acf6-9a75ba7d933c tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Acquired lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 717.264369] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-3d3acd2e-756a-442f-acf6-9a75ba7d933c tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 717.264644] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-495700aa-1d8b-477a-a387-4612fdefa899 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.272808] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-3d3acd2e-756a-442f-acf6-9a75ba7d933c tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 717.273075] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-3d3acd2e-756a-442f-acf6-9a75ba7d933c tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 717.276147] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-91e25a0d-8717-4d43-850e-a02d1a254eac {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.283537] env[62066]: DEBUG oslo_vmware.api [None req-3d3acd2e-756a-442f-acf6-9a75ba7d933c tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Waiting for the task: (returnval){ [ 717.283537] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]523a97bb-58d1-3816-51f5-1c1faa736f6f" [ 717.283537] env[62066]: _type = "Task" [ 717.283537] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 717.300566] env[62066]: DEBUG oslo_vmware.api [None req-3d3acd2e-756a-442f-acf6-9a75ba7d933c tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]523a97bb-58d1-3816-51f5-1c1faa736f6f, 'name': SearchDatastore_Task, 'duration_secs': 0.011251} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 717.301562] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1866e389-ef51-4125-9341-7bf849364f61 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.306380] env[62066]: DEBUG oslo_vmware.api [None req-3d3acd2e-756a-442f-acf6-9a75ba7d933c tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Waiting for the task: (returnval){ [ 717.306380] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]52d2016e-0360-cc9f-10f2-5e92662ce0e5" [ 717.306380] env[62066]: _type = "Task" [ 717.306380] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 717.318137] env[62066]: DEBUG oslo_vmware.api [None req-3d3acd2e-756a-442f-acf6-9a75ba7d933c tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52d2016e-0360-cc9f-10f2-5e92662ce0e5, 'name': SearchDatastore_Task} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 717.318137] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3d3acd2e-756a-442f-acf6-9a75ba7d933c tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Releasing lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 717.318137] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d3acd2e-756a-442f-acf6-9a75ba7d933c tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk to [datastore2] 0ff2f9f4-6d85-4465-8e0a-fb39f2a25c9b/0ff2f9f4-6d85-4465-8e0a-fb39f2a25c9b.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 717.318137] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7c821ce1-bcb5-4c5a-a022-f9be4f4f090c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.324194] env[62066]: DEBUG oslo_vmware.api [None req-3d3acd2e-756a-442f-acf6-9a75ba7d933c tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Waiting for the task: (returnval){ [ 717.324194] env[62066]: value = "task-1340678" [ 717.324194] env[62066]: _type = "Task" [ 717.324194] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 717.331792] env[62066]: DEBUG oslo_vmware.api [None req-3d3acd2e-756a-442f-acf6-9a75ba7d933c tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Task: {'id': task-1340678, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 717.337613] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b75991f-9b75-4342-adf4-0e9978f94a21 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.343947] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e5485af-ba30-45a5-99e7-66d1044dac0b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.375050] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5020edbc-49d9-4077-ac6e-a36bfc9df9cf {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.382487] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6fd9458-008a-4573-be8c-65bbfeaffdd0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.396454] env[62066]: DEBUG nova.compute.provider_tree [None req-b2b06eaa-4237-476b-9167-ed8be81ba3b9 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 717.484598] env[62066]: DEBUG nova.compute.manager [None req-5b5d7867-70eb-472d-8d1a-c2bb035ec6db tempest-InstanceActionsV221TestJSON-1098427857 tempest-InstanceActionsV221TestJSON-1098427857-project-member] [instance: d7de90c0-95e9-4899-92a0-81658831713a] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 717.510736] env[62066]: DEBUG nova.network.neutron [None req-5b5d7867-70eb-472d-8d1a-c2bb035ec6db tempest-InstanceActionsV221TestJSON-1098427857 tempest-InstanceActionsV221TestJSON-1098427857-project-member] [instance: d7de90c0-95e9-4899-92a0-81658831713a] Successfully created port: 389df780-4dc9-49a5-b7d6-c0cbbcd46c35 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 717.834860] env[62066]: DEBUG oslo_vmware.api [None req-3d3acd2e-756a-442f-acf6-9a75ba7d933c tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Task: {'id': task-1340678, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.479447} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 717.835102] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d3acd2e-756a-442f-acf6-9a75ba7d933c tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk to [datastore2] 0ff2f9f4-6d85-4465-8e0a-fb39f2a25c9b/0ff2f9f4-6d85-4465-8e0a-fb39f2a25c9b.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 717.836101] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-3d3acd2e-756a-442f-acf6-9a75ba7d933c tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] [instance: 0ff2f9f4-6d85-4465-8e0a-fb39f2a25c9b] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 717.836435] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-10948d69-7ff6-47db-8bc1-ddac43499e05 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.843859] env[62066]: DEBUG oslo_vmware.api [None req-3d3acd2e-756a-442f-acf6-9a75ba7d933c tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Waiting for the task: (returnval){ [ 717.843859] env[62066]: value = "task-1340679" [ 717.843859] env[62066]: _type = "Task" [ 717.843859] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 717.854242] env[62066]: DEBUG oslo_vmware.api [None req-3d3acd2e-756a-442f-acf6-9a75ba7d933c tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Task: {'id': task-1340679, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 717.900636] env[62066]: DEBUG nova.scheduler.client.report [None req-b2b06eaa-4237-476b-9167-ed8be81ba3b9 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 718.359248] env[62066]: DEBUG oslo_vmware.api [None req-3d3acd2e-756a-442f-acf6-9a75ba7d933c tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Task: {'id': task-1340679, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066172} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 718.359654] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-3d3acd2e-756a-442f-acf6-9a75ba7d933c tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] [instance: 0ff2f9f4-6d85-4465-8e0a-fb39f2a25c9b] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 718.360386] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13d51d7d-d94e-472a-888d-adb86bbae286 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.386189] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-3d3acd2e-756a-442f-acf6-9a75ba7d933c tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] [instance: 0ff2f9f4-6d85-4465-8e0a-fb39f2a25c9b] Reconfiguring VM instance instance-0000002d to attach disk [datastore2] 0ff2f9f4-6d85-4465-8e0a-fb39f2a25c9b/0ff2f9f4-6d85-4465-8e0a-fb39f2a25c9b.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 718.386539] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-367eb089-ed5c-4120-b6ab-07a3d31c2f4c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.408082] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b2b06eaa-4237-476b-9167-ed8be81ba3b9 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.936s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 718.408820] env[62066]: ERROR nova.compute.manager [None req-b2b06eaa-4237-476b-9167-ed8be81ba3b9 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port eb8ff994-60f5-461b-b984-ff082644aebc, please check neutron logs for more information. [ 718.408820] env[62066]: ERROR nova.compute.manager [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] Traceback (most recent call last): [ 718.408820] env[62066]: ERROR nova.compute.manager [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 718.408820] env[62066]: ERROR nova.compute.manager [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] self.driver.spawn(context, instance, image_meta, [ 718.408820] env[62066]: ERROR nova.compute.manager [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 718.408820] env[62066]: ERROR nova.compute.manager [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] self._vmops.spawn(context, instance, image_meta, injected_files, [ 718.408820] env[62066]: ERROR nova.compute.manager [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 718.408820] env[62066]: ERROR nova.compute.manager [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] vm_ref = self.build_virtual_machine(instance, [ 718.408820] env[62066]: ERROR nova.compute.manager [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 718.408820] env[62066]: ERROR nova.compute.manager [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] vif_infos = vmwarevif.get_vif_info(self._session, [ 718.408820] env[62066]: ERROR nova.compute.manager [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 718.409208] env[62066]: ERROR nova.compute.manager [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] for vif in network_info: [ 718.409208] env[62066]: ERROR nova.compute.manager [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 718.409208] env[62066]: ERROR nova.compute.manager [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] return self._sync_wrapper(fn, *args, **kwargs) [ 718.409208] env[62066]: ERROR nova.compute.manager [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 718.409208] env[62066]: ERROR nova.compute.manager [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] self.wait() [ 718.409208] env[62066]: ERROR nova.compute.manager [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 718.409208] env[62066]: ERROR nova.compute.manager [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] self[:] = self._gt.wait() [ 718.409208] env[62066]: ERROR nova.compute.manager [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 718.409208] env[62066]: ERROR nova.compute.manager [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] return self._exit_event.wait() [ 718.409208] env[62066]: ERROR nova.compute.manager [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 718.409208] env[62066]: ERROR nova.compute.manager [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] result = hub.switch() [ 718.409208] env[62066]: ERROR nova.compute.manager [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 718.409208] env[62066]: ERROR nova.compute.manager [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] return self.greenlet.switch() [ 718.409596] env[62066]: ERROR nova.compute.manager [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 718.409596] env[62066]: ERROR nova.compute.manager [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] result = function(*args, **kwargs) [ 718.409596] env[62066]: ERROR nova.compute.manager [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 718.409596] env[62066]: ERROR nova.compute.manager [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] return func(*args, **kwargs) [ 718.409596] env[62066]: ERROR nova.compute.manager [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 718.409596] env[62066]: ERROR nova.compute.manager [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] raise e [ 718.409596] env[62066]: ERROR nova.compute.manager [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 718.409596] env[62066]: ERROR nova.compute.manager [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] nwinfo = self.network_api.allocate_for_instance( [ 718.409596] env[62066]: ERROR nova.compute.manager [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 718.409596] env[62066]: ERROR nova.compute.manager [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] created_port_ids = self._update_ports_for_instance( [ 718.409596] env[62066]: ERROR nova.compute.manager [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 718.409596] env[62066]: ERROR nova.compute.manager [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] with excutils.save_and_reraise_exception(): [ 718.409596] env[62066]: ERROR nova.compute.manager [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 718.409959] env[62066]: ERROR nova.compute.manager [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] self.force_reraise() [ 718.409959] env[62066]: ERROR nova.compute.manager [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 718.409959] env[62066]: ERROR nova.compute.manager [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] raise self.value [ 718.409959] env[62066]: ERROR nova.compute.manager [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 718.409959] env[62066]: ERROR nova.compute.manager [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] updated_port = self._update_port( [ 718.409959] env[62066]: ERROR nova.compute.manager [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 718.409959] env[62066]: ERROR nova.compute.manager [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] _ensure_no_port_binding_failure(port) [ 718.409959] env[62066]: ERROR nova.compute.manager [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 718.409959] env[62066]: ERROR nova.compute.manager [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] raise exception.PortBindingFailed(port_id=port['id']) [ 718.409959] env[62066]: ERROR nova.compute.manager [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] nova.exception.PortBindingFailed: Binding failed for port eb8ff994-60f5-461b-b984-ff082644aebc, please check neutron logs for more information. [ 718.409959] env[62066]: ERROR nova.compute.manager [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] [ 718.410293] env[62066]: DEBUG nova.compute.utils [None req-b2b06eaa-4237-476b-9167-ed8be81ba3b9 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] Binding failed for port eb8ff994-60f5-461b-b984-ff082644aebc, please check neutron logs for more information. {{(pid=62066) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 718.413400] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5bf8bcd1-f041-4648-9b3c-4febff568fa9 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.796s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 718.415326] env[62066]: INFO nova.compute.claims [None req-5bf8bcd1-f041-4648-9b3c-4febff568fa9 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] [instance: 3624ca43-9ee3-418a-9205-debb699295df] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 718.416623] env[62066]: DEBUG oslo_vmware.api [None req-3d3acd2e-756a-442f-acf6-9a75ba7d933c tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Waiting for the task: (returnval){ [ 718.416623] env[62066]: value = "task-1340680" [ 718.416623] env[62066]: _type = "Task" [ 718.416623] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 718.417069] env[62066]: DEBUG nova.compute.manager [None req-b2b06eaa-4237-476b-9167-ed8be81ba3b9 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] Build of instance f9a8e863-6e7f-4f00-b54a-78802659cd59 was re-scheduled: Binding failed for port eb8ff994-60f5-461b-b984-ff082644aebc, please check neutron logs for more information. {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 718.417465] env[62066]: DEBUG nova.compute.manager [None req-b2b06eaa-4237-476b-9167-ed8be81ba3b9 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] Unplugging VIFs for instance {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 718.417690] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b2b06eaa-4237-476b-9167-ed8be81ba3b9 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Acquiring lock "refresh_cache-f9a8e863-6e7f-4f00-b54a-78802659cd59" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 718.417942] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b2b06eaa-4237-476b-9167-ed8be81ba3b9 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Acquired lock "refresh_cache-f9a8e863-6e7f-4f00-b54a-78802659cd59" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 718.418031] env[62066]: DEBUG nova.network.neutron [None req-b2b06eaa-4237-476b-9167-ed8be81ba3b9 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 718.428387] env[62066]: DEBUG oslo_vmware.api [None req-3d3acd2e-756a-442f-acf6-9a75ba7d933c tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Task: {'id': task-1340680, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 718.500534] env[62066]: DEBUG oslo_concurrency.lockutils [None req-90f5cbab-7c0d-4b69-92aa-09bf24d9f7da tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] Acquiring lock "050040cb-6ee3-4ad2-960a-fcebb53ac394" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 718.500818] env[62066]: DEBUG oslo_concurrency.lockutils [None req-90f5cbab-7c0d-4b69-92aa-09bf24d9f7da tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] Lock "050040cb-6ee3-4ad2-960a-fcebb53ac394" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 718.505120] env[62066]: DEBUG nova.compute.manager [None req-5b5d7867-70eb-472d-8d1a-c2bb035ec6db tempest-InstanceActionsV221TestJSON-1098427857 tempest-InstanceActionsV221TestJSON-1098427857-project-member] [instance: d7de90c0-95e9-4899-92a0-81658831713a] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 718.540490] env[62066]: DEBUG nova.virt.hardware [None req-5b5d7867-70eb-472d-8d1a-c2bb035ec6db tempest-InstanceActionsV221TestJSON-1098427857 tempest-InstanceActionsV221TestJSON-1098427857-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-23T13:40:41Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-23T13:40:23Z,direct_url=,disk_format='vmdk',id=50ff584c-3b50-4395-af07-3e66769bc9f7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='eb52f7069a374c61ae946f052007c6d9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-23T13:40:24Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 718.541492] env[62066]: DEBUG nova.virt.hardware [None req-5b5d7867-70eb-472d-8d1a-c2bb035ec6db tempest-InstanceActionsV221TestJSON-1098427857 tempest-InstanceActionsV221TestJSON-1098427857-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 718.541492] env[62066]: DEBUG nova.virt.hardware [None req-5b5d7867-70eb-472d-8d1a-c2bb035ec6db tempest-InstanceActionsV221TestJSON-1098427857 tempest-InstanceActionsV221TestJSON-1098427857-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 718.541492] env[62066]: DEBUG nova.virt.hardware [None req-5b5d7867-70eb-472d-8d1a-c2bb035ec6db tempest-InstanceActionsV221TestJSON-1098427857 tempest-InstanceActionsV221TestJSON-1098427857-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 718.541492] env[62066]: DEBUG nova.virt.hardware [None req-5b5d7867-70eb-472d-8d1a-c2bb035ec6db tempest-InstanceActionsV221TestJSON-1098427857 tempest-InstanceActionsV221TestJSON-1098427857-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 718.541492] env[62066]: DEBUG nova.virt.hardware [None req-5b5d7867-70eb-472d-8d1a-c2bb035ec6db tempest-InstanceActionsV221TestJSON-1098427857 tempest-InstanceActionsV221TestJSON-1098427857-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 718.541684] env[62066]: DEBUG nova.virt.hardware [None req-5b5d7867-70eb-472d-8d1a-c2bb035ec6db tempest-InstanceActionsV221TestJSON-1098427857 tempest-InstanceActionsV221TestJSON-1098427857-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 718.542335] env[62066]: DEBUG nova.virt.hardware [None req-5b5d7867-70eb-472d-8d1a-c2bb035ec6db tempest-InstanceActionsV221TestJSON-1098427857 tempest-InstanceActionsV221TestJSON-1098427857-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 718.542335] env[62066]: DEBUG nova.virt.hardware [None req-5b5d7867-70eb-472d-8d1a-c2bb035ec6db tempest-InstanceActionsV221TestJSON-1098427857 tempest-InstanceActionsV221TestJSON-1098427857-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 718.542335] env[62066]: DEBUG nova.virt.hardware [None req-5b5d7867-70eb-472d-8d1a-c2bb035ec6db tempest-InstanceActionsV221TestJSON-1098427857 tempest-InstanceActionsV221TestJSON-1098427857-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 718.542335] env[62066]: DEBUG nova.virt.hardware [None req-5b5d7867-70eb-472d-8d1a-c2bb035ec6db tempest-InstanceActionsV221TestJSON-1098427857 tempest-InstanceActionsV221TestJSON-1098427857-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 718.543175] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18a10df4-d724-487c-8aa2-6012139608ed {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.551489] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8ededca-5006-4751-8164-a1759c1e6d22 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.645457] env[62066]: DEBUG nova.compute.manager [req-4efd7092-1575-4374-a39b-9b6520eae0c0 req-6d0283ea-5aac-45f6-9286-69693b82aa10 service nova] [instance: d7de90c0-95e9-4899-92a0-81658831713a] Received event network-changed-389df780-4dc9-49a5-b7d6-c0cbbcd46c35 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 718.645773] env[62066]: DEBUG nova.compute.manager [req-4efd7092-1575-4374-a39b-9b6520eae0c0 req-6d0283ea-5aac-45f6-9286-69693b82aa10 service nova] [instance: d7de90c0-95e9-4899-92a0-81658831713a] Refreshing instance network info cache due to event network-changed-389df780-4dc9-49a5-b7d6-c0cbbcd46c35. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 718.646065] env[62066]: DEBUG oslo_concurrency.lockutils [req-4efd7092-1575-4374-a39b-9b6520eae0c0 req-6d0283ea-5aac-45f6-9286-69693b82aa10 service nova] Acquiring lock "refresh_cache-d7de90c0-95e9-4899-92a0-81658831713a" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 718.646217] env[62066]: DEBUG oslo_concurrency.lockutils [req-4efd7092-1575-4374-a39b-9b6520eae0c0 req-6d0283ea-5aac-45f6-9286-69693b82aa10 service nova] Acquired lock "refresh_cache-d7de90c0-95e9-4899-92a0-81658831713a" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 718.646378] env[62066]: DEBUG nova.network.neutron [req-4efd7092-1575-4374-a39b-9b6520eae0c0 req-6d0283ea-5aac-45f6-9286-69693b82aa10 service nova] [instance: d7de90c0-95e9-4899-92a0-81658831713a] Refreshing network info cache for port 389df780-4dc9-49a5-b7d6-c0cbbcd46c35 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 718.934665] env[62066]: ERROR nova.compute.manager [None req-5b5d7867-70eb-472d-8d1a-c2bb035ec6db tempest-InstanceActionsV221TestJSON-1098427857 tempest-InstanceActionsV221TestJSON-1098427857-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 389df780-4dc9-49a5-b7d6-c0cbbcd46c35, please check neutron logs for more information. [ 718.934665] env[62066]: ERROR nova.compute.manager Traceback (most recent call last): [ 718.934665] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 718.934665] env[62066]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 718.934665] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 718.934665] env[62066]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 718.934665] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 718.934665] env[62066]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 718.934665] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 718.934665] env[62066]: ERROR nova.compute.manager self.force_reraise() [ 718.934665] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 718.934665] env[62066]: ERROR nova.compute.manager raise self.value [ 718.934665] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 718.934665] env[62066]: ERROR nova.compute.manager updated_port = self._update_port( [ 718.934665] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 718.934665] env[62066]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 718.935161] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 718.935161] env[62066]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 718.935161] env[62066]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 389df780-4dc9-49a5-b7d6-c0cbbcd46c35, please check neutron logs for more information. [ 718.935161] env[62066]: ERROR nova.compute.manager [ 718.935161] env[62066]: Traceback (most recent call last): [ 718.935161] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 718.935161] env[62066]: listener.cb(fileno) [ 718.935161] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 718.935161] env[62066]: result = function(*args, **kwargs) [ 718.935161] env[62066]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 718.935161] env[62066]: return func(*args, **kwargs) [ 718.935161] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 718.935161] env[62066]: raise e [ 718.935161] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 718.935161] env[62066]: nwinfo = self.network_api.allocate_for_instance( [ 718.935161] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 718.935161] env[62066]: created_port_ids = self._update_ports_for_instance( [ 718.935161] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 718.935161] env[62066]: with excutils.save_and_reraise_exception(): [ 718.935161] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 718.935161] env[62066]: self.force_reraise() [ 718.935161] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 718.935161] env[62066]: raise self.value [ 718.935161] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 718.935161] env[62066]: updated_port = self._update_port( [ 718.935161] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 718.935161] env[62066]: _ensure_no_port_binding_failure(port) [ 718.935161] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 718.935161] env[62066]: raise exception.PortBindingFailed(port_id=port['id']) [ 718.935945] env[62066]: nova.exception.PortBindingFailed: Binding failed for port 389df780-4dc9-49a5-b7d6-c0cbbcd46c35, please check neutron logs for more information. [ 718.935945] env[62066]: Removing descriptor: 20 [ 718.935945] env[62066]: ERROR nova.compute.manager [None req-5b5d7867-70eb-472d-8d1a-c2bb035ec6db tempest-InstanceActionsV221TestJSON-1098427857 tempest-InstanceActionsV221TestJSON-1098427857-project-member] [instance: d7de90c0-95e9-4899-92a0-81658831713a] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 389df780-4dc9-49a5-b7d6-c0cbbcd46c35, please check neutron logs for more information. [ 718.935945] env[62066]: ERROR nova.compute.manager [instance: d7de90c0-95e9-4899-92a0-81658831713a] Traceback (most recent call last): [ 718.935945] env[62066]: ERROR nova.compute.manager [instance: d7de90c0-95e9-4899-92a0-81658831713a] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 718.935945] env[62066]: ERROR nova.compute.manager [instance: d7de90c0-95e9-4899-92a0-81658831713a] yield resources [ 718.935945] env[62066]: ERROR nova.compute.manager [instance: d7de90c0-95e9-4899-92a0-81658831713a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 718.935945] env[62066]: ERROR nova.compute.manager [instance: d7de90c0-95e9-4899-92a0-81658831713a] self.driver.spawn(context, instance, image_meta, [ 718.935945] env[62066]: ERROR nova.compute.manager [instance: d7de90c0-95e9-4899-92a0-81658831713a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 718.935945] env[62066]: ERROR nova.compute.manager [instance: d7de90c0-95e9-4899-92a0-81658831713a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 718.935945] env[62066]: ERROR nova.compute.manager [instance: d7de90c0-95e9-4899-92a0-81658831713a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 718.935945] env[62066]: ERROR nova.compute.manager [instance: d7de90c0-95e9-4899-92a0-81658831713a] vm_ref = self.build_virtual_machine(instance, [ 718.936309] env[62066]: ERROR nova.compute.manager [instance: d7de90c0-95e9-4899-92a0-81658831713a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 718.936309] env[62066]: ERROR nova.compute.manager [instance: d7de90c0-95e9-4899-92a0-81658831713a] vif_infos = vmwarevif.get_vif_info(self._session, [ 718.936309] env[62066]: ERROR nova.compute.manager [instance: d7de90c0-95e9-4899-92a0-81658831713a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 718.936309] env[62066]: ERROR nova.compute.manager [instance: d7de90c0-95e9-4899-92a0-81658831713a] for vif in network_info: [ 718.936309] env[62066]: ERROR nova.compute.manager [instance: d7de90c0-95e9-4899-92a0-81658831713a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 718.936309] env[62066]: ERROR nova.compute.manager [instance: d7de90c0-95e9-4899-92a0-81658831713a] return self._sync_wrapper(fn, *args, **kwargs) [ 718.936309] env[62066]: ERROR nova.compute.manager [instance: d7de90c0-95e9-4899-92a0-81658831713a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 718.936309] env[62066]: ERROR nova.compute.manager [instance: d7de90c0-95e9-4899-92a0-81658831713a] self.wait() [ 718.936309] env[62066]: ERROR nova.compute.manager [instance: d7de90c0-95e9-4899-92a0-81658831713a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 718.936309] env[62066]: ERROR nova.compute.manager [instance: d7de90c0-95e9-4899-92a0-81658831713a] self[:] = self._gt.wait() [ 718.936309] env[62066]: ERROR nova.compute.manager [instance: d7de90c0-95e9-4899-92a0-81658831713a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 718.936309] env[62066]: ERROR nova.compute.manager [instance: d7de90c0-95e9-4899-92a0-81658831713a] return self._exit_event.wait() [ 718.936309] env[62066]: ERROR nova.compute.manager [instance: d7de90c0-95e9-4899-92a0-81658831713a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 718.936688] env[62066]: ERROR nova.compute.manager [instance: d7de90c0-95e9-4899-92a0-81658831713a] result = hub.switch() [ 718.936688] env[62066]: ERROR nova.compute.manager [instance: d7de90c0-95e9-4899-92a0-81658831713a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 718.936688] env[62066]: ERROR nova.compute.manager [instance: d7de90c0-95e9-4899-92a0-81658831713a] return self.greenlet.switch() [ 718.936688] env[62066]: ERROR nova.compute.manager [instance: d7de90c0-95e9-4899-92a0-81658831713a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 718.936688] env[62066]: ERROR nova.compute.manager [instance: d7de90c0-95e9-4899-92a0-81658831713a] result = function(*args, **kwargs) [ 718.936688] env[62066]: ERROR nova.compute.manager [instance: d7de90c0-95e9-4899-92a0-81658831713a] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 718.936688] env[62066]: ERROR nova.compute.manager [instance: d7de90c0-95e9-4899-92a0-81658831713a] return func(*args, **kwargs) [ 718.936688] env[62066]: ERROR nova.compute.manager [instance: d7de90c0-95e9-4899-92a0-81658831713a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 718.936688] env[62066]: ERROR nova.compute.manager [instance: d7de90c0-95e9-4899-92a0-81658831713a] raise e [ 718.936688] env[62066]: ERROR nova.compute.manager [instance: d7de90c0-95e9-4899-92a0-81658831713a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 718.936688] env[62066]: ERROR nova.compute.manager [instance: d7de90c0-95e9-4899-92a0-81658831713a] nwinfo = self.network_api.allocate_for_instance( [ 718.936688] env[62066]: ERROR nova.compute.manager [instance: d7de90c0-95e9-4899-92a0-81658831713a] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 718.936688] env[62066]: ERROR nova.compute.manager [instance: d7de90c0-95e9-4899-92a0-81658831713a] created_port_ids = self._update_ports_for_instance( [ 718.937086] env[62066]: ERROR nova.compute.manager [instance: d7de90c0-95e9-4899-92a0-81658831713a] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 718.937086] env[62066]: ERROR nova.compute.manager [instance: d7de90c0-95e9-4899-92a0-81658831713a] with excutils.save_and_reraise_exception(): [ 718.937086] env[62066]: ERROR nova.compute.manager [instance: d7de90c0-95e9-4899-92a0-81658831713a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 718.937086] env[62066]: ERROR nova.compute.manager [instance: d7de90c0-95e9-4899-92a0-81658831713a] self.force_reraise() [ 718.937086] env[62066]: ERROR nova.compute.manager [instance: d7de90c0-95e9-4899-92a0-81658831713a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 718.937086] env[62066]: ERROR nova.compute.manager [instance: d7de90c0-95e9-4899-92a0-81658831713a] raise self.value [ 718.937086] env[62066]: ERROR nova.compute.manager [instance: d7de90c0-95e9-4899-92a0-81658831713a] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 718.937086] env[62066]: ERROR nova.compute.manager [instance: d7de90c0-95e9-4899-92a0-81658831713a] updated_port = self._update_port( [ 718.937086] env[62066]: ERROR nova.compute.manager [instance: d7de90c0-95e9-4899-92a0-81658831713a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 718.937086] env[62066]: ERROR nova.compute.manager [instance: d7de90c0-95e9-4899-92a0-81658831713a] _ensure_no_port_binding_failure(port) [ 718.937086] env[62066]: ERROR nova.compute.manager [instance: d7de90c0-95e9-4899-92a0-81658831713a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 718.937086] env[62066]: ERROR nova.compute.manager [instance: d7de90c0-95e9-4899-92a0-81658831713a] raise exception.PortBindingFailed(port_id=port['id']) [ 718.937461] env[62066]: ERROR nova.compute.manager [instance: d7de90c0-95e9-4899-92a0-81658831713a] nova.exception.PortBindingFailed: Binding failed for port 389df780-4dc9-49a5-b7d6-c0cbbcd46c35, please check neutron logs for more information. [ 718.937461] env[62066]: ERROR nova.compute.manager [instance: d7de90c0-95e9-4899-92a0-81658831713a] [ 718.937461] env[62066]: INFO nova.compute.manager [None req-5b5d7867-70eb-472d-8d1a-c2bb035ec6db tempest-InstanceActionsV221TestJSON-1098427857 tempest-InstanceActionsV221TestJSON-1098427857-project-member] [instance: d7de90c0-95e9-4899-92a0-81658831713a] Terminating instance [ 718.940286] env[62066]: DEBUG oslo_vmware.api [None req-3d3acd2e-756a-442f-acf6-9a75ba7d933c tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Task: {'id': task-1340680, 'name': ReconfigVM_Task, 'duration_secs': 0.285356} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 718.941278] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5b5d7867-70eb-472d-8d1a-c2bb035ec6db tempest-InstanceActionsV221TestJSON-1098427857 tempest-InstanceActionsV221TestJSON-1098427857-project-member] Acquiring lock "refresh_cache-d7de90c0-95e9-4899-92a0-81658831713a" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 718.941517] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-3d3acd2e-756a-442f-acf6-9a75ba7d933c tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] [instance: 0ff2f9f4-6d85-4465-8e0a-fb39f2a25c9b] Reconfigured VM instance instance-0000002d to attach disk [datastore2] 0ff2f9f4-6d85-4465-8e0a-fb39f2a25c9b/0ff2f9f4-6d85-4465-8e0a-fb39f2a25c9b.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 718.942321] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4a150130-1d76-4e70-9033-affea1612602 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.949227] env[62066]: DEBUG oslo_vmware.api [None req-3d3acd2e-756a-442f-acf6-9a75ba7d933c tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Waiting for the task: (returnval){ [ 718.949227] env[62066]: value = "task-1340681" [ 718.949227] env[62066]: _type = "Task" [ 718.949227] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 718.949944] env[62066]: DEBUG nova.network.neutron [None req-b2b06eaa-4237-476b-9167-ed8be81ba3b9 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 718.960535] env[62066]: DEBUG oslo_vmware.api [None req-3d3acd2e-756a-442f-acf6-9a75ba7d933c tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Task: {'id': task-1340681, 'name': Rename_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 719.027857] env[62066]: DEBUG nova.network.neutron [None req-b2b06eaa-4237-476b-9167-ed8be81ba3b9 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 719.173143] env[62066]: DEBUG nova.network.neutron [req-4efd7092-1575-4374-a39b-9b6520eae0c0 req-6d0283ea-5aac-45f6-9286-69693b82aa10 service nova] [instance: d7de90c0-95e9-4899-92a0-81658831713a] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 719.252630] env[62066]: DEBUG nova.network.neutron [req-4efd7092-1575-4374-a39b-9b6520eae0c0 req-6d0283ea-5aac-45f6-9286-69693b82aa10 service nova] [instance: d7de90c0-95e9-4899-92a0-81658831713a] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 719.466012] env[62066]: DEBUG oslo_vmware.api [None req-3d3acd2e-756a-442f-acf6-9a75ba7d933c tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Task: {'id': task-1340681, 'name': Rename_Task, 'duration_secs': 0.137589} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 719.466012] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d3acd2e-756a-442f-acf6-9a75ba7d933c tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] [instance: 0ff2f9f4-6d85-4465-8e0a-fb39f2a25c9b] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 719.466012] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4e85e279-b30a-454d-9593-cd9e8ba9d9f3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.472148] env[62066]: DEBUG oslo_vmware.api [None req-3d3acd2e-756a-442f-acf6-9a75ba7d933c tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Waiting for the task: (returnval){ [ 719.472148] env[62066]: value = "task-1340682" [ 719.472148] env[62066]: _type = "Task" [ 719.472148] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 719.485124] env[62066]: DEBUG oslo_vmware.api [None req-3d3acd2e-756a-442f-acf6-9a75ba7d933c tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Task: {'id': task-1340682, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 719.532724] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b2b06eaa-4237-476b-9167-ed8be81ba3b9 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Releasing lock "refresh_cache-f9a8e863-6e7f-4f00-b54a-78802659cd59" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 719.532866] env[62066]: DEBUG nova.compute.manager [None req-b2b06eaa-4237-476b-9167-ed8be81ba3b9 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 719.533057] env[62066]: DEBUG nova.compute.manager [None req-b2b06eaa-4237-476b-9167-ed8be81ba3b9 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 719.533228] env[62066]: DEBUG nova.network.neutron [None req-b2b06eaa-4237-476b-9167-ed8be81ba3b9 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 719.548963] env[62066]: DEBUG nova.network.neutron [None req-b2b06eaa-4237-476b-9167-ed8be81ba3b9 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 719.755401] env[62066]: DEBUG oslo_concurrency.lockutils [req-4efd7092-1575-4374-a39b-9b6520eae0c0 req-6d0283ea-5aac-45f6-9286-69693b82aa10 service nova] Releasing lock "refresh_cache-d7de90c0-95e9-4899-92a0-81658831713a" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 719.756862] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5b5d7867-70eb-472d-8d1a-c2bb035ec6db tempest-InstanceActionsV221TestJSON-1098427857 tempest-InstanceActionsV221TestJSON-1098427857-project-member] Acquired lock "refresh_cache-d7de90c0-95e9-4899-92a0-81658831713a" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 719.757152] env[62066]: DEBUG nova.network.neutron [None req-5b5d7867-70eb-472d-8d1a-c2bb035ec6db tempest-InstanceActionsV221TestJSON-1098427857 tempest-InstanceActionsV221TestJSON-1098427857-project-member] [instance: d7de90c0-95e9-4899-92a0-81658831713a] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 719.793612] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0188668d-25e1-420a-be60-fb5771349282 tempest-ServerAddressesTestJSON-1053293248 tempest-ServerAddressesTestJSON-1053293248-project-member] Acquiring lock "49bc1350-0095-406b-bc68-005eb6b681a8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 719.793841] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0188668d-25e1-420a-be60-fb5771349282 tempest-ServerAddressesTestJSON-1053293248 tempest-ServerAddressesTestJSON-1053293248-project-member] Lock "49bc1350-0095-406b-bc68-005eb6b681a8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 719.794606] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b90821f0-011b-47ea-a5fa-f7e3ca452af6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.802857] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-064f3ed9-290e-419f-a5c9-a33d331f87b8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.834637] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e299f66-d8e3-430b-952f-37118c8ecf0d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.841301] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d935ff6-cc09-4522-8a33-efcfdfe5e384 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.857161] env[62066]: DEBUG nova.compute.provider_tree [None req-5bf8bcd1-f041-4648-9b3c-4febff568fa9 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 719.983136] env[62066]: DEBUG oslo_vmware.api [None req-3d3acd2e-756a-442f-acf6-9a75ba7d933c tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Task: {'id': task-1340682, 'name': PowerOnVM_Task, 'duration_secs': 0.421599} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 719.983500] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d3acd2e-756a-442f-acf6-9a75ba7d933c tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] [instance: 0ff2f9f4-6d85-4465-8e0a-fb39f2a25c9b] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 719.983640] env[62066]: INFO nova.compute.manager [None req-3d3acd2e-756a-442f-acf6-9a75ba7d933c tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] [instance: 0ff2f9f4-6d85-4465-8e0a-fb39f2a25c9b] Took 3.86 seconds to spawn the instance on the hypervisor. [ 719.983825] env[62066]: DEBUG nova.compute.manager [None req-3d3acd2e-756a-442f-acf6-9a75ba7d933c tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] [instance: 0ff2f9f4-6d85-4465-8e0a-fb39f2a25c9b] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 719.984681] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce094041-d4f6-4627-ac0e-28a4f0d1eb3f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.053855] env[62066]: DEBUG nova.network.neutron [None req-b2b06eaa-4237-476b-9167-ed8be81ba3b9 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 720.279229] env[62066]: DEBUG nova.network.neutron [None req-5b5d7867-70eb-472d-8d1a-c2bb035ec6db tempest-InstanceActionsV221TestJSON-1098427857 tempest-InstanceActionsV221TestJSON-1098427857-project-member] [instance: d7de90c0-95e9-4899-92a0-81658831713a] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 720.361240] env[62066]: DEBUG nova.scheduler.client.report [None req-5bf8bcd1-f041-4648-9b3c-4febff568fa9 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 720.372516] env[62066]: DEBUG nova.network.neutron [None req-5b5d7867-70eb-472d-8d1a-c2bb035ec6db tempest-InstanceActionsV221TestJSON-1098427857 tempest-InstanceActionsV221TestJSON-1098427857-project-member] [instance: d7de90c0-95e9-4899-92a0-81658831713a] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 720.505018] env[62066]: INFO nova.compute.manager [None req-3d3acd2e-756a-442f-acf6-9a75ba7d933c tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] [instance: 0ff2f9f4-6d85-4465-8e0a-fb39f2a25c9b] Took 23.57 seconds to build instance. [ 720.557391] env[62066]: INFO nova.compute.manager [None req-b2b06eaa-4237-476b-9167-ed8be81ba3b9 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] [instance: f9a8e863-6e7f-4f00-b54a-78802659cd59] Took 1.02 seconds to deallocate network for instance. [ 720.671310] env[62066]: DEBUG nova.compute.manager [req-633183b4-f665-460b-8dea-8fbac510b2eb req-39200c9a-d03a-45c5-93bb-fe1dab8621ec service nova] [instance: d7de90c0-95e9-4899-92a0-81658831713a] Received event network-vif-deleted-389df780-4dc9-49a5-b7d6-c0cbbcd46c35 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 720.867023] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5bf8bcd1-f041-4648-9b3c-4febff568fa9 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.454s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 720.867023] env[62066]: DEBUG nova.compute.manager [None req-5bf8bcd1-f041-4648-9b3c-4febff568fa9 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] [instance: 3624ca43-9ee3-418a-9205-debb699295df] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 720.870328] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7d1d419b-6e61-4f85-9698-fe0c6e4f3988 tempest-AttachInterfacesV270Test-236309384 tempest-AttachInterfacesV270Test-236309384-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.292s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 720.871786] env[62066]: INFO nova.compute.claims [None req-7d1d419b-6e61-4f85-9698-fe0c6e4f3988 tempest-AttachInterfacesV270Test-236309384 tempest-AttachInterfacesV270Test-236309384-project-member] [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 720.875951] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5b5d7867-70eb-472d-8d1a-c2bb035ec6db tempest-InstanceActionsV221TestJSON-1098427857 tempest-InstanceActionsV221TestJSON-1098427857-project-member] Releasing lock "refresh_cache-d7de90c0-95e9-4899-92a0-81658831713a" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 720.876370] env[62066]: DEBUG nova.compute.manager [None req-5b5d7867-70eb-472d-8d1a-c2bb035ec6db tempest-InstanceActionsV221TestJSON-1098427857 tempest-InstanceActionsV221TestJSON-1098427857-project-member] [instance: d7de90c0-95e9-4899-92a0-81658831713a] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 720.876589] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-5b5d7867-70eb-472d-8d1a-c2bb035ec6db tempest-InstanceActionsV221TestJSON-1098427857 tempest-InstanceActionsV221TestJSON-1098427857-project-member] [instance: d7de90c0-95e9-4899-92a0-81658831713a] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 720.876860] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-aabf9029-83ad-4f0c-897f-cbc07cbabe48 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.885560] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf090c7a-8b3e-4f28-bd7e-a8d891e6b7f3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.909362] env[62066]: WARNING nova.virt.vmwareapi.vmops [None req-5b5d7867-70eb-472d-8d1a-c2bb035ec6db tempest-InstanceActionsV221TestJSON-1098427857 tempest-InstanceActionsV221TestJSON-1098427857-project-member] [instance: d7de90c0-95e9-4899-92a0-81658831713a] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance d7de90c0-95e9-4899-92a0-81658831713a could not be found. [ 720.909596] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-5b5d7867-70eb-472d-8d1a-c2bb035ec6db tempest-InstanceActionsV221TestJSON-1098427857 tempest-InstanceActionsV221TestJSON-1098427857-project-member] [instance: d7de90c0-95e9-4899-92a0-81658831713a] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 720.909775] env[62066]: INFO nova.compute.manager [None req-5b5d7867-70eb-472d-8d1a-c2bb035ec6db tempest-InstanceActionsV221TestJSON-1098427857 tempest-InstanceActionsV221TestJSON-1098427857-project-member] [instance: d7de90c0-95e9-4899-92a0-81658831713a] Took 0.03 seconds to destroy the instance on the hypervisor. [ 720.910020] env[62066]: DEBUG oslo.service.loopingcall [None req-5b5d7867-70eb-472d-8d1a-c2bb035ec6db tempest-InstanceActionsV221TestJSON-1098427857 tempest-InstanceActionsV221TestJSON-1098427857-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 720.910546] env[62066]: DEBUG nova.compute.manager [-] [instance: d7de90c0-95e9-4899-92a0-81658831713a] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 720.910658] env[62066]: DEBUG nova.network.neutron [-] [instance: d7de90c0-95e9-4899-92a0-81658831713a] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 720.925900] env[62066]: DEBUG nova.network.neutron [-] [instance: d7de90c0-95e9-4899-92a0-81658831713a] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 721.007096] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3d3acd2e-756a-442f-acf6-9a75ba7d933c tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Lock "0ff2f9f4-6d85-4465-8e0a-fb39f2a25c9b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 151.830s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 721.243741] env[62066]: INFO nova.compute.manager [None req-fbd85018-c9a9-4c9f-87ff-96cad2b995f9 tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] [instance: 0ff2f9f4-6d85-4465-8e0a-fb39f2a25c9b] Rebuilding instance [ 721.289028] env[62066]: DEBUG nova.compute.manager [None req-fbd85018-c9a9-4c9f-87ff-96cad2b995f9 tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] [instance: 0ff2f9f4-6d85-4465-8e0a-fb39f2a25c9b] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 721.290041] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a82240b-636f-43c3-b306-28f371cf2207 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.372651] env[62066]: DEBUG nova.compute.utils [None req-5bf8bcd1-f041-4648-9b3c-4febff568fa9 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 721.374063] env[62066]: DEBUG nova.compute.manager [None req-5bf8bcd1-f041-4648-9b3c-4febff568fa9 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] [instance: 3624ca43-9ee3-418a-9205-debb699295df] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 721.375032] env[62066]: DEBUG nova.network.neutron [None req-5bf8bcd1-f041-4648-9b3c-4febff568fa9 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] [instance: 3624ca43-9ee3-418a-9205-debb699295df] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 721.417364] env[62066]: DEBUG nova.policy [None req-5bf8bcd1-f041-4648-9b3c-4febff568fa9 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9d11ab2ea74442369fa1d93daf98392b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'dd30584efc254610b3939d1ba3806693', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 721.428032] env[62066]: DEBUG nova.network.neutron [-] [instance: d7de90c0-95e9-4899-92a0-81658831713a] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 721.510026] env[62066]: DEBUG nova.compute.manager [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] [instance: d973b09f-7727-4055-b05e-123079df9ce4] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 721.589186] env[62066]: INFO nova.scheduler.client.report [None req-b2b06eaa-4237-476b-9167-ed8be81ba3b9 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Deleted allocations for instance f9a8e863-6e7f-4f00-b54a-78802659cd59 [ 721.697175] env[62066]: DEBUG nova.network.neutron [None req-5bf8bcd1-f041-4648-9b3c-4febff568fa9 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] [instance: 3624ca43-9ee3-418a-9205-debb699295df] Successfully created port: bf6b4191-9489-4988-8fd5-320a5a2607a1 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 721.800886] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-fbd85018-c9a9-4c9f-87ff-96cad2b995f9 tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] [instance: 0ff2f9f4-6d85-4465-8e0a-fb39f2a25c9b] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 721.801234] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-731350a2-cfec-4256-84f8-a2737b3dcd47 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.808351] env[62066]: DEBUG oslo_vmware.api [None req-fbd85018-c9a9-4c9f-87ff-96cad2b995f9 tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Waiting for the task: (returnval){ [ 721.808351] env[62066]: value = "task-1340683" [ 721.808351] env[62066]: _type = "Task" [ 721.808351] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 721.816307] env[62066]: DEBUG oslo_vmware.api [None req-fbd85018-c9a9-4c9f-87ff-96cad2b995f9 tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Task: {'id': task-1340683, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 721.876865] env[62066]: DEBUG nova.compute.manager [None req-5bf8bcd1-f041-4648-9b3c-4febff568fa9 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] [instance: 3624ca43-9ee3-418a-9205-debb699295df] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 721.931809] env[62066]: INFO nova.compute.manager [-] [instance: d7de90c0-95e9-4899-92a0-81658831713a] Took 1.02 seconds to deallocate network for instance. [ 721.936264] env[62066]: DEBUG nova.compute.claims [None req-5b5d7867-70eb-472d-8d1a-c2bb035ec6db tempest-InstanceActionsV221TestJSON-1098427857 tempest-InstanceActionsV221TestJSON-1098427857-project-member] [instance: d7de90c0-95e9-4899-92a0-81658831713a] Aborting claim: {{(pid=62066) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 721.936264] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5b5d7867-70eb-472d-8d1a-c2bb035ec6db tempest-InstanceActionsV221TestJSON-1098427857 tempest-InstanceActionsV221TestJSON-1098427857-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 722.027921] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 722.096569] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b2b06eaa-4237-476b-9167-ed8be81ba3b9 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Lock "f9a8e863-6e7f-4f00-b54a-78802659cd59" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 161.224s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 722.247873] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9076b9f5-e3bd-48ab-b96f-803421d1f2c2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.256358] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac4ebd47-cd4a-4e20-bdcd-9c57eeb74414 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.288052] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ae4dac5-55f8-435c-96fb-ff2c70b404be {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.296829] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2953e6a8-f1df-49ad-9d15-29ca42b1bc26 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.311967] env[62066]: DEBUG nova.compute.provider_tree [None req-7d1d419b-6e61-4f85-9698-fe0c6e4f3988 tempest-AttachInterfacesV270Test-236309384 tempest-AttachInterfacesV270Test-236309384-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 722.327145] env[62066]: DEBUG oslo_vmware.api [None req-fbd85018-c9a9-4c9f-87ff-96cad2b995f9 tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Task: {'id': task-1340683, 'name': PowerOffVM_Task, 'duration_secs': 0.203929} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 722.328020] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-fbd85018-c9a9-4c9f-87ff-96cad2b995f9 tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] [instance: 0ff2f9f4-6d85-4465-8e0a-fb39f2a25c9b] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 722.328425] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-fbd85018-c9a9-4c9f-87ff-96cad2b995f9 tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] [instance: 0ff2f9f4-6d85-4465-8e0a-fb39f2a25c9b] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 722.329138] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6a9d4a7-de6e-45ac-b10d-e417dd4c4009 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.337171] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-fbd85018-c9a9-4c9f-87ff-96cad2b995f9 tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] [instance: 0ff2f9f4-6d85-4465-8e0a-fb39f2a25c9b] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 722.337408] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b2402ed6-537e-4b28-bac3-45c635ac46c7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.364663] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-fbd85018-c9a9-4c9f-87ff-96cad2b995f9 tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] [instance: 0ff2f9f4-6d85-4465-8e0a-fb39f2a25c9b] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 722.364891] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-fbd85018-c9a9-4c9f-87ff-96cad2b995f9 tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] [instance: 0ff2f9f4-6d85-4465-8e0a-fb39f2a25c9b] Deleting contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 722.365083] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-fbd85018-c9a9-4c9f-87ff-96cad2b995f9 tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Deleting the datastore file [datastore2] 0ff2f9f4-6d85-4465-8e0a-fb39f2a25c9b {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 722.365347] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f6f77ef2-4146-4b58-977b-f881b330f170 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.372402] env[62066]: DEBUG oslo_vmware.api [None req-fbd85018-c9a9-4c9f-87ff-96cad2b995f9 tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Waiting for the task: (returnval){ [ 722.372402] env[62066]: value = "task-1340685" [ 722.372402] env[62066]: _type = "Task" [ 722.372402] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 722.382633] env[62066]: DEBUG oslo_vmware.api [None req-fbd85018-c9a9-4c9f-87ff-96cad2b995f9 tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Task: {'id': task-1340685, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 722.405977] env[62066]: DEBUG nova.compute.manager [req-1f70c524-b322-45bb-b3cf-d3cb1c62f82c req-09d13711-1cda-45d9-bdac-91365fe87383 service nova] [instance: 3624ca43-9ee3-418a-9205-debb699295df] Received event network-changed-bf6b4191-9489-4988-8fd5-320a5a2607a1 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 722.405977] env[62066]: DEBUG nova.compute.manager [req-1f70c524-b322-45bb-b3cf-d3cb1c62f82c req-09d13711-1cda-45d9-bdac-91365fe87383 service nova] [instance: 3624ca43-9ee3-418a-9205-debb699295df] Refreshing instance network info cache due to event network-changed-bf6b4191-9489-4988-8fd5-320a5a2607a1. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 722.405977] env[62066]: DEBUG oslo_concurrency.lockutils [req-1f70c524-b322-45bb-b3cf-d3cb1c62f82c req-09d13711-1cda-45d9-bdac-91365fe87383 service nova] Acquiring lock "refresh_cache-3624ca43-9ee3-418a-9205-debb699295df" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 722.405977] env[62066]: DEBUG oslo_concurrency.lockutils [req-1f70c524-b322-45bb-b3cf-d3cb1c62f82c req-09d13711-1cda-45d9-bdac-91365fe87383 service nova] Acquired lock "refresh_cache-3624ca43-9ee3-418a-9205-debb699295df" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 722.405977] env[62066]: DEBUG nova.network.neutron [req-1f70c524-b322-45bb-b3cf-d3cb1c62f82c req-09d13711-1cda-45d9-bdac-91365fe87383 service nova] [instance: 3624ca43-9ee3-418a-9205-debb699295df] Refreshing network info cache for port bf6b4191-9489-4988-8fd5-320a5a2607a1 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 722.570179] env[62066]: ERROR nova.compute.manager [None req-5bf8bcd1-f041-4648-9b3c-4febff568fa9 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port bf6b4191-9489-4988-8fd5-320a5a2607a1, please check neutron logs for more information. [ 722.570179] env[62066]: ERROR nova.compute.manager Traceback (most recent call last): [ 722.570179] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 722.570179] env[62066]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 722.570179] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 722.570179] env[62066]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 722.570179] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 722.570179] env[62066]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 722.570179] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 722.570179] env[62066]: ERROR nova.compute.manager self.force_reraise() [ 722.570179] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 722.570179] env[62066]: ERROR nova.compute.manager raise self.value [ 722.570179] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 722.570179] env[62066]: ERROR nova.compute.manager updated_port = self._update_port( [ 722.570179] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 722.570179] env[62066]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 722.571161] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 722.571161] env[62066]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 722.571161] env[62066]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port bf6b4191-9489-4988-8fd5-320a5a2607a1, please check neutron logs for more information. [ 722.571161] env[62066]: ERROR nova.compute.manager [ 722.571161] env[62066]: Traceback (most recent call last): [ 722.571161] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 722.571161] env[62066]: listener.cb(fileno) [ 722.571161] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 722.571161] env[62066]: result = function(*args, **kwargs) [ 722.571161] env[62066]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 722.571161] env[62066]: return func(*args, **kwargs) [ 722.571161] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 722.571161] env[62066]: raise e [ 722.571161] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 722.571161] env[62066]: nwinfo = self.network_api.allocate_for_instance( [ 722.571161] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 722.571161] env[62066]: created_port_ids = self._update_ports_for_instance( [ 722.571161] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 722.571161] env[62066]: with excutils.save_and_reraise_exception(): [ 722.571161] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 722.571161] env[62066]: self.force_reraise() [ 722.571161] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 722.571161] env[62066]: raise self.value [ 722.571161] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 722.571161] env[62066]: updated_port = self._update_port( [ 722.571161] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 722.571161] env[62066]: _ensure_no_port_binding_failure(port) [ 722.571161] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 722.571161] env[62066]: raise exception.PortBindingFailed(port_id=port['id']) [ 722.572120] env[62066]: nova.exception.PortBindingFailed: Binding failed for port bf6b4191-9489-4988-8fd5-320a5a2607a1, please check neutron logs for more information. [ 722.572120] env[62066]: Removing descriptor: 15 [ 722.600994] env[62066]: DEBUG nova.compute.manager [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 722.822457] env[62066]: DEBUG nova.scheduler.client.report [None req-7d1d419b-6e61-4f85-9698-fe0c6e4f3988 tempest-AttachInterfacesV270Test-236309384 tempest-AttachInterfacesV270Test-236309384-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 722.885127] env[62066]: DEBUG oslo_vmware.api [None req-fbd85018-c9a9-4c9f-87ff-96cad2b995f9 tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Task: {'id': task-1340685, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.09961} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 722.885312] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-fbd85018-c9a9-4c9f-87ff-96cad2b995f9 tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 722.885494] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-fbd85018-c9a9-4c9f-87ff-96cad2b995f9 tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] [instance: 0ff2f9f4-6d85-4465-8e0a-fb39f2a25c9b] Deleted contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 722.885662] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-fbd85018-c9a9-4c9f-87ff-96cad2b995f9 tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] [instance: 0ff2f9f4-6d85-4465-8e0a-fb39f2a25c9b] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 722.891205] env[62066]: DEBUG nova.compute.manager [None req-5bf8bcd1-f041-4648-9b3c-4febff568fa9 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] [instance: 3624ca43-9ee3-418a-9205-debb699295df] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 722.920850] env[62066]: DEBUG nova.virt.hardware [None req-5bf8bcd1-f041-4648-9b3c-4febff568fa9 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-23T13:40:41Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-23T13:40:23Z,direct_url=,disk_format='vmdk',id=50ff584c-3b50-4395-af07-3e66769bc9f7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='eb52f7069a374c61ae946f052007c6d9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-23T13:40:24Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 722.921068] env[62066]: DEBUG nova.virt.hardware [None req-5bf8bcd1-f041-4648-9b3c-4febff568fa9 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 722.921274] env[62066]: DEBUG nova.virt.hardware [None req-5bf8bcd1-f041-4648-9b3c-4febff568fa9 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 722.921511] env[62066]: DEBUG nova.virt.hardware [None req-5bf8bcd1-f041-4648-9b3c-4febff568fa9 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 722.921667] env[62066]: DEBUG nova.virt.hardware [None req-5bf8bcd1-f041-4648-9b3c-4febff568fa9 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 722.921811] env[62066]: DEBUG nova.virt.hardware [None req-5bf8bcd1-f041-4648-9b3c-4febff568fa9 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 722.922058] env[62066]: DEBUG nova.virt.hardware [None req-5bf8bcd1-f041-4648-9b3c-4febff568fa9 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 722.922256] env[62066]: DEBUG nova.virt.hardware [None req-5bf8bcd1-f041-4648-9b3c-4febff568fa9 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 722.922459] env[62066]: DEBUG nova.virt.hardware [None req-5bf8bcd1-f041-4648-9b3c-4febff568fa9 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 722.922655] env[62066]: DEBUG nova.virt.hardware [None req-5bf8bcd1-f041-4648-9b3c-4febff568fa9 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 722.922870] env[62066]: DEBUG nova.virt.hardware [None req-5bf8bcd1-f041-4648-9b3c-4febff568fa9 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 722.923806] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00328d8e-a376-47cf-8a47-d1c3d1e33bab {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.929135] env[62066]: DEBUG nova.network.neutron [req-1f70c524-b322-45bb-b3cf-d3cb1c62f82c req-09d13711-1cda-45d9-bdac-91365fe87383 service nova] [instance: 3624ca43-9ee3-418a-9205-debb699295df] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 722.935722] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2586e018-ad40-43e7-a9c1-b8424a94191f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.950775] env[62066]: ERROR nova.compute.manager [None req-5bf8bcd1-f041-4648-9b3c-4febff568fa9 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] [instance: 3624ca43-9ee3-418a-9205-debb699295df] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port bf6b4191-9489-4988-8fd5-320a5a2607a1, please check neutron logs for more information. [ 722.950775] env[62066]: ERROR nova.compute.manager [instance: 3624ca43-9ee3-418a-9205-debb699295df] Traceback (most recent call last): [ 722.950775] env[62066]: ERROR nova.compute.manager [instance: 3624ca43-9ee3-418a-9205-debb699295df] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 722.950775] env[62066]: ERROR nova.compute.manager [instance: 3624ca43-9ee3-418a-9205-debb699295df] yield resources [ 722.950775] env[62066]: ERROR nova.compute.manager [instance: 3624ca43-9ee3-418a-9205-debb699295df] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 722.950775] env[62066]: ERROR nova.compute.manager [instance: 3624ca43-9ee3-418a-9205-debb699295df] self.driver.spawn(context, instance, image_meta, [ 722.950775] env[62066]: ERROR nova.compute.manager [instance: 3624ca43-9ee3-418a-9205-debb699295df] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 722.950775] env[62066]: ERROR nova.compute.manager [instance: 3624ca43-9ee3-418a-9205-debb699295df] self._vmops.spawn(context, instance, image_meta, injected_files, [ 722.950775] env[62066]: ERROR nova.compute.manager [instance: 3624ca43-9ee3-418a-9205-debb699295df] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 722.950775] env[62066]: ERROR nova.compute.manager [instance: 3624ca43-9ee3-418a-9205-debb699295df] vm_ref = self.build_virtual_machine(instance, [ 722.950775] env[62066]: ERROR nova.compute.manager [instance: 3624ca43-9ee3-418a-9205-debb699295df] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 722.951232] env[62066]: ERROR nova.compute.manager [instance: 3624ca43-9ee3-418a-9205-debb699295df] vif_infos = vmwarevif.get_vif_info(self._session, [ 722.951232] env[62066]: ERROR nova.compute.manager [instance: 3624ca43-9ee3-418a-9205-debb699295df] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 722.951232] env[62066]: ERROR nova.compute.manager [instance: 3624ca43-9ee3-418a-9205-debb699295df] for vif in network_info: [ 722.951232] env[62066]: ERROR nova.compute.manager [instance: 3624ca43-9ee3-418a-9205-debb699295df] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 722.951232] env[62066]: ERROR nova.compute.manager [instance: 3624ca43-9ee3-418a-9205-debb699295df] return self._sync_wrapper(fn, *args, **kwargs) [ 722.951232] env[62066]: ERROR nova.compute.manager [instance: 3624ca43-9ee3-418a-9205-debb699295df] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 722.951232] env[62066]: ERROR nova.compute.manager [instance: 3624ca43-9ee3-418a-9205-debb699295df] self.wait() [ 722.951232] env[62066]: ERROR nova.compute.manager [instance: 3624ca43-9ee3-418a-9205-debb699295df] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 722.951232] env[62066]: ERROR nova.compute.manager [instance: 3624ca43-9ee3-418a-9205-debb699295df] self[:] = self._gt.wait() [ 722.951232] env[62066]: ERROR nova.compute.manager [instance: 3624ca43-9ee3-418a-9205-debb699295df] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 722.951232] env[62066]: ERROR nova.compute.manager [instance: 3624ca43-9ee3-418a-9205-debb699295df] return self._exit_event.wait() [ 722.951232] env[62066]: ERROR nova.compute.manager [instance: 3624ca43-9ee3-418a-9205-debb699295df] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 722.951232] env[62066]: ERROR nova.compute.manager [instance: 3624ca43-9ee3-418a-9205-debb699295df] current.throw(*self._exc) [ 722.951642] env[62066]: ERROR nova.compute.manager [instance: 3624ca43-9ee3-418a-9205-debb699295df] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 722.951642] env[62066]: ERROR nova.compute.manager [instance: 3624ca43-9ee3-418a-9205-debb699295df] result = function(*args, **kwargs) [ 722.951642] env[62066]: ERROR nova.compute.manager [instance: 3624ca43-9ee3-418a-9205-debb699295df] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 722.951642] env[62066]: ERROR nova.compute.manager [instance: 3624ca43-9ee3-418a-9205-debb699295df] return func(*args, **kwargs) [ 722.951642] env[62066]: ERROR nova.compute.manager [instance: 3624ca43-9ee3-418a-9205-debb699295df] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 722.951642] env[62066]: ERROR nova.compute.manager [instance: 3624ca43-9ee3-418a-9205-debb699295df] raise e [ 722.951642] env[62066]: ERROR nova.compute.manager [instance: 3624ca43-9ee3-418a-9205-debb699295df] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 722.951642] env[62066]: ERROR nova.compute.manager [instance: 3624ca43-9ee3-418a-9205-debb699295df] nwinfo = self.network_api.allocate_for_instance( [ 722.951642] env[62066]: ERROR nova.compute.manager [instance: 3624ca43-9ee3-418a-9205-debb699295df] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 722.951642] env[62066]: ERROR nova.compute.manager [instance: 3624ca43-9ee3-418a-9205-debb699295df] created_port_ids = self._update_ports_for_instance( [ 722.951642] env[62066]: ERROR nova.compute.manager [instance: 3624ca43-9ee3-418a-9205-debb699295df] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 722.951642] env[62066]: ERROR nova.compute.manager [instance: 3624ca43-9ee3-418a-9205-debb699295df] with excutils.save_and_reraise_exception(): [ 722.951642] env[62066]: ERROR nova.compute.manager [instance: 3624ca43-9ee3-418a-9205-debb699295df] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 722.952054] env[62066]: ERROR nova.compute.manager [instance: 3624ca43-9ee3-418a-9205-debb699295df] self.force_reraise() [ 722.952054] env[62066]: ERROR nova.compute.manager [instance: 3624ca43-9ee3-418a-9205-debb699295df] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 722.952054] env[62066]: ERROR nova.compute.manager [instance: 3624ca43-9ee3-418a-9205-debb699295df] raise self.value [ 722.952054] env[62066]: ERROR nova.compute.manager [instance: 3624ca43-9ee3-418a-9205-debb699295df] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 722.952054] env[62066]: ERROR nova.compute.manager [instance: 3624ca43-9ee3-418a-9205-debb699295df] updated_port = self._update_port( [ 722.952054] env[62066]: ERROR nova.compute.manager [instance: 3624ca43-9ee3-418a-9205-debb699295df] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 722.952054] env[62066]: ERROR nova.compute.manager [instance: 3624ca43-9ee3-418a-9205-debb699295df] _ensure_no_port_binding_failure(port) [ 722.952054] env[62066]: ERROR nova.compute.manager [instance: 3624ca43-9ee3-418a-9205-debb699295df] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 722.952054] env[62066]: ERROR nova.compute.manager [instance: 3624ca43-9ee3-418a-9205-debb699295df] raise exception.PortBindingFailed(port_id=port['id']) [ 722.952054] env[62066]: ERROR nova.compute.manager [instance: 3624ca43-9ee3-418a-9205-debb699295df] nova.exception.PortBindingFailed: Binding failed for port bf6b4191-9489-4988-8fd5-320a5a2607a1, please check neutron logs for more information. [ 722.952054] env[62066]: ERROR nova.compute.manager [instance: 3624ca43-9ee3-418a-9205-debb699295df] [ 722.952054] env[62066]: INFO nova.compute.manager [None req-5bf8bcd1-f041-4648-9b3c-4febff568fa9 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] [instance: 3624ca43-9ee3-418a-9205-debb699295df] Terminating instance [ 722.953052] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5bf8bcd1-f041-4648-9b3c-4febff568fa9 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Acquiring lock "refresh_cache-3624ca43-9ee3-418a-9205-debb699295df" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 723.033015] env[62066]: DEBUG nova.network.neutron [req-1f70c524-b322-45bb-b3cf-d3cb1c62f82c req-09d13711-1cda-45d9-bdac-91365fe87383 service nova] [instance: 3624ca43-9ee3-418a-9205-debb699295df] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 723.123696] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 723.326951] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7d1d419b-6e61-4f85-9698-fe0c6e4f3988 tempest-AttachInterfacesV270Test-236309384 tempest-AttachInterfacesV270Test-236309384-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.456s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 723.327537] env[62066]: DEBUG nova.compute.manager [None req-7d1d419b-6e61-4f85-9698-fe0c6e4f3988 tempest-AttachInterfacesV270Test-236309384 tempest-AttachInterfacesV270Test-236309384-project-member] [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 723.330974] env[62066]: DEBUG oslo_concurrency.lockutils [None req-84147b21-d606-4daf-b0b9-f581072be1cb tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.583s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 723.536082] env[62066]: DEBUG oslo_concurrency.lockutils [req-1f70c524-b322-45bb-b3cf-d3cb1c62f82c req-09d13711-1cda-45d9-bdac-91365fe87383 service nova] Releasing lock "refresh_cache-3624ca43-9ee3-418a-9205-debb699295df" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 723.536566] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5bf8bcd1-f041-4648-9b3c-4febff568fa9 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Acquired lock "refresh_cache-3624ca43-9ee3-418a-9205-debb699295df" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 723.536811] env[62066]: DEBUG nova.network.neutron [None req-5bf8bcd1-f041-4648-9b3c-4febff568fa9 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] [instance: 3624ca43-9ee3-418a-9205-debb699295df] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 723.833061] env[62066]: DEBUG nova.compute.utils [None req-7d1d419b-6e61-4f85-9698-fe0c6e4f3988 tempest-AttachInterfacesV270Test-236309384 tempest-AttachInterfacesV270Test-236309384-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 723.836964] env[62066]: DEBUG nova.compute.manager [None req-7d1d419b-6e61-4f85-9698-fe0c6e4f3988 tempest-AttachInterfacesV270Test-236309384 tempest-AttachInterfacesV270Test-236309384-project-member] [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 723.837289] env[62066]: DEBUG nova.network.neutron [None req-7d1d419b-6e61-4f85-9698-fe0c6e4f3988 tempest-AttachInterfacesV270Test-236309384 tempest-AttachInterfacesV270Test-236309384-project-member] [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 723.909942] env[62066]: DEBUG nova.policy [None req-7d1d419b-6e61-4f85-9698-fe0c6e4f3988 tempest-AttachInterfacesV270Test-236309384 tempest-AttachInterfacesV270Test-236309384-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b1fdb8b4b0b94e579194e5d18d06bd81', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '50161cb62d674467aa16a5150984be9a', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 723.927673] env[62066]: DEBUG nova.virt.hardware [None req-fbd85018-c9a9-4c9f-87ff-96cad2b995f9 tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-23T13:40:41Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-23T13:40:23Z,direct_url=,disk_format='vmdk',id=50ff584c-3b50-4395-af07-3e66769bc9f7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='eb52f7069a374c61ae946f052007c6d9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-23T13:40:24Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 723.927887] env[62066]: DEBUG nova.virt.hardware [None req-fbd85018-c9a9-4c9f-87ff-96cad2b995f9 tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 723.928052] env[62066]: DEBUG nova.virt.hardware [None req-fbd85018-c9a9-4c9f-87ff-96cad2b995f9 tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 723.928243] env[62066]: DEBUG nova.virt.hardware [None req-fbd85018-c9a9-4c9f-87ff-96cad2b995f9 tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 723.928437] env[62066]: DEBUG nova.virt.hardware [None req-fbd85018-c9a9-4c9f-87ff-96cad2b995f9 tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 723.928563] env[62066]: DEBUG nova.virt.hardware [None req-fbd85018-c9a9-4c9f-87ff-96cad2b995f9 tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 723.928809] env[62066]: DEBUG nova.virt.hardware [None req-fbd85018-c9a9-4c9f-87ff-96cad2b995f9 tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 723.929510] env[62066]: DEBUG nova.virt.hardware [None req-fbd85018-c9a9-4c9f-87ff-96cad2b995f9 tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 723.929744] env[62066]: DEBUG nova.virt.hardware [None req-fbd85018-c9a9-4c9f-87ff-96cad2b995f9 tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 723.929921] env[62066]: DEBUG nova.virt.hardware [None req-fbd85018-c9a9-4c9f-87ff-96cad2b995f9 tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 723.930116] env[62066]: DEBUG nova.virt.hardware [None req-fbd85018-c9a9-4c9f-87ff-96cad2b995f9 tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 723.931374] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b01fd024-d1f4-48f9-9ffb-972f268b52f9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.942288] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80d558f3-a5be-4988-9936-af77fb4ff11d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.958786] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-fbd85018-c9a9-4c9f-87ff-96cad2b995f9 tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] [instance: 0ff2f9f4-6d85-4465-8e0a-fb39f2a25c9b] Instance VIF info [] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 723.965099] env[62066]: DEBUG oslo.service.loopingcall [None req-fbd85018-c9a9-4c9f-87ff-96cad2b995f9 tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 723.965262] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0ff2f9f4-6d85-4465-8e0a-fb39f2a25c9b] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 723.965489] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1c5c11be-7bd2-4b74-9f5b-6b648652aeed {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.985419] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 723.985419] env[62066]: value = "task-1340686" [ 723.985419] env[62066]: _type = "Task" [ 723.985419] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 723.992937] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1340686, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 724.063155] env[62066]: DEBUG nova.network.neutron [None req-5bf8bcd1-f041-4648-9b3c-4febff568fa9 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] [instance: 3624ca43-9ee3-418a-9205-debb699295df] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 724.198149] env[62066]: DEBUG nova.network.neutron [None req-5bf8bcd1-f041-4648-9b3c-4febff568fa9 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] [instance: 3624ca43-9ee3-418a-9205-debb699295df] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 724.238117] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01d289a1-5598-4bcd-b865-9b80f7fe8b1a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.247238] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6dd1906-16f7-442f-a08f-b0ded1b5b14d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.277646] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cc0a1eb-7fc9-4b82-85ef-200e09513dd7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.285242] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e45693d-f0fc-42c8-9cf1-bd4ffe97c218 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.298896] env[62066]: DEBUG nova.compute.provider_tree [None req-84147b21-d606-4daf-b0b9-f581072be1cb tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 724.337906] env[62066]: DEBUG nova.compute.manager [None req-7d1d419b-6e61-4f85-9698-fe0c6e4f3988 tempest-AttachInterfacesV270Test-236309384 tempest-AttachInterfacesV270Test-236309384-project-member] [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 724.433544] env[62066]: DEBUG nova.compute.manager [req-d7dab555-9703-4bd9-b944-163f850a7d1b req-17b7a4b3-c69d-406d-9bcb-15fa664cdd6b service nova] [instance: 3624ca43-9ee3-418a-9205-debb699295df] Received event network-vif-deleted-bf6b4191-9489-4988-8fd5-320a5a2607a1 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 724.473240] env[62066]: DEBUG nova.network.neutron [None req-7d1d419b-6e61-4f85-9698-fe0c6e4f3988 tempest-AttachInterfacesV270Test-236309384 tempest-AttachInterfacesV270Test-236309384-project-member] [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] Successfully created port: a3e86a4a-9e56-428c-a598-e42c247ab0e2 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 724.495681] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1340686, 'name': CreateVM_Task, 'duration_secs': 0.295154} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 724.495845] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0ff2f9f4-6d85-4465-8e0a-fb39f2a25c9b] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 724.496813] env[62066]: DEBUG oslo_concurrency.lockutils [None req-fbd85018-c9a9-4c9f-87ff-96cad2b995f9 tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 724.496946] env[62066]: DEBUG oslo_concurrency.lockutils [None req-fbd85018-c9a9-4c9f-87ff-96cad2b995f9 tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Acquired lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 724.497693] env[62066]: DEBUG oslo_concurrency.lockutils [None req-fbd85018-c9a9-4c9f-87ff-96cad2b995f9 tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 724.497693] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3555ede4-0cc1-4240-9a64-f3f421415dd1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.502289] env[62066]: DEBUG oslo_vmware.api [None req-fbd85018-c9a9-4c9f-87ff-96cad2b995f9 tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Waiting for the task: (returnval){ [ 724.502289] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]52ceb6a0-fcb2-f733-bbe7-86c3584e952e" [ 724.502289] env[62066]: _type = "Task" [ 724.502289] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 724.511311] env[62066]: DEBUG oslo_vmware.api [None req-fbd85018-c9a9-4c9f-87ff-96cad2b995f9 tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52ceb6a0-fcb2-f733-bbe7-86c3584e952e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 724.701904] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5bf8bcd1-f041-4648-9b3c-4febff568fa9 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Releasing lock "refresh_cache-3624ca43-9ee3-418a-9205-debb699295df" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 724.702212] env[62066]: DEBUG nova.compute.manager [None req-5bf8bcd1-f041-4648-9b3c-4febff568fa9 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] [instance: 3624ca43-9ee3-418a-9205-debb699295df] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 724.702412] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-5bf8bcd1-f041-4648-9b3c-4febff568fa9 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] [instance: 3624ca43-9ee3-418a-9205-debb699295df] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 724.703066] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-fe1aaafb-773b-4abe-8909-f45e231d1faa {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.712976] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7531582-da83-4085-8cc7-77d859f94613 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.737805] env[62066]: WARNING nova.virt.vmwareapi.vmops [None req-5bf8bcd1-f041-4648-9b3c-4febff568fa9 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] [instance: 3624ca43-9ee3-418a-9205-debb699295df] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 3624ca43-9ee3-418a-9205-debb699295df could not be found. [ 724.738356] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-5bf8bcd1-f041-4648-9b3c-4febff568fa9 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] [instance: 3624ca43-9ee3-418a-9205-debb699295df] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 724.738356] env[62066]: INFO nova.compute.manager [None req-5bf8bcd1-f041-4648-9b3c-4febff568fa9 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] [instance: 3624ca43-9ee3-418a-9205-debb699295df] Took 0.04 seconds to destroy the instance on the hypervisor. [ 724.738680] env[62066]: DEBUG oslo.service.loopingcall [None req-5bf8bcd1-f041-4648-9b3c-4febff568fa9 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 724.738782] env[62066]: DEBUG nova.compute.manager [-] [instance: 3624ca43-9ee3-418a-9205-debb699295df] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 724.738859] env[62066]: DEBUG nova.network.neutron [-] [instance: 3624ca43-9ee3-418a-9205-debb699295df] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 724.761185] env[62066]: DEBUG nova.network.neutron [-] [instance: 3624ca43-9ee3-418a-9205-debb699295df] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 724.802213] env[62066]: DEBUG nova.scheduler.client.report [None req-84147b21-d606-4daf-b0b9-f581072be1cb tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 725.012021] env[62066]: DEBUG oslo_vmware.api [None req-fbd85018-c9a9-4c9f-87ff-96cad2b995f9 tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52ceb6a0-fcb2-f733-bbe7-86c3584e952e, 'name': SearchDatastore_Task, 'duration_secs': 0.009546} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 725.013039] env[62066]: DEBUG oslo_concurrency.lockutils [None req-fbd85018-c9a9-4c9f-87ff-96cad2b995f9 tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Releasing lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 725.013445] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-fbd85018-c9a9-4c9f-87ff-96cad2b995f9 tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] [instance: 0ff2f9f4-6d85-4465-8e0a-fb39f2a25c9b] Processing image 50ff584c-3b50-4395-af07-3e66769bc9f7 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 725.013947] env[62066]: DEBUG oslo_concurrency.lockutils [None req-fbd85018-c9a9-4c9f-87ff-96cad2b995f9 tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 725.014143] env[62066]: DEBUG oslo_concurrency.lockutils [None req-fbd85018-c9a9-4c9f-87ff-96cad2b995f9 tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Acquired lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 725.014334] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-fbd85018-c9a9-4c9f-87ff-96cad2b995f9 tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 725.014870] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-cfeb255b-c5e3-4341-aa02-a8ce1db4d73d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.023285] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-fbd85018-c9a9-4c9f-87ff-96cad2b995f9 tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 725.023460] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-fbd85018-c9a9-4c9f-87ff-96cad2b995f9 tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 725.024207] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0ddaf4fe-60f4-4df0-95b2-d2e89dc76212 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.031325] env[62066]: DEBUG oslo_vmware.api [None req-fbd85018-c9a9-4c9f-87ff-96cad2b995f9 tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Waiting for the task: (returnval){ [ 725.031325] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]520968ae-459f-d62a-c205-59796d26e557" [ 725.031325] env[62066]: _type = "Task" [ 725.031325] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 725.044860] env[62066]: DEBUG oslo_vmware.api [None req-fbd85018-c9a9-4c9f-87ff-96cad2b995f9 tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]520968ae-459f-d62a-c205-59796d26e557, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 725.263127] env[62066]: DEBUG nova.network.neutron [-] [instance: 3624ca43-9ee3-418a-9205-debb699295df] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 725.307827] env[62066]: DEBUG oslo_concurrency.lockutils [None req-84147b21-d606-4daf-b0b9-f581072be1cb tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.976s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 725.308071] env[62066]: ERROR nova.compute.manager [None req-84147b21-d606-4daf-b0b9-f581072be1cb tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 11527383-8bd1-4760-9702-997440e384c8, please check neutron logs for more information. [ 725.308071] env[62066]: ERROR nova.compute.manager [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] Traceback (most recent call last): [ 725.308071] env[62066]: ERROR nova.compute.manager [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 725.308071] env[62066]: ERROR nova.compute.manager [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] self.driver.spawn(context, instance, image_meta, [ 725.308071] env[62066]: ERROR nova.compute.manager [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 725.308071] env[62066]: ERROR nova.compute.manager [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] self._vmops.spawn(context, instance, image_meta, injected_files, [ 725.308071] env[62066]: ERROR nova.compute.manager [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 725.308071] env[62066]: ERROR nova.compute.manager [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] vm_ref = self.build_virtual_machine(instance, [ 725.308071] env[62066]: ERROR nova.compute.manager [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 725.308071] env[62066]: ERROR nova.compute.manager [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] vif_infos = vmwarevif.get_vif_info(self._session, [ 725.308071] env[62066]: ERROR nova.compute.manager [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 725.308436] env[62066]: ERROR nova.compute.manager [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] for vif in network_info: [ 725.308436] env[62066]: ERROR nova.compute.manager [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 725.308436] env[62066]: ERROR nova.compute.manager [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] return self._sync_wrapper(fn, *args, **kwargs) [ 725.308436] env[62066]: ERROR nova.compute.manager [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 725.308436] env[62066]: ERROR nova.compute.manager [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] self.wait() [ 725.308436] env[62066]: ERROR nova.compute.manager [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 725.308436] env[62066]: ERROR nova.compute.manager [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] self[:] = self._gt.wait() [ 725.308436] env[62066]: ERROR nova.compute.manager [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 725.308436] env[62066]: ERROR nova.compute.manager [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] return self._exit_event.wait() [ 725.308436] env[62066]: ERROR nova.compute.manager [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 725.308436] env[62066]: ERROR nova.compute.manager [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] result = hub.switch() [ 725.308436] env[62066]: ERROR nova.compute.manager [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 725.308436] env[62066]: ERROR nova.compute.manager [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] return self.greenlet.switch() [ 725.308902] env[62066]: ERROR nova.compute.manager [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 725.308902] env[62066]: ERROR nova.compute.manager [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] result = function(*args, **kwargs) [ 725.308902] env[62066]: ERROR nova.compute.manager [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 725.308902] env[62066]: ERROR nova.compute.manager [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] return func(*args, **kwargs) [ 725.308902] env[62066]: ERROR nova.compute.manager [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 725.308902] env[62066]: ERROR nova.compute.manager [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] raise e [ 725.308902] env[62066]: ERROR nova.compute.manager [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 725.308902] env[62066]: ERROR nova.compute.manager [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] nwinfo = self.network_api.allocate_for_instance( [ 725.308902] env[62066]: ERROR nova.compute.manager [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 725.308902] env[62066]: ERROR nova.compute.manager [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] created_port_ids = self._update_ports_for_instance( [ 725.308902] env[62066]: ERROR nova.compute.manager [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 725.308902] env[62066]: ERROR nova.compute.manager [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] with excutils.save_and_reraise_exception(): [ 725.308902] env[62066]: ERROR nova.compute.manager [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 725.309350] env[62066]: ERROR nova.compute.manager [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] self.force_reraise() [ 725.309350] env[62066]: ERROR nova.compute.manager [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 725.309350] env[62066]: ERROR nova.compute.manager [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] raise self.value [ 725.309350] env[62066]: ERROR nova.compute.manager [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 725.309350] env[62066]: ERROR nova.compute.manager [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] updated_port = self._update_port( [ 725.309350] env[62066]: ERROR nova.compute.manager [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 725.309350] env[62066]: ERROR nova.compute.manager [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] _ensure_no_port_binding_failure(port) [ 725.309350] env[62066]: ERROR nova.compute.manager [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 725.309350] env[62066]: ERROR nova.compute.manager [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] raise exception.PortBindingFailed(port_id=port['id']) [ 725.309350] env[62066]: ERROR nova.compute.manager [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] nova.exception.PortBindingFailed: Binding failed for port 11527383-8bd1-4760-9702-997440e384c8, please check neutron logs for more information. [ 725.309350] env[62066]: ERROR nova.compute.manager [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] [ 725.309733] env[62066]: DEBUG nova.compute.utils [None req-84147b21-d606-4daf-b0b9-f581072be1cb tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] Binding failed for port 11527383-8bd1-4760-9702-997440e384c8, please check neutron logs for more information. {{(pid=62066) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 725.310415] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a4d93091-69b6-4b35-887c-e7b927c2300a tempest-ServerMetadataTestJSON-1590103242 tempest-ServerMetadataTestJSON-1590103242-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.199s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 725.312041] env[62066]: INFO nova.compute.claims [None req-a4d93091-69b6-4b35-887c-e7b927c2300a tempest-ServerMetadataTestJSON-1590103242 tempest-ServerMetadataTestJSON-1590103242-project-member] [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 725.315499] env[62066]: DEBUG nova.compute.manager [None req-84147b21-d606-4daf-b0b9-f581072be1cb tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] Build of instance 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3 was re-scheduled: Binding failed for port 11527383-8bd1-4760-9702-997440e384c8, please check neutron logs for more information. {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 725.315942] env[62066]: DEBUG nova.compute.manager [None req-84147b21-d606-4daf-b0b9-f581072be1cb tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] Unplugging VIFs for instance {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 725.316178] env[62066]: DEBUG oslo_concurrency.lockutils [None req-84147b21-d606-4daf-b0b9-f581072be1cb tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] Acquiring lock "refresh_cache-2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 725.316324] env[62066]: DEBUG oslo_concurrency.lockutils [None req-84147b21-d606-4daf-b0b9-f581072be1cb tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] Acquired lock "refresh_cache-2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 725.316480] env[62066]: DEBUG nova.network.neutron [None req-84147b21-d606-4daf-b0b9-f581072be1cb tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 725.325587] env[62066]: ERROR nova.compute.manager [None req-7d1d419b-6e61-4f85-9698-fe0c6e4f3988 tempest-AttachInterfacesV270Test-236309384 tempest-AttachInterfacesV270Test-236309384-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port a3e86a4a-9e56-428c-a598-e42c247ab0e2, please check neutron logs for more information. [ 725.325587] env[62066]: ERROR nova.compute.manager Traceback (most recent call last): [ 725.325587] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 725.325587] env[62066]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 725.325587] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 725.325587] env[62066]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 725.325587] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 725.325587] env[62066]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 725.325587] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 725.325587] env[62066]: ERROR nova.compute.manager self.force_reraise() [ 725.325587] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 725.325587] env[62066]: ERROR nova.compute.manager raise self.value [ 725.325587] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 725.325587] env[62066]: ERROR nova.compute.manager updated_port = self._update_port( [ 725.325587] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 725.325587] env[62066]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 725.326191] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 725.326191] env[62066]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 725.326191] env[62066]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port a3e86a4a-9e56-428c-a598-e42c247ab0e2, please check neutron logs for more information. [ 725.326191] env[62066]: ERROR nova.compute.manager [ 725.326191] env[62066]: Traceback (most recent call last): [ 725.326191] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 725.326191] env[62066]: listener.cb(fileno) [ 725.326191] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 725.326191] env[62066]: result = function(*args, **kwargs) [ 725.326191] env[62066]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 725.326191] env[62066]: return func(*args, **kwargs) [ 725.326191] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 725.326191] env[62066]: raise e [ 725.326191] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 725.326191] env[62066]: nwinfo = self.network_api.allocate_for_instance( [ 725.326191] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 725.326191] env[62066]: created_port_ids = self._update_ports_for_instance( [ 725.326191] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 725.326191] env[62066]: with excutils.save_and_reraise_exception(): [ 725.326191] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 725.326191] env[62066]: self.force_reraise() [ 725.326191] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 725.326191] env[62066]: raise self.value [ 725.326191] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 725.326191] env[62066]: updated_port = self._update_port( [ 725.326191] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 725.326191] env[62066]: _ensure_no_port_binding_failure(port) [ 725.326191] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 725.326191] env[62066]: raise exception.PortBindingFailed(port_id=port['id']) [ 725.327052] env[62066]: nova.exception.PortBindingFailed: Binding failed for port a3e86a4a-9e56-428c-a598-e42c247ab0e2, please check neutron logs for more information. [ 725.327052] env[62066]: Removing descriptor: 15 [ 725.346708] env[62066]: DEBUG nova.compute.manager [None req-7d1d419b-6e61-4f85-9698-fe0c6e4f3988 tempest-AttachInterfacesV270Test-236309384 tempest-AttachInterfacesV270Test-236309384-project-member] [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 725.370955] env[62066]: DEBUG nova.virt.hardware [None req-7d1d419b-6e61-4f85-9698-fe0c6e4f3988 tempest-AttachInterfacesV270Test-236309384 tempest-AttachInterfacesV270Test-236309384-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-23T13:40:41Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-23T13:40:23Z,direct_url=,disk_format='vmdk',id=50ff584c-3b50-4395-af07-3e66769bc9f7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='eb52f7069a374c61ae946f052007c6d9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-23T13:40:24Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 725.371217] env[62066]: DEBUG nova.virt.hardware [None req-7d1d419b-6e61-4f85-9698-fe0c6e4f3988 tempest-AttachInterfacesV270Test-236309384 tempest-AttachInterfacesV270Test-236309384-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 725.371373] env[62066]: DEBUG nova.virt.hardware [None req-7d1d419b-6e61-4f85-9698-fe0c6e4f3988 tempest-AttachInterfacesV270Test-236309384 tempest-AttachInterfacesV270Test-236309384-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 725.371554] env[62066]: DEBUG nova.virt.hardware [None req-7d1d419b-6e61-4f85-9698-fe0c6e4f3988 tempest-AttachInterfacesV270Test-236309384 tempest-AttachInterfacesV270Test-236309384-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 725.372018] env[62066]: DEBUG nova.virt.hardware [None req-7d1d419b-6e61-4f85-9698-fe0c6e4f3988 tempest-AttachInterfacesV270Test-236309384 tempest-AttachInterfacesV270Test-236309384-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 725.372018] env[62066]: DEBUG nova.virt.hardware [None req-7d1d419b-6e61-4f85-9698-fe0c6e4f3988 tempest-AttachInterfacesV270Test-236309384 tempest-AttachInterfacesV270Test-236309384-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 725.372168] env[62066]: DEBUG nova.virt.hardware [None req-7d1d419b-6e61-4f85-9698-fe0c6e4f3988 tempest-AttachInterfacesV270Test-236309384 tempest-AttachInterfacesV270Test-236309384-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 725.372334] env[62066]: DEBUG nova.virt.hardware [None req-7d1d419b-6e61-4f85-9698-fe0c6e4f3988 tempest-AttachInterfacesV270Test-236309384 tempest-AttachInterfacesV270Test-236309384-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 725.372498] env[62066]: DEBUG nova.virt.hardware [None req-7d1d419b-6e61-4f85-9698-fe0c6e4f3988 tempest-AttachInterfacesV270Test-236309384 tempest-AttachInterfacesV270Test-236309384-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 725.372657] env[62066]: DEBUG nova.virt.hardware [None req-7d1d419b-6e61-4f85-9698-fe0c6e4f3988 tempest-AttachInterfacesV270Test-236309384 tempest-AttachInterfacesV270Test-236309384-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 725.372823] env[62066]: DEBUG nova.virt.hardware [None req-7d1d419b-6e61-4f85-9698-fe0c6e4f3988 tempest-AttachInterfacesV270Test-236309384 tempest-AttachInterfacesV270Test-236309384-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 725.373663] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f105ad25-eb4c-400c-9b51-3971779df93a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.381649] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19c35e92-7009-4008-95cf-27da634e7fe0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.395741] env[62066]: ERROR nova.compute.manager [None req-7d1d419b-6e61-4f85-9698-fe0c6e4f3988 tempest-AttachInterfacesV270Test-236309384 tempest-AttachInterfacesV270Test-236309384-project-member] [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port a3e86a4a-9e56-428c-a598-e42c247ab0e2, please check neutron logs for more information. [ 725.395741] env[62066]: ERROR nova.compute.manager [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] Traceback (most recent call last): [ 725.395741] env[62066]: ERROR nova.compute.manager [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 725.395741] env[62066]: ERROR nova.compute.manager [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] yield resources [ 725.395741] env[62066]: ERROR nova.compute.manager [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 725.395741] env[62066]: ERROR nova.compute.manager [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] self.driver.spawn(context, instance, image_meta, [ 725.395741] env[62066]: ERROR nova.compute.manager [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 725.395741] env[62066]: ERROR nova.compute.manager [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 725.395741] env[62066]: ERROR nova.compute.manager [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 725.395741] env[62066]: ERROR nova.compute.manager [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] vm_ref = self.build_virtual_machine(instance, [ 725.395741] env[62066]: ERROR nova.compute.manager [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 725.396187] env[62066]: ERROR nova.compute.manager [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] vif_infos = vmwarevif.get_vif_info(self._session, [ 725.396187] env[62066]: ERROR nova.compute.manager [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 725.396187] env[62066]: ERROR nova.compute.manager [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] for vif in network_info: [ 725.396187] env[62066]: ERROR nova.compute.manager [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 725.396187] env[62066]: ERROR nova.compute.manager [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] return self._sync_wrapper(fn, *args, **kwargs) [ 725.396187] env[62066]: ERROR nova.compute.manager [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 725.396187] env[62066]: ERROR nova.compute.manager [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] self.wait() [ 725.396187] env[62066]: ERROR nova.compute.manager [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 725.396187] env[62066]: ERROR nova.compute.manager [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] self[:] = self._gt.wait() [ 725.396187] env[62066]: ERROR nova.compute.manager [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 725.396187] env[62066]: ERROR nova.compute.manager [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] return self._exit_event.wait() [ 725.396187] env[62066]: ERROR nova.compute.manager [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 725.396187] env[62066]: ERROR nova.compute.manager [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] current.throw(*self._exc) [ 725.396637] env[62066]: ERROR nova.compute.manager [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 725.396637] env[62066]: ERROR nova.compute.manager [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] result = function(*args, **kwargs) [ 725.396637] env[62066]: ERROR nova.compute.manager [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 725.396637] env[62066]: ERROR nova.compute.manager [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] return func(*args, **kwargs) [ 725.396637] env[62066]: ERROR nova.compute.manager [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 725.396637] env[62066]: ERROR nova.compute.manager [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] raise e [ 725.396637] env[62066]: ERROR nova.compute.manager [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 725.396637] env[62066]: ERROR nova.compute.manager [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] nwinfo = self.network_api.allocate_for_instance( [ 725.396637] env[62066]: ERROR nova.compute.manager [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 725.396637] env[62066]: ERROR nova.compute.manager [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] created_port_ids = self._update_ports_for_instance( [ 725.396637] env[62066]: ERROR nova.compute.manager [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 725.396637] env[62066]: ERROR nova.compute.manager [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] with excutils.save_and_reraise_exception(): [ 725.396637] env[62066]: ERROR nova.compute.manager [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 725.397088] env[62066]: ERROR nova.compute.manager [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] self.force_reraise() [ 725.397088] env[62066]: ERROR nova.compute.manager [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 725.397088] env[62066]: ERROR nova.compute.manager [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] raise self.value [ 725.397088] env[62066]: ERROR nova.compute.manager [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 725.397088] env[62066]: ERROR nova.compute.manager [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] updated_port = self._update_port( [ 725.397088] env[62066]: ERROR nova.compute.manager [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 725.397088] env[62066]: ERROR nova.compute.manager [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] _ensure_no_port_binding_failure(port) [ 725.397088] env[62066]: ERROR nova.compute.manager [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 725.397088] env[62066]: ERROR nova.compute.manager [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] raise exception.PortBindingFailed(port_id=port['id']) [ 725.397088] env[62066]: ERROR nova.compute.manager [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] nova.exception.PortBindingFailed: Binding failed for port a3e86a4a-9e56-428c-a598-e42c247ab0e2, please check neutron logs for more information. [ 725.397088] env[62066]: ERROR nova.compute.manager [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] [ 725.397088] env[62066]: INFO nova.compute.manager [None req-7d1d419b-6e61-4f85-9698-fe0c6e4f3988 tempest-AttachInterfacesV270Test-236309384 tempest-AttachInterfacesV270Test-236309384-project-member] [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] Terminating instance [ 725.398020] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7d1d419b-6e61-4f85-9698-fe0c6e4f3988 tempest-AttachInterfacesV270Test-236309384 tempest-AttachInterfacesV270Test-236309384-project-member] Acquiring lock "refresh_cache-c908a3ed-88a6-4e79-8c27-2e16bf2bdca4" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 725.398183] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7d1d419b-6e61-4f85-9698-fe0c6e4f3988 tempest-AttachInterfacesV270Test-236309384 tempest-AttachInterfacesV270Test-236309384-project-member] Acquired lock "refresh_cache-c908a3ed-88a6-4e79-8c27-2e16bf2bdca4" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 725.398348] env[62066]: DEBUG nova.network.neutron [None req-7d1d419b-6e61-4f85-9698-fe0c6e4f3988 tempest-AttachInterfacesV270Test-236309384 tempest-AttachInterfacesV270Test-236309384-project-member] [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 725.478127] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9e052131-3359-4888-9d78-89440b639dcc tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Acquiring lock "8ff6fbb9-c90f-498d-9a85-d220a8c2f794" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 725.478290] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9e052131-3359-4888-9d78-89440b639dcc tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Lock "8ff6fbb9-c90f-498d-9a85-d220a8c2f794" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 725.541673] env[62066]: DEBUG oslo_vmware.api [None req-fbd85018-c9a9-4c9f-87ff-96cad2b995f9 tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]520968ae-459f-d62a-c205-59796d26e557, 'name': SearchDatastore_Task, 'duration_secs': 0.007991} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 725.542974] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-11b4972a-91f1-4948-96da-51fa3b0b81d3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.548077] env[62066]: DEBUG oslo_vmware.api [None req-fbd85018-c9a9-4c9f-87ff-96cad2b995f9 tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Waiting for the task: (returnval){ [ 725.548077] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]52b8ae5e-2419-b200-a137-fec680223a97" [ 725.548077] env[62066]: _type = "Task" [ 725.548077] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 725.557014] env[62066]: DEBUG oslo_vmware.api [None req-fbd85018-c9a9-4c9f-87ff-96cad2b995f9 tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52b8ae5e-2419-b200-a137-fec680223a97, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 725.769928] env[62066]: INFO nova.compute.manager [-] [instance: 3624ca43-9ee3-418a-9205-debb699295df] Took 1.03 seconds to deallocate network for instance. [ 725.772379] env[62066]: DEBUG nova.compute.claims [None req-5bf8bcd1-f041-4648-9b3c-4febff568fa9 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] [instance: 3624ca43-9ee3-418a-9205-debb699295df] Aborting claim: {{(pid=62066) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 725.772562] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5bf8bcd1-f041-4648-9b3c-4febff568fa9 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 725.835044] env[62066]: DEBUG nova.network.neutron [None req-84147b21-d606-4daf-b0b9-f581072be1cb tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 725.912056] env[62066]: DEBUG nova.network.neutron [None req-84147b21-d606-4daf-b0b9-f581072be1cb tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 725.918982] env[62066]: DEBUG nova.network.neutron [None req-7d1d419b-6e61-4f85-9698-fe0c6e4f3988 tempest-AttachInterfacesV270Test-236309384 tempest-AttachInterfacesV270Test-236309384-project-member] [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 726.009292] env[62066]: DEBUG nova.network.neutron [None req-7d1d419b-6e61-4f85-9698-fe0c6e4f3988 tempest-AttachInterfacesV270Test-236309384 tempest-AttachInterfacesV270Test-236309384-project-member] [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 726.058370] env[62066]: DEBUG oslo_vmware.api [None req-fbd85018-c9a9-4c9f-87ff-96cad2b995f9 tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52b8ae5e-2419-b200-a137-fec680223a97, 'name': SearchDatastore_Task, 'duration_secs': 0.008577} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 726.058692] env[62066]: DEBUG oslo_concurrency.lockutils [None req-fbd85018-c9a9-4c9f-87ff-96cad2b995f9 tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Releasing lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 726.058925] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-fbd85018-c9a9-4c9f-87ff-96cad2b995f9 tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk to [datastore2] 0ff2f9f4-6d85-4465-8e0a-fb39f2a25c9b/0ff2f9f4-6d85-4465-8e0a-fb39f2a25c9b.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 726.059193] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3d40a64e-5e86-4269-96d7-c89b42d501ed {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.065637] env[62066]: DEBUG oslo_vmware.api [None req-fbd85018-c9a9-4c9f-87ff-96cad2b995f9 tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Waiting for the task: (returnval){ [ 726.065637] env[62066]: value = "task-1340687" [ 726.065637] env[62066]: _type = "Task" [ 726.065637] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 726.074046] env[62066]: DEBUG oslo_vmware.api [None req-fbd85018-c9a9-4c9f-87ff-96cad2b995f9 tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Task: {'id': task-1340687, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 726.418030] env[62066]: DEBUG oslo_concurrency.lockutils [None req-84147b21-d606-4daf-b0b9-f581072be1cb tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] Releasing lock "refresh_cache-2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 726.418030] env[62066]: DEBUG nova.compute.manager [None req-84147b21-d606-4daf-b0b9-f581072be1cb tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 726.418030] env[62066]: DEBUG nova.compute.manager [None req-84147b21-d606-4daf-b0b9-f581072be1cb tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 726.418030] env[62066]: DEBUG nova.network.neutron [None req-84147b21-d606-4daf-b0b9-f581072be1cb tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 726.443291] env[62066]: DEBUG nova.network.neutron [None req-84147b21-d606-4daf-b0b9-f581072be1cb tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 726.477256] env[62066]: DEBUG nova.compute.manager [req-44d70fd0-b76c-4314-9624-e7dd608dd562 req-de06a757-e91f-42da-bf34-e3393b1a8938 service nova] [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] Received event network-changed-a3e86a4a-9e56-428c-a598-e42c247ab0e2 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 726.477535] env[62066]: DEBUG nova.compute.manager [req-44d70fd0-b76c-4314-9624-e7dd608dd562 req-de06a757-e91f-42da-bf34-e3393b1a8938 service nova] [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] Refreshing instance network info cache due to event network-changed-a3e86a4a-9e56-428c-a598-e42c247ab0e2. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 726.477676] env[62066]: DEBUG oslo_concurrency.lockutils [req-44d70fd0-b76c-4314-9624-e7dd608dd562 req-de06a757-e91f-42da-bf34-e3393b1a8938 service nova] Acquiring lock "refresh_cache-c908a3ed-88a6-4e79-8c27-2e16bf2bdca4" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 726.511621] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7d1d419b-6e61-4f85-9698-fe0c6e4f3988 tempest-AttachInterfacesV270Test-236309384 tempest-AttachInterfacesV270Test-236309384-project-member] Releasing lock "refresh_cache-c908a3ed-88a6-4e79-8c27-2e16bf2bdca4" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 726.512070] env[62066]: DEBUG nova.compute.manager [None req-7d1d419b-6e61-4f85-9698-fe0c6e4f3988 tempest-AttachInterfacesV270Test-236309384 tempest-AttachInterfacesV270Test-236309384-project-member] [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 726.512267] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-7d1d419b-6e61-4f85-9698-fe0c6e4f3988 tempest-AttachInterfacesV270Test-236309384 tempest-AttachInterfacesV270Test-236309384-project-member] [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 726.512570] env[62066]: DEBUG oslo_concurrency.lockutils [req-44d70fd0-b76c-4314-9624-e7dd608dd562 req-de06a757-e91f-42da-bf34-e3393b1a8938 service nova] Acquired lock "refresh_cache-c908a3ed-88a6-4e79-8c27-2e16bf2bdca4" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 726.512743] env[62066]: DEBUG nova.network.neutron [req-44d70fd0-b76c-4314-9624-e7dd608dd562 req-de06a757-e91f-42da-bf34-e3393b1a8938 service nova] [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] Refreshing network info cache for port a3e86a4a-9e56-428c-a598-e42c247ab0e2 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 726.514170] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ffc0469b-32ed-4f02-adc6-4a4d2e6183c2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.527528] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5795faa-f721-405a-ae70-69951dd0643a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.552763] env[62066]: WARNING nova.virt.vmwareapi.vmops [None req-7d1d419b-6e61-4f85-9698-fe0c6e4f3988 tempest-AttachInterfacesV270Test-236309384 tempest-AttachInterfacesV270Test-236309384-project-member] [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance c908a3ed-88a6-4e79-8c27-2e16bf2bdca4 could not be found. [ 726.553189] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-7d1d419b-6e61-4f85-9698-fe0c6e4f3988 tempest-AttachInterfacesV270Test-236309384 tempest-AttachInterfacesV270Test-236309384-project-member] [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 726.553365] env[62066]: INFO nova.compute.manager [None req-7d1d419b-6e61-4f85-9698-fe0c6e4f3988 tempest-AttachInterfacesV270Test-236309384 tempest-AttachInterfacesV270Test-236309384-project-member] [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] Took 0.04 seconds to destroy the instance on the hypervisor. [ 726.553550] env[62066]: DEBUG oslo.service.loopingcall [None req-7d1d419b-6e61-4f85-9698-fe0c6e4f3988 tempest-AttachInterfacesV270Test-236309384 tempest-AttachInterfacesV270Test-236309384-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 726.556186] env[62066]: DEBUG nova.compute.manager [-] [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 726.556282] env[62066]: DEBUG nova.network.neutron [-] [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 726.571613] env[62066]: DEBUG nova.network.neutron [-] [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 726.577261] env[62066]: DEBUG oslo_vmware.api [None req-fbd85018-c9a9-4c9f-87ff-96cad2b995f9 tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Task: {'id': task-1340687, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.473721} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 726.577504] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-fbd85018-c9a9-4c9f-87ff-96cad2b995f9 tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk to [datastore2] 0ff2f9f4-6d85-4465-8e0a-fb39f2a25c9b/0ff2f9f4-6d85-4465-8e0a-fb39f2a25c9b.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 726.577710] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-fbd85018-c9a9-4c9f-87ff-96cad2b995f9 tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] [instance: 0ff2f9f4-6d85-4465-8e0a-fb39f2a25c9b] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 726.578395] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ec512b5a-4b2b-43d1-8a3e-56542a26444b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.586150] env[62066]: DEBUG oslo_vmware.api [None req-fbd85018-c9a9-4c9f-87ff-96cad2b995f9 tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Waiting for the task: (returnval){ [ 726.586150] env[62066]: value = "task-1340688" [ 726.586150] env[62066]: _type = "Task" [ 726.586150] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 726.595331] env[62066]: DEBUG oslo_vmware.api [None req-fbd85018-c9a9-4c9f-87ff-96cad2b995f9 tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Task: {'id': task-1340688, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 726.701363] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-186b10bd-eb89-4bc1-8640-26e359168322 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.710207] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0db1c11-7aab-496b-a822-67d99b812c17 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.738717] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed1cfa5e-ddd6-4404-b6f3-bb8ed9810e28 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.745377] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e82d7215-b1f7-4035-8d89-fab7e95ab257 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.758857] env[62066]: DEBUG nova.compute.provider_tree [None req-a4d93091-69b6-4b35-887c-e7b927c2300a tempest-ServerMetadataTestJSON-1590103242 tempest-ServerMetadataTestJSON-1590103242-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 726.946047] env[62066]: DEBUG nova.network.neutron [None req-84147b21-d606-4daf-b0b9-f581072be1cb tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 727.046167] env[62066]: DEBUG nova.network.neutron [req-44d70fd0-b76c-4314-9624-e7dd608dd562 req-de06a757-e91f-42da-bf34-e3393b1a8938 service nova] [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 727.075028] env[62066]: DEBUG nova.network.neutron [-] [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 727.095743] env[62066]: DEBUG oslo_vmware.api [None req-fbd85018-c9a9-4c9f-87ff-96cad2b995f9 tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Task: {'id': task-1340688, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070932} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 727.096133] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-fbd85018-c9a9-4c9f-87ff-96cad2b995f9 tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] [instance: 0ff2f9f4-6d85-4465-8e0a-fb39f2a25c9b] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 727.096913] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3de2b5dd-45f0-47ce-bb7b-bce8fe82cec3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.117403] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-fbd85018-c9a9-4c9f-87ff-96cad2b995f9 tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] [instance: 0ff2f9f4-6d85-4465-8e0a-fb39f2a25c9b] Reconfiguring VM instance instance-0000002d to attach disk [datastore2] 0ff2f9f4-6d85-4465-8e0a-fb39f2a25c9b/0ff2f9f4-6d85-4465-8e0a-fb39f2a25c9b.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 727.118769] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0de9420d-f298-46ac-b17f-86c548c64d96 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.137548] env[62066]: DEBUG oslo_vmware.api [None req-fbd85018-c9a9-4c9f-87ff-96cad2b995f9 tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Waiting for the task: (returnval){ [ 727.137548] env[62066]: value = "task-1340689" [ 727.137548] env[62066]: _type = "Task" [ 727.137548] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 727.146902] env[62066]: DEBUG oslo_vmware.api [None req-fbd85018-c9a9-4c9f-87ff-96cad2b995f9 tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Task: {'id': task-1340689, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 727.182419] env[62066]: DEBUG nova.network.neutron [req-44d70fd0-b76c-4314-9624-e7dd608dd562 req-de06a757-e91f-42da-bf34-e3393b1a8938 service nova] [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 727.261930] env[62066]: DEBUG nova.scheduler.client.report [None req-a4d93091-69b6-4b35-887c-e7b927c2300a tempest-ServerMetadataTestJSON-1590103242 tempest-ServerMetadataTestJSON-1590103242-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 727.448301] env[62066]: INFO nova.compute.manager [None req-84147b21-d606-4daf-b0b9-f581072be1cb tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] [instance: 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3] Took 1.03 seconds to deallocate network for instance. [ 727.576754] env[62066]: INFO nova.compute.manager [-] [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] Took 1.02 seconds to deallocate network for instance. [ 727.579039] env[62066]: DEBUG nova.compute.claims [None req-7d1d419b-6e61-4f85-9698-fe0c6e4f3988 tempest-AttachInterfacesV270Test-236309384 tempest-AttachInterfacesV270Test-236309384-project-member] [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] Aborting claim: {{(pid=62066) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 727.579218] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7d1d419b-6e61-4f85-9698-fe0c6e4f3988 tempest-AttachInterfacesV270Test-236309384 tempest-AttachInterfacesV270Test-236309384-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 727.646281] env[62066]: DEBUG oslo_vmware.api [None req-fbd85018-c9a9-4c9f-87ff-96cad2b995f9 tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Task: {'id': task-1340689, 'name': ReconfigVM_Task, 'duration_secs': 0.259673} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 727.646537] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-fbd85018-c9a9-4c9f-87ff-96cad2b995f9 tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] [instance: 0ff2f9f4-6d85-4465-8e0a-fb39f2a25c9b] Reconfigured VM instance instance-0000002d to attach disk [datastore2] 0ff2f9f4-6d85-4465-8e0a-fb39f2a25c9b/0ff2f9f4-6d85-4465-8e0a-fb39f2a25c9b.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 727.647125] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-848b2359-d82d-440f-b5de-b8e47cab339e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.653423] env[62066]: DEBUG oslo_vmware.api [None req-fbd85018-c9a9-4c9f-87ff-96cad2b995f9 tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Waiting for the task: (returnval){ [ 727.653423] env[62066]: value = "task-1340690" [ 727.653423] env[62066]: _type = "Task" [ 727.653423] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 727.660962] env[62066]: DEBUG oslo_vmware.api [None req-fbd85018-c9a9-4c9f-87ff-96cad2b995f9 tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Task: {'id': task-1340690, 'name': Rename_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 727.684534] env[62066]: DEBUG oslo_concurrency.lockutils [req-44d70fd0-b76c-4314-9624-e7dd608dd562 req-de06a757-e91f-42da-bf34-e3393b1a8938 service nova] Releasing lock "refresh_cache-c908a3ed-88a6-4e79-8c27-2e16bf2bdca4" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 727.684841] env[62066]: DEBUG nova.compute.manager [req-44d70fd0-b76c-4314-9624-e7dd608dd562 req-de06a757-e91f-42da-bf34-e3393b1a8938 service nova] [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] Received event network-vif-deleted-a3e86a4a-9e56-428c-a598-e42c247ab0e2 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 727.767586] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a4d93091-69b6-4b35-887c-e7b927c2300a tempest-ServerMetadataTestJSON-1590103242 tempest-ServerMetadataTestJSON-1590103242-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.457s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 727.768134] env[62066]: DEBUG nova.compute.manager [None req-a4d93091-69b6-4b35-887c-e7b927c2300a tempest-ServerMetadataTestJSON-1590103242 tempest-ServerMetadataTestJSON-1590103242-project-member] [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 727.770832] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e9793203-98b3-4ca0-8a49-79767b1f52f7 tempest-ServersTestBootFromVolume-1535558002 tempest-ServersTestBootFromVolume-1535558002-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.801s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 727.772463] env[62066]: INFO nova.compute.claims [None req-e9793203-98b3-4ca0-8a49-79767b1f52f7 tempest-ServersTestBootFromVolume-1535558002 tempest-ServersTestBootFromVolume-1535558002-project-member] [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 728.163460] env[62066]: DEBUG oslo_vmware.api [None req-fbd85018-c9a9-4c9f-87ff-96cad2b995f9 tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Task: {'id': task-1340690, 'name': Rename_Task, 'duration_secs': 0.125645} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 728.163727] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-fbd85018-c9a9-4c9f-87ff-96cad2b995f9 tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] [instance: 0ff2f9f4-6d85-4465-8e0a-fb39f2a25c9b] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 728.163966] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e219eb5c-aefb-44c4-a33f-6ada0fd8177f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.171019] env[62066]: DEBUG oslo_vmware.api [None req-fbd85018-c9a9-4c9f-87ff-96cad2b995f9 tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Waiting for the task: (returnval){ [ 728.171019] env[62066]: value = "task-1340691" [ 728.171019] env[62066]: _type = "Task" [ 728.171019] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 728.177738] env[62066]: DEBUG oslo_vmware.api [None req-fbd85018-c9a9-4c9f-87ff-96cad2b995f9 tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Task: {'id': task-1340691, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 728.276946] env[62066]: DEBUG nova.compute.utils [None req-a4d93091-69b6-4b35-887c-e7b927c2300a tempest-ServerMetadataTestJSON-1590103242 tempest-ServerMetadataTestJSON-1590103242-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 728.280579] env[62066]: DEBUG nova.compute.manager [None req-a4d93091-69b6-4b35-887c-e7b927c2300a tempest-ServerMetadataTestJSON-1590103242 tempest-ServerMetadataTestJSON-1590103242-project-member] [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 728.280779] env[62066]: DEBUG nova.network.neutron [None req-a4d93091-69b6-4b35-887c-e7b927c2300a tempest-ServerMetadataTestJSON-1590103242 tempest-ServerMetadataTestJSON-1590103242-project-member] [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 728.341800] env[62066]: DEBUG nova.policy [None req-a4d93091-69b6-4b35-887c-e7b927c2300a tempest-ServerMetadataTestJSON-1590103242 tempest-ServerMetadataTestJSON-1590103242-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1fd6dc70febb41e38d4d6d13160fd3c2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e5d045f27f034d0fa2ef2b1ea879118e', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 728.481046] env[62066]: INFO nova.scheduler.client.report [None req-84147b21-d606-4daf-b0b9-f581072be1cb tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] Deleted allocations for instance 2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3 [ 728.680978] env[62066]: DEBUG oslo_vmware.api [None req-fbd85018-c9a9-4c9f-87ff-96cad2b995f9 tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Task: {'id': task-1340691, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 728.689182] env[62066]: DEBUG nova.network.neutron [None req-a4d93091-69b6-4b35-887c-e7b927c2300a tempest-ServerMetadataTestJSON-1590103242 tempest-ServerMetadataTestJSON-1590103242-project-member] [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] Successfully created port: 91b620d9-65dc-4e47-87b9-3747a6897b6f {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 728.787020] env[62066]: DEBUG nova.compute.manager [None req-a4d93091-69b6-4b35-887c-e7b927c2300a tempest-ServerMetadataTestJSON-1590103242 tempest-ServerMetadataTestJSON-1590103242-project-member] [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 728.988772] env[62066]: DEBUG oslo_concurrency.lockutils [None req-84147b21-d606-4daf-b0b9-f581072be1cb tempest-VolumesAdminNegativeTest-1848663090 tempest-VolumesAdminNegativeTest-1848663090-project-member] Lock "2e0f2d4e-aa3d-4b1a-8c8c-9338f75a07e3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 167.014s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 729.087527] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d99d545-8548-4735-b95a-6dd2bd071836 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.094865] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5937267-3026-44db-b63e-651d854c77c6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.124483] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4105ea45-1bb0-4557-bcc8-0723f2db6ebe {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.132019] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35b81e61-f41f-4af1-a52e-d7f39a582eb6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.146097] env[62066]: DEBUG nova.compute.provider_tree [None req-e9793203-98b3-4ca0-8a49-79767b1f52f7 tempest-ServersTestBootFromVolume-1535558002 tempest-ServersTestBootFromVolume-1535558002-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 729.180216] env[62066]: DEBUG oslo_vmware.api [None req-fbd85018-c9a9-4c9f-87ff-96cad2b995f9 tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Task: {'id': task-1340691, 'name': PowerOnVM_Task, 'duration_secs': 0.62168} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 729.180493] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-fbd85018-c9a9-4c9f-87ff-96cad2b995f9 tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] [instance: 0ff2f9f4-6d85-4465-8e0a-fb39f2a25c9b] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 729.180669] env[62066]: DEBUG nova.compute.manager [None req-fbd85018-c9a9-4c9f-87ff-96cad2b995f9 tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] [instance: 0ff2f9f4-6d85-4465-8e0a-fb39f2a25c9b] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 729.181451] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f669ce06-e888-437d-a244-86547fa03ea1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.493312] env[62066]: DEBUG nova.compute.manager [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 729.649751] env[62066]: DEBUG nova.scheduler.client.report [None req-e9793203-98b3-4ca0-8a49-79767b1f52f7 tempest-ServersTestBootFromVolume-1535558002 tempest-ServersTestBootFromVolume-1535558002-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 729.697619] env[62066]: DEBUG oslo_concurrency.lockutils [None req-fbd85018-c9a9-4c9f-87ff-96cad2b995f9 tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 729.738400] env[62066]: DEBUG nova.compute.manager [req-3c9f96e0-2bb4-446e-8736-aac58fb68f24 req-94e34b36-7845-4be2-a3f9-b5a753c4719b service nova] [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] Received event network-changed-91b620d9-65dc-4e47-87b9-3747a6897b6f {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 729.738400] env[62066]: DEBUG nova.compute.manager [req-3c9f96e0-2bb4-446e-8736-aac58fb68f24 req-94e34b36-7845-4be2-a3f9-b5a753c4719b service nova] [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] Refreshing instance network info cache due to event network-changed-91b620d9-65dc-4e47-87b9-3747a6897b6f. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 729.738400] env[62066]: DEBUG oslo_concurrency.lockutils [req-3c9f96e0-2bb4-446e-8736-aac58fb68f24 req-94e34b36-7845-4be2-a3f9-b5a753c4719b service nova] Acquiring lock "refresh_cache-0ed07533-9e79-416d-a495-a8e20b97e108" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 729.739776] env[62066]: DEBUG oslo_concurrency.lockutils [req-3c9f96e0-2bb4-446e-8736-aac58fb68f24 req-94e34b36-7845-4be2-a3f9-b5a753c4719b service nova] Acquired lock "refresh_cache-0ed07533-9e79-416d-a495-a8e20b97e108" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 729.742466] env[62066]: DEBUG nova.network.neutron [req-3c9f96e0-2bb4-446e-8736-aac58fb68f24 req-94e34b36-7845-4be2-a3f9-b5a753c4719b service nova] [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] Refreshing network info cache for port 91b620d9-65dc-4e47-87b9-3747a6897b6f {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 729.800416] env[62066]: DEBUG nova.compute.manager [None req-a4d93091-69b6-4b35-887c-e7b927c2300a tempest-ServerMetadataTestJSON-1590103242 tempest-ServerMetadataTestJSON-1590103242-project-member] [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 729.838236] env[62066]: DEBUG nova.virt.hardware [None req-a4d93091-69b6-4b35-887c-e7b927c2300a tempest-ServerMetadataTestJSON-1590103242 tempest-ServerMetadataTestJSON-1590103242-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-23T13:40:41Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-23T13:40:23Z,direct_url=,disk_format='vmdk',id=50ff584c-3b50-4395-af07-3e66769bc9f7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='eb52f7069a374c61ae946f052007c6d9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-23T13:40:24Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 729.838236] env[62066]: DEBUG nova.virt.hardware [None req-a4d93091-69b6-4b35-887c-e7b927c2300a tempest-ServerMetadataTestJSON-1590103242 tempest-ServerMetadataTestJSON-1590103242-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 729.838236] env[62066]: DEBUG nova.virt.hardware [None req-a4d93091-69b6-4b35-887c-e7b927c2300a tempest-ServerMetadataTestJSON-1590103242 tempest-ServerMetadataTestJSON-1590103242-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 729.838419] env[62066]: DEBUG nova.virt.hardware [None req-a4d93091-69b6-4b35-887c-e7b927c2300a tempest-ServerMetadataTestJSON-1590103242 tempest-ServerMetadataTestJSON-1590103242-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 729.838419] env[62066]: DEBUG nova.virt.hardware [None req-a4d93091-69b6-4b35-887c-e7b927c2300a tempest-ServerMetadataTestJSON-1590103242 tempest-ServerMetadataTestJSON-1590103242-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 729.838419] env[62066]: DEBUG nova.virt.hardware [None req-a4d93091-69b6-4b35-887c-e7b927c2300a tempest-ServerMetadataTestJSON-1590103242 tempest-ServerMetadataTestJSON-1590103242-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 729.838419] env[62066]: DEBUG nova.virt.hardware [None req-a4d93091-69b6-4b35-887c-e7b927c2300a tempest-ServerMetadataTestJSON-1590103242 tempest-ServerMetadataTestJSON-1590103242-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 729.838419] env[62066]: DEBUG nova.virt.hardware [None req-a4d93091-69b6-4b35-887c-e7b927c2300a tempest-ServerMetadataTestJSON-1590103242 tempest-ServerMetadataTestJSON-1590103242-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 729.838648] env[62066]: DEBUG nova.virt.hardware [None req-a4d93091-69b6-4b35-887c-e7b927c2300a tempest-ServerMetadataTestJSON-1590103242 tempest-ServerMetadataTestJSON-1590103242-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 729.839923] env[62066]: DEBUG nova.virt.hardware [None req-a4d93091-69b6-4b35-887c-e7b927c2300a tempest-ServerMetadataTestJSON-1590103242 tempest-ServerMetadataTestJSON-1590103242-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 729.839923] env[62066]: DEBUG nova.virt.hardware [None req-a4d93091-69b6-4b35-887c-e7b927c2300a tempest-ServerMetadataTestJSON-1590103242 tempest-ServerMetadataTestJSON-1590103242-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 729.839923] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea266063-4d98-4927-8ed2-b6d44eacbbb3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.843398] env[62066]: ERROR nova.compute.manager [None req-a4d93091-69b6-4b35-887c-e7b927c2300a tempest-ServerMetadataTestJSON-1590103242 tempest-ServerMetadataTestJSON-1590103242-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 91b620d9-65dc-4e47-87b9-3747a6897b6f, please check neutron logs for more information. [ 729.843398] env[62066]: ERROR nova.compute.manager Traceback (most recent call last): [ 729.843398] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 729.843398] env[62066]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 729.843398] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 729.843398] env[62066]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 729.843398] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 729.843398] env[62066]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 729.843398] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 729.843398] env[62066]: ERROR nova.compute.manager self.force_reraise() [ 729.843398] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 729.843398] env[62066]: ERROR nova.compute.manager raise self.value [ 729.843398] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 729.843398] env[62066]: ERROR nova.compute.manager updated_port = self._update_port( [ 729.843398] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 729.843398] env[62066]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 729.844598] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 729.844598] env[62066]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 729.844598] env[62066]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 91b620d9-65dc-4e47-87b9-3747a6897b6f, please check neutron logs for more information. [ 729.844598] env[62066]: ERROR nova.compute.manager [ 729.844598] env[62066]: Traceback (most recent call last): [ 729.844598] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 729.844598] env[62066]: listener.cb(fileno) [ 729.844598] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 729.844598] env[62066]: result = function(*args, **kwargs) [ 729.844598] env[62066]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 729.844598] env[62066]: return func(*args, **kwargs) [ 729.844598] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 729.844598] env[62066]: raise e [ 729.844598] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 729.844598] env[62066]: nwinfo = self.network_api.allocate_for_instance( [ 729.844598] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 729.844598] env[62066]: created_port_ids = self._update_ports_for_instance( [ 729.844598] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 729.844598] env[62066]: with excutils.save_and_reraise_exception(): [ 729.844598] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 729.844598] env[62066]: self.force_reraise() [ 729.844598] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 729.844598] env[62066]: raise self.value [ 729.844598] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 729.844598] env[62066]: updated_port = self._update_port( [ 729.844598] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 729.844598] env[62066]: _ensure_no_port_binding_failure(port) [ 729.844598] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 729.844598] env[62066]: raise exception.PortBindingFailed(port_id=port['id']) [ 729.845708] env[62066]: nova.exception.PortBindingFailed: Binding failed for port 91b620d9-65dc-4e47-87b9-3747a6897b6f, please check neutron logs for more information. [ 729.845708] env[62066]: Removing descriptor: 20 [ 729.849289] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe8be1e2-6970-4f30-876e-f7194176e3ea {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.862949] env[62066]: ERROR nova.compute.manager [None req-a4d93091-69b6-4b35-887c-e7b927c2300a tempest-ServerMetadataTestJSON-1590103242 tempest-ServerMetadataTestJSON-1590103242-project-member] [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 91b620d9-65dc-4e47-87b9-3747a6897b6f, please check neutron logs for more information. [ 729.862949] env[62066]: ERROR nova.compute.manager [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] Traceback (most recent call last): [ 729.862949] env[62066]: ERROR nova.compute.manager [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 729.862949] env[62066]: ERROR nova.compute.manager [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] yield resources [ 729.862949] env[62066]: ERROR nova.compute.manager [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 729.862949] env[62066]: ERROR nova.compute.manager [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] self.driver.spawn(context, instance, image_meta, [ 729.862949] env[62066]: ERROR nova.compute.manager [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 729.862949] env[62066]: ERROR nova.compute.manager [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] self._vmops.spawn(context, instance, image_meta, injected_files, [ 729.862949] env[62066]: ERROR nova.compute.manager [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 729.862949] env[62066]: ERROR nova.compute.manager [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] vm_ref = self.build_virtual_machine(instance, [ 729.862949] env[62066]: ERROR nova.compute.manager [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 729.863375] env[62066]: ERROR nova.compute.manager [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] vif_infos = vmwarevif.get_vif_info(self._session, [ 729.863375] env[62066]: ERROR nova.compute.manager [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 729.863375] env[62066]: ERROR nova.compute.manager [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] for vif in network_info: [ 729.863375] env[62066]: ERROR nova.compute.manager [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 729.863375] env[62066]: ERROR nova.compute.manager [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] return self._sync_wrapper(fn, *args, **kwargs) [ 729.863375] env[62066]: ERROR nova.compute.manager [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 729.863375] env[62066]: ERROR nova.compute.manager [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] self.wait() [ 729.863375] env[62066]: ERROR nova.compute.manager [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 729.863375] env[62066]: ERROR nova.compute.manager [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] self[:] = self._gt.wait() [ 729.863375] env[62066]: ERROR nova.compute.manager [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 729.863375] env[62066]: ERROR nova.compute.manager [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] return self._exit_event.wait() [ 729.863375] env[62066]: ERROR nova.compute.manager [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 729.863375] env[62066]: ERROR nova.compute.manager [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] current.throw(*self._exc) [ 729.863814] env[62066]: ERROR nova.compute.manager [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 729.863814] env[62066]: ERROR nova.compute.manager [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] result = function(*args, **kwargs) [ 729.863814] env[62066]: ERROR nova.compute.manager [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 729.863814] env[62066]: ERROR nova.compute.manager [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] return func(*args, **kwargs) [ 729.863814] env[62066]: ERROR nova.compute.manager [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 729.863814] env[62066]: ERROR nova.compute.manager [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] raise e [ 729.863814] env[62066]: ERROR nova.compute.manager [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 729.863814] env[62066]: ERROR nova.compute.manager [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] nwinfo = self.network_api.allocate_for_instance( [ 729.863814] env[62066]: ERROR nova.compute.manager [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 729.863814] env[62066]: ERROR nova.compute.manager [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] created_port_ids = self._update_ports_for_instance( [ 729.863814] env[62066]: ERROR nova.compute.manager [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 729.863814] env[62066]: ERROR nova.compute.manager [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] with excutils.save_and_reraise_exception(): [ 729.863814] env[62066]: ERROR nova.compute.manager [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 729.864254] env[62066]: ERROR nova.compute.manager [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] self.force_reraise() [ 729.864254] env[62066]: ERROR nova.compute.manager [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 729.864254] env[62066]: ERROR nova.compute.manager [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] raise self.value [ 729.864254] env[62066]: ERROR nova.compute.manager [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 729.864254] env[62066]: ERROR nova.compute.manager [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] updated_port = self._update_port( [ 729.864254] env[62066]: ERROR nova.compute.manager [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 729.864254] env[62066]: ERROR nova.compute.manager [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] _ensure_no_port_binding_failure(port) [ 729.864254] env[62066]: ERROR nova.compute.manager [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 729.864254] env[62066]: ERROR nova.compute.manager [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] raise exception.PortBindingFailed(port_id=port['id']) [ 729.864254] env[62066]: ERROR nova.compute.manager [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] nova.exception.PortBindingFailed: Binding failed for port 91b620d9-65dc-4e47-87b9-3747a6897b6f, please check neutron logs for more information. [ 729.864254] env[62066]: ERROR nova.compute.manager [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] [ 729.864254] env[62066]: INFO nova.compute.manager [None req-a4d93091-69b6-4b35-887c-e7b927c2300a tempest-ServerMetadataTestJSON-1590103242 tempest-ServerMetadataTestJSON-1590103242-project-member] [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] Terminating instance [ 729.865344] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a4d93091-69b6-4b35-887c-e7b927c2300a tempest-ServerMetadataTestJSON-1590103242 tempest-ServerMetadataTestJSON-1590103242-project-member] Acquiring lock "refresh_cache-0ed07533-9e79-416d-a495-a8e20b97e108" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 730.018271] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 730.155690] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e9793203-98b3-4ca0-8a49-79767b1f52f7 tempest-ServersTestBootFromVolume-1535558002 tempest-ServersTestBootFromVolume-1535558002-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.385s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 730.156093] env[62066]: DEBUG nova.compute.manager [None req-e9793203-98b3-4ca0-8a49-79767b1f52f7 tempest-ServersTestBootFromVolume-1535558002 tempest-ServersTestBootFromVolume-1535558002-project-member] [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 730.159088] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 17.300s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 730.159263] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 730.159441] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62066) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 730.159742] env[62066]: DEBUG oslo_concurrency.lockutils [None req-8d368cc7-6ac0-44cd-b71e-9136d8337cb4 tempest-AttachInterfacesUnderV243Test-1832696447 tempest-AttachInterfacesUnderV243Test-1832696447-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.055s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 730.163599] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19a22255-e601-4b2c-89dd-45c9a5ae0c5b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.172045] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6156f90e-046d-4e26-833c-ed27050342a8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.185754] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f09c89b-71b1-4528-a0f8-3e21cb37dc14 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.193734] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0fc3de3-456c-4fa5-b941-4dce65073854 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.225204] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181419MB free_disk=154GB free_vcpus=48 pci_devices=None {{(pid=62066) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 730.226114] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 730.258844] env[62066]: DEBUG nova.network.neutron [req-3c9f96e0-2bb4-446e-8736-aac58fb68f24 req-94e34b36-7845-4be2-a3f9-b5a753c4719b service nova] [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 730.362799] env[62066]: DEBUG nova.network.neutron [req-3c9f96e0-2bb4-446e-8736-aac58fb68f24 req-94e34b36-7845-4be2-a3f9-b5a753c4719b service nova] [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 730.402137] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a2d866c9-e011-47d1-a21e-f9569fa299fe tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Acquiring lock "0ff2f9f4-6d85-4465-8e0a-fb39f2a25c9b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 730.402390] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a2d866c9-e011-47d1-a21e-f9569fa299fe tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Lock "0ff2f9f4-6d85-4465-8e0a-fb39f2a25c9b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 730.402591] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a2d866c9-e011-47d1-a21e-f9569fa299fe tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Acquiring lock "0ff2f9f4-6d85-4465-8e0a-fb39f2a25c9b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 730.402859] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a2d866c9-e011-47d1-a21e-f9569fa299fe tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Lock "0ff2f9f4-6d85-4465-8e0a-fb39f2a25c9b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 730.402945] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a2d866c9-e011-47d1-a21e-f9569fa299fe tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Lock "0ff2f9f4-6d85-4465-8e0a-fb39f2a25c9b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 730.405411] env[62066]: INFO nova.compute.manager [None req-a2d866c9-e011-47d1-a21e-f9569fa299fe tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] [instance: 0ff2f9f4-6d85-4465-8e0a-fb39f2a25c9b] Terminating instance [ 730.407043] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a2d866c9-e011-47d1-a21e-f9569fa299fe tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Acquiring lock "refresh_cache-0ff2f9f4-6d85-4465-8e0a-fb39f2a25c9b" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 730.407294] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a2d866c9-e011-47d1-a21e-f9569fa299fe tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Acquired lock "refresh_cache-0ff2f9f4-6d85-4465-8e0a-fb39f2a25c9b" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 730.407962] env[62066]: DEBUG nova.network.neutron [None req-a2d866c9-e011-47d1-a21e-f9569fa299fe tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] [instance: 0ff2f9f4-6d85-4465-8e0a-fb39f2a25c9b] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 730.667476] env[62066]: DEBUG nova.compute.utils [None req-e9793203-98b3-4ca0-8a49-79767b1f52f7 tempest-ServersTestBootFromVolume-1535558002 tempest-ServersTestBootFromVolume-1535558002-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 730.668994] env[62066]: DEBUG nova.compute.manager [None req-e9793203-98b3-4ca0-8a49-79767b1f52f7 tempest-ServersTestBootFromVolume-1535558002 tempest-ServersTestBootFromVolume-1535558002-project-member] [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 730.669177] env[62066]: DEBUG nova.network.neutron [None req-e9793203-98b3-4ca0-8a49-79767b1f52f7 tempest-ServersTestBootFromVolume-1535558002 tempest-ServersTestBootFromVolume-1535558002-project-member] [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 730.710436] env[62066]: DEBUG nova.policy [None req-e9793203-98b3-4ca0-8a49-79767b1f52f7 tempest-ServersTestBootFromVolume-1535558002 tempest-ServersTestBootFromVolume-1535558002-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1accc1330eee4f3e94bb79924c2e597d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '32fcf91353da4fcd978d20348493a522', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 730.865895] env[62066]: DEBUG oslo_concurrency.lockutils [req-3c9f96e0-2bb4-446e-8736-aac58fb68f24 req-94e34b36-7845-4be2-a3f9-b5a753c4719b service nova] Releasing lock "refresh_cache-0ed07533-9e79-416d-a495-a8e20b97e108" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 730.868046] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a4d93091-69b6-4b35-887c-e7b927c2300a tempest-ServerMetadataTestJSON-1590103242 tempest-ServerMetadataTestJSON-1590103242-project-member] Acquired lock "refresh_cache-0ed07533-9e79-416d-a495-a8e20b97e108" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 730.868046] env[62066]: DEBUG nova.network.neutron [None req-a4d93091-69b6-4b35-887c-e7b927c2300a tempest-ServerMetadataTestJSON-1590103242 tempest-ServerMetadataTestJSON-1590103242-project-member] [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 730.927020] env[62066]: DEBUG nova.network.neutron [None req-a2d866c9-e011-47d1-a21e-f9569fa299fe tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] [instance: 0ff2f9f4-6d85-4465-8e0a-fb39f2a25c9b] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 731.002255] env[62066]: DEBUG nova.network.neutron [None req-a2d866c9-e011-47d1-a21e-f9569fa299fe tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] [instance: 0ff2f9f4-6d85-4465-8e0a-fb39f2a25c9b] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 731.020169] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9047d361-acf2-4ff0-b536-1fd4804a5094 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.029271] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8502d914-aaf4-4fdb-8e8a-d9278d946247 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.063626] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3b75bcb-67b4-4dfe-a6ef-79dc66217b78 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.074713] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8eb069a6-9db5-4674-9cee-3903bf0a94d2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.096377] env[62066]: DEBUG nova.compute.provider_tree [None req-8d368cc7-6ac0-44cd-b71e-9136d8337cb4 tempest-AttachInterfacesUnderV243Test-1832696447 tempest-AttachInterfacesUnderV243Test-1832696447-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 731.173364] env[62066]: DEBUG nova.compute.manager [None req-e9793203-98b3-4ca0-8a49-79767b1f52f7 tempest-ServersTestBootFromVolume-1535558002 tempest-ServersTestBootFromVolume-1535558002-project-member] [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 731.318735] env[62066]: DEBUG nova.network.neutron [None req-e9793203-98b3-4ca0-8a49-79767b1f52f7 tempest-ServersTestBootFromVolume-1535558002 tempest-ServersTestBootFromVolume-1535558002-project-member] [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] Successfully created port: c44654b4-f098-40c1-93d9-b524a3b29c8c {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 731.386373] env[62066]: DEBUG nova.network.neutron [None req-a4d93091-69b6-4b35-887c-e7b927c2300a tempest-ServerMetadataTestJSON-1590103242 tempest-ServerMetadataTestJSON-1590103242-project-member] [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 731.564200] env[62066]: DEBUG nova.network.neutron [None req-a4d93091-69b6-4b35-887c-e7b927c2300a tempest-ServerMetadataTestJSON-1590103242 tempest-ServerMetadataTestJSON-1590103242-project-member] [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 731.564200] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a2d866c9-e011-47d1-a21e-f9569fa299fe tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Releasing lock "refresh_cache-0ff2f9f4-6d85-4465-8e0a-fb39f2a25c9b" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 731.564200] env[62066]: DEBUG nova.compute.manager [None req-a2d866c9-e011-47d1-a21e-f9569fa299fe tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] [instance: 0ff2f9f4-6d85-4465-8e0a-fb39f2a25c9b] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 731.564200] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-a2d866c9-e011-47d1-a21e-f9569fa299fe tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] [instance: 0ff2f9f4-6d85-4465-8e0a-fb39f2a25c9b] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 731.564200] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-345cb2a3-8351-49db-ac38-9c1e413a9e26 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.565413] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-a2d866c9-e011-47d1-a21e-f9569fa299fe tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] [instance: 0ff2f9f4-6d85-4465-8e0a-fb39f2a25c9b] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 731.565413] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0bfc2808-9e6b-4306-ae22-7838a0b084c4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.565413] env[62066]: DEBUG oslo_vmware.api [None req-a2d866c9-e011-47d1-a21e-f9569fa299fe tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Waiting for the task: (returnval){ [ 731.565413] env[62066]: value = "task-1340692" [ 731.565413] env[62066]: _type = "Task" [ 731.565413] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 731.565413] env[62066]: DEBUG oslo_vmware.api [None req-a2d866c9-e011-47d1-a21e-f9569fa299fe tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Task: {'id': task-1340692, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 731.598680] env[62066]: DEBUG nova.scheduler.client.report [None req-8d368cc7-6ac0-44cd-b71e-9136d8337cb4 tempest-AttachInterfacesUnderV243Test-1832696447 tempest-AttachInterfacesUnderV243Test-1832696447-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 731.678793] env[62066]: INFO nova.virt.block_device [None req-e9793203-98b3-4ca0-8a49-79767b1f52f7 tempest-ServersTestBootFromVolume-1535558002 tempest-ServersTestBootFromVolume-1535558002-project-member] [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] Booting with volume f03f24e7-ae56-4f25-b04f-01c97563c055 at /dev/sda [ 731.774771] env[62066]: DEBUG nova.compute.manager [req-373bfe97-bdab-4997-8463-55caa03ec0dc req-960d6b1a-73bc-4ba9-a038-7c0bce293115 service nova] [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] Received event network-vif-deleted-91b620d9-65dc-4e47-87b9-3747a6897b6f {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 731.807747] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-94582fd8-5342-48f0-977c-efb658773b5b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.817876] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b206678-f0fd-4840-beae-aca9b7d77d56 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.843071] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-004807b1-bff6-4a15-8079-97c95cdebe77 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.852558] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-676a8347-3e6e-4c64-817f-2d17ff115bda {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.883222] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a0079a2-38dd-4d91-9f6d-799cc4ad734d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.891607] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbb8844c-2771-41e9-8a54-84b2768f69dd {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.904558] env[62066]: DEBUG nova.virt.block_device [None req-e9793203-98b3-4ca0-8a49-79767b1f52f7 tempest-ServersTestBootFromVolume-1535558002 tempest-ServersTestBootFromVolume-1535558002-project-member] [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] Updating existing volume attachment record: 8eeca3f3-c8cd-454b-99a8-439e91ffbcb2 {{(pid=62066) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 732.010139] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a4d93091-69b6-4b35-887c-e7b927c2300a tempest-ServerMetadataTestJSON-1590103242 tempest-ServerMetadataTestJSON-1590103242-project-member] Releasing lock "refresh_cache-0ed07533-9e79-416d-a495-a8e20b97e108" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 732.010548] env[62066]: DEBUG nova.compute.manager [None req-a4d93091-69b6-4b35-887c-e7b927c2300a tempest-ServerMetadataTestJSON-1590103242 tempest-ServerMetadataTestJSON-1590103242-project-member] [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 732.010742] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-a4d93091-69b6-4b35-887c-e7b927c2300a tempest-ServerMetadataTestJSON-1590103242 tempest-ServerMetadataTestJSON-1590103242-project-member] [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 732.011060] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e20863b8-a9c4-4c2b-bbee-9bd27615d114 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.019610] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f5a11ef-130c-4d20-9de6-2e5736c3c7ae {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.038407] env[62066]: DEBUG oslo_vmware.api [None req-a2d866c9-e011-47d1-a21e-f9569fa299fe tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Task: {'id': task-1340692, 'name': PowerOffVM_Task, 'duration_secs': 0.132917} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 732.038676] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-a2d866c9-e011-47d1-a21e-f9569fa299fe tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] [instance: 0ff2f9f4-6d85-4465-8e0a-fb39f2a25c9b] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 732.038850] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-a2d866c9-e011-47d1-a21e-f9569fa299fe tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] [instance: 0ff2f9f4-6d85-4465-8e0a-fb39f2a25c9b] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 732.039094] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-fa4bf0ca-e795-4edc-a4b5-a2fa9a614ccb {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.044850] env[62066]: WARNING nova.virt.vmwareapi.vmops [None req-a4d93091-69b6-4b35-887c-e7b927c2300a tempest-ServerMetadataTestJSON-1590103242 tempest-ServerMetadataTestJSON-1590103242-project-member] [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 0ed07533-9e79-416d-a495-a8e20b97e108 could not be found. [ 732.045077] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-a4d93091-69b6-4b35-887c-e7b927c2300a tempest-ServerMetadataTestJSON-1590103242 tempest-ServerMetadataTestJSON-1590103242-project-member] [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 732.045384] env[62066]: INFO nova.compute.manager [None req-a4d93091-69b6-4b35-887c-e7b927c2300a tempest-ServerMetadataTestJSON-1590103242 tempest-ServerMetadataTestJSON-1590103242-project-member] [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] Took 0.03 seconds to destroy the instance on the hypervisor. [ 732.045608] env[62066]: DEBUG oslo.service.loopingcall [None req-a4d93091-69b6-4b35-887c-e7b927c2300a tempest-ServerMetadataTestJSON-1590103242 tempest-ServerMetadataTestJSON-1590103242-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 732.046195] env[62066]: DEBUG nova.compute.manager [-] [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 732.046321] env[62066]: DEBUG nova.network.neutron [-] [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 732.065893] env[62066]: DEBUG nova.network.neutron [-] [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 732.072552] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-a2d866c9-e011-47d1-a21e-f9569fa299fe tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] [instance: 0ff2f9f4-6d85-4465-8e0a-fb39f2a25c9b] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 732.072762] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-a2d866c9-e011-47d1-a21e-f9569fa299fe tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] [instance: 0ff2f9f4-6d85-4465-8e0a-fb39f2a25c9b] Deleting contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 732.072974] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-a2d866c9-e011-47d1-a21e-f9569fa299fe tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Deleting the datastore file [datastore2] 0ff2f9f4-6d85-4465-8e0a-fb39f2a25c9b {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 732.073261] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d4e8a00d-25bd-47b8-9c38-ea1ff5077b18 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.081419] env[62066]: DEBUG oslo_vmware.api [None req-a2d866c9-e011-47d1-a21e-f9569fa299fe tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Waiting for the task: (returnval){ [ 732.081419] env[62066]: value = "task-1340694" [ 732.081419] env[62066]: _type = "Task" [ 732.081419] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 732.089788] env[62066]: DEBUG oslo_vmware.api [None req-a2d866c9-e011-47d1-a21e-f9569fa299fe tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Task: {'id': task-1340694, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 732.107576] env[62066]: DEBUG oslo_concurrency.lockutils [None req-8d368cc7-6ac0-44cd-b71e-9136d8337cb4 tempest-AttachInterfacesUnderV243Test-1832696447 tempest-AttachInterfacesUnderV243Test-1832696447-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.948s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 732.108177] env[62066]: ERROR nova.compute.manager [None req-8d368cc7-6ac0-44cd-b71e-9136d8337cb4 tempest-AttachInterfacesUnderV243Test-1832696447 tempest-AttachInterfacesUnderV243Test-1832696447-project-member] [instance: f91a90f7-be73-424c-966c-1be6f37a0864] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 01c976c4-59c2-4855-942f-61915dbd9fe3, please check neutron logs for more information. [ 732.108177] env[62066]: ERROR nova.compute.manager [instance: f91a90f7-be73-424c-966c-1be6f37a0864] Traceback (most recent call last): [ 732.108177] env[62066]: ERROR nova.compute.manager [instance: f91a90f7-be73-424c-966c-1be6f37a0864] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 732.108177] env[62066]: ERROR nova.compute.manager [instance: f91a90f7-be73-424c-966c-1be6f37a0864] self.driver.spawn(context, instance, image_meta, [ 732.108177] env[62066]: ERROR nova.compute.manager [instance: f91a90f7-be73-424c-966c-1be6f37a0864] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 732.108177] env[62066]: ERROR nova.compute.manager [instance: f91a90f7-be73-424c-966c-1be6f37a0864] self._vmops.spawn(context, instance, image_meta, injected_files, [ 732.108177] env[62066]: ERROR nova.compute.manager [instance: f91a90f7-be73-424c-966c-1be6f37a0864] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 732.108177] env[62066]: ERROR nova.compute.manager [instance: f91a90f7-be73-424c-966c-1be6f37a0864] vm_ref = self.build_virtual_machine(instance, [ 732.108177] env[62066]: ERROR nova.compute.manager [instance: f91a90f7-be73-424c-966c-1be6f37a0864] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 732.108177] env[62066]: ERROR nova.compute.manager [instance: f91a90f7-be73-424c-966c-1be6f37a0864] vif_infos = vmwarevif.get_vif_info(self._session, [ 732.108177] env[62066]: ERROR nova.compute.manager [instance: f91a90f7-be73-424c-966c-1be6f37a0864] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 732.108486] env[62066]: ERROR nova.compute.manager [instance: f91a90f7-be73-424c-966c-1be6f37a0864] for vif in network_info: [ 732.108486] env[62066]: ERROR nova.compute.manager [instance: f91a90f7-be73-424c-966c-1be6f37a0864] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 732.108486] env[62066]: ERROR nova.compute.manager [instance: f91a90f7-be73-424c-966c-1be6f37a0864] return self._sync_wrapper(fn, *args, **kwargs) [ 732.108486] env[62066]: ERROR nova.compute.manager [instance: f91a90f7-be73-424c-966c-1be6f37a0864] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 732.108486] env[62066]: ERROR nova.compute.manager [instance: f91a90f7-be73-424c-966c-1be6f37a0864] self.wait() [ 732.108486] env[62066]: ERROR nova.compute.manager [instance: f91a90f7-be73-424c-966c-1be6f37a0864] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 732.108486] env[62066]: ERROR nova.compute.manager [instance: f91a90f7-be73-424c-966c-1be6f37a0864] self[:] = self._gt.wait() [ 732.108486] env[62066]: ERROR nova.compute.manager [instance: f91a90f7-be73-424c-966c-1be6f37a0864] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 732.108486] env[62066]: ERROR nova.compute.manager [instance: f91a90f7-be73-424c-966c-1be6f37a0864] return self._exit_event.wait() [ 732.108486] env[62066]: ERROR nova.compute.manager [instance: f91a90f7-be73-424c-966c-1be6f37a0864] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 732.108486] env[62066]: ERROR nova.compute.manager [instance: f91a90f7-be73-424c-966c-1be6f37a0864] result = hub.switch() [ 732.108486] env[62066]: ERROR nova.compute.manager [instance: f91a90f7-be73-424c-966c-1be6f37a0864] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 732.108486] env[62066]: ERROR nova.compute.manager [instance: f91a90f7-be73-424c-966c-1be6f37a0864] return self.greenlet.switch() [ 732.108807] env[62066]: ERROR nova.compute.manager [instance: f91a90f7-be73-424c-966c-1be6f37a0864] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 732.108807] env[62066]: ERROR nova.compute.manager [instance: f91a90f7-be73-424c-966c-1be6f37a0864] result = function(*args, **kwargs) [ 732.108807] env[62066]: ERROR nova.compute.manager [instance: f91a90f7-be73-424c-966c-1be6f37a0864] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 732.108807] env[62066]: ERROR nova.compute.manager [instance: f91a90f7-be73-424c-966c-1be6f37a0864] return func(*args, **kwargs) [ 732.108807] env[62066]: ERROR nova.compute.manager [instance: f91a90f7-be73-424c-966c-1be6f37a0864] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 732.108807] env[62066]: ERROR nova.compute.manager [instance: f91a90f7-be73-424c-966c-1be6f37a0864] raise e [ 732.108807] env[62066]: ERROR nova.compute.manager [instance: f91a90f7-be73-424c-966c-1be6f37a0864] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 732.108807] env[62066]: ERROR nova.compute.manager [instance: f91a90f7-be73-424c-966c-1be6f37a0864] nwinfo = self.network_api.allocate_for_instance( [ 732.108807] env[62066]: ERROR nova.compute.manager [instance: f91a90f7-be73-424c-966c-1be6f37a0864] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 732.108807] env[62066]: ERROR nova.compute.manager [instance: f91a90f7-be73-424c-966c-1be6f37a0864] created_port_ids = self._update_ports_for_instance( [ 732.108807] env[62066]: ERROR nova.compute.manager [instance: f91a90f7-be73-424c-966c-1be6f37a0864] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 732.108807] env[62066]: ERROR nova.compute.manager [instance: f91a90f7-be73-424c-966c-1be6f37a0864] with excutils.save_and_reraise_exception(): [ 732.108807] env[62066]: ERROR nova.compute.manager [instance: f91a90f7-be73-424c-966c-1be6f37a0864] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 732.109160] env[62066]: ERROR nova.compute.manager [instance: f91a90f7-be73-424c-966c-1be6f37a0864] self.force_reraise() [ 732.109160] env[62066]: ERROR nova.compute.manager [instance: f91a90f7-be73-424c-966c-1be6f37a0864] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 732.109160] env[62066]: ERROR nova.compute.manager [instance: f91a90f7-be73-424c-966c-1be6f37a0864] raise self.value [ 732.109160] env[62066]: ERROR nova.compute.manager [instance: f91a90f7-be73-424c-966c-1be6f37a0864] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 732.109160] env[62066]: ERROR nova.compute.manager [instance: f91a90f7-be73-424c-966c-1be6f37a0864] updated_port = self._update_port( [ 732.109160] env[62066]: ERROR nova.compute.manager [instance: f91a90f7-be73-424c-966c-1be6f37a0864] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 732.109160] env[62066]: ERROR nova.compute.manager [instance: f91a90f7-be73-424c-966c-1be6f37a0864] _ensure_no_port_binding_failure(port) [ 732.109160] env[62066]: ERROR nova.compute.manager [instance: f91a90f7-be73-424c-966c-1be6f37a0864] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 732.109160] env[62066]: ERROR nova.compute.manager [instance: f91a90f7-be73-424c-966c-1be6f37a0864] raise exception.PortBindingFailed(port_id=port['id']) [ 732.109160] env[62066]: ERROR nova.compute.manager [instance: f91a90f7-be73-424c-966c-1be6f37a0864] nova.exception.PortBindingFailed: Binding failed for port 01c976c4-59c2-4855-942f-61915dbd9fe3, please check neutron logs for more information. [ 732.109160] env[62066]: ERROR nova.compute.manager [instance: f91a90f7-be73-424c-966c-1be6f37a0864] [ 732.109397] env[62066]: DEBUG nova.compute.utils [None req-8d368cc7-6ac0-44cd-b71e-9136d8337cb4 tempest-AttachInterfacesUnderV243Test-1832696447 tempest-AttachInterfacesUnderV243Test-1832696447-project-member] [instance: f91a90f7-be73-424c-966c-1be6f37a0864] Binding failed for port 01c976c4-59c2-4855-942f-61915dbd9fe3, please check neutron logs for more information. {{(pid=62066) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 732.110526] env[62066]: DEBUG oslo_concurrency.lockutils [None req-478d4566-e74c-4f71-b94d-419dc0a31446 tempest-ServerPasswordTestJSON-849130511 tempest-ServerPasswordTestJSON-849130511-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.704s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 732.111831] env[62066]: INFO nova.compute.claims [None req-478d4566-e74c-4f71-b94d-419dc0a31446 tempest-ServerPasswordTestJSON-849130511 tempest-ServerPasswordTestJSON-849130511-project-member] [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 732.114385] env[62066]: DEBUG nova.compute.manager [None req-8d368cc7-6ac0-44cd-b71e-9136d8337cb4 tempest-AttachInterfacesUnderV243Test-1832696447 tempest-AttachInterfacesUnderV243Test-1832696447-project-member] [instance: f91a90f7-be73-424c-966c-1be6f37a0864] Build of instance f91a90f7-be73-424c-966c-1be6f37a0864 was re-scheduled: Binding failed for port 01c976c4-59c2-4855-942f-61915dbd9fe3, please check neutron logs for more information. {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 732.114855] env[62066]: DEBUG nova.compute.manager [None req-8d368cc7-6ac0-44cd-b71e-9136d8337cb4 tempest-AttachInterfacesUnderV243Test-1832696447 tempest-AttachInterfacesUnderV243Test-1832696447-project-member] [instance: f91a90f7-be73-424c-966c-1be6f37a0864] Unplugging VIFs for instance {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 732.115121] env[62066]: DEBUG oslo_concurrency.lockutils [None req-8d368cc7-6ac0-44cd-b71e-9136d8337cb4 tempest-AttachInterfacesUnderV243Test-1832696447 tempest-AttachInterfacesUnderV243Test-1832696447-project-member] Acquiring lock "refresh_cache-f91a90f7-be73-424c-966c-1be6f37a0864" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 732.115287] env[62066]: DEBUG oslo_concurrency.lockutils [None req-8d368cc7-6ac0-44cd-b71e-9136d8337cb4 tempest-AttachInterfacesUnderV243Test-1832696447 tempest-AttachInterfacesUnderV243Test-1832696447-project-member] Acquired lock "refresh_cache-f91a90f7-be73-424c-966c-1be6f37a0864" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 732.115456] env[62066]: DEBUG nova.network.neutron [None req-8d368cc7-6ac0-44cd-b71e-9136d8337cb4 tempest-AttachInterfacesUnderV243Test-1832696447 tempest-AttachInterfacesUnderV243Test-1832696447-project-member] [instance: f91a90f7-be73-424c-966c-1be6f37a0864] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 732.566792] env[62066]: ERROR nova.compute.manager [None req-e9793203-98b3-4ca0-8a49-79767b1f52f7 tempest-ServersTestBootFromVolume-1535558002 tempest-ServersTestBootFromVolume-1535558002-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port c44654b4-f098-40c1-93d9-b524a3b29c8c, please check neutron logs for more information. [ 732.566792] env[62066]: ERROR nova.compute.manager Traceback (most recent call last): [ 732.566792] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 732.566792] env[62066]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 732.566792] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 732.566792] env[62066]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 732.566792] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 732.566792] env[62066]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 732.566792] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 732.566792] env[62066]: ERROR nova.compute.manager self.force_reraise() [ 732.566792] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 732.566792] env[62066]: ERROR nova.compute.manager raise self.value [ 732.566792] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 732.566792] env[62066]: ERROR nova.compute.manager updated_port = self._update_port( [ 732.566792] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 732.566792] env[62066]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 732.567504] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 732.567504] env[62066]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 732.567504] env[62066]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port c44654b4-f098-40c1-93d9-b524a3b29c8c, please check neutron logs for more information. [ 732.567504] env[62066]: ERROR nova.compute.manager [ 732.567504] env[62066]: Traceback (most recent call last): [ 732.567504] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 732.567504] env[62066]: listener.cb(fileno) [ 732.567504] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 732.567504] env[62066]: result = function(*args, **kwargs) [ 732.567504] env[62066]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 732.567504] env[62066]: return func(*args, **kwargs) [ 732.567504] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 732.567504] env[62066]: raise e [ 732.567504] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 732.567504] env[62066]: nwinfo = self.network_api.allocate_for_instance( [ 732.567504] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 732.567504] env[62066]: created_port_ids = self._update_ports_for_instance( [ 732.567504] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 732.567504] env[62066]: with excutils.save_and_reraise_exception(): [ 732.567504] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 732.567504] env[62066]: self.force_reraise() [ 732.567504] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 732.567504] env[62066]: raise self.value [ 732.567504] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 732.567504] env[62066]: updated_port = self._update_port( [ 732.567504] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 732.567504] env[62066]: _ensure_no_port_binding_failure(port) [ 732.567504] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 732.567504] env[62066]: raise exception.PortBindingFailed(port_id=port['id']) [ 732.568211] env[62066]: nova.exception.PortBindingFailed: Binding failed for port c44654b4-f098-40c1-93d9-b524a3b29c8c, please check neutron logs for more information. [ 732.568211] env[62066]: Removing descriptor: 20 [ 732.574971] env[62066]: DEBUG nova.network.neutron [-] [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 732.593503] env[62066]: DEBUG oslo_vmware.api [None req-a2d866c9-e011-47d1-a21e-f9569fa299fe tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Task: {'id': task-1340694, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.130515} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 732.593834] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-a2d866c9-e011-47d1-a21e-f9569fa299fe tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 732.594064] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-a2d866c9-e011-47d1-a21e-f9569fa299fe tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] [instance: 0ff2f9f4-6d85-4465-8e0a-fb39f2a25c9b] Deleted contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 732.594332] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-a2d866c9-e011-47d1-a21e-f9569fa299fe tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] [instance: 0ff2f9f4-6d85-4465-8e0a-fb39f2a25c9b] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 732.594719] env[62066]: INFO nova.compute.manager [None req-a2d866c9-e011-47d1-a21e-f9569fa299fe tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] [instance: 0ff2f9f4-6d85-4465-8e0a-fb39f2a25c9b] Took 1.09 seconds to destroy the instance on the hypervisor. [ 732.595043] env[62066]: DEBUG oslo.service.loopingcall [None req-a2d866c9-e011-47d1-a21e-f9569fa299fe tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 732.595370] env[62066]: DEBUG nova.compute.manager [-] [instance: 0ff2f9f4-6d85-4465-8e0a-fb39f2a25c9b] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 732.595502] env[62066]: DEBUG nova.network.neutron [-] [instance: 0ff2f9f4-6d85-4465-8e0a-fb39f2a25c9b] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 732.625582] env[62066]: DEBUG nova.network.neutron [-] [instance: 0ff2f9f4-6d85-4465-8e0a-fb39f2a25c9b] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 732.647990] env[62066]: DEBUG nova.network.neutron [None req-8d368cc7-6ac0-44cd-b71e-9136d8337cb4 tempest-AttachInterfacesUnderV243Test-1832696447 tempest-AttachInterfacesUnderV243Test-1832696447-project-member] [instance: f91a90f7-be73-424c-966c-1be6f37a0864] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 732.774207] env[62066]: DEBUG nova.network.neutron [None req-8d368cc7-6ac0-44cd-b71e-9136d8337cb4 tempest-AttachInterfacesUnderV243Test-1832696447 tempest-AttachInterfacesUnderV243Test-1832696447-project-member] [instance: f91a90f7-be73-424c-966c-1be6f37a0864] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 733.079639] env[62066]: INFO nova.compute.manager [-] [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] Took 1.03 seconds to deallocate network for instance. [ 733.082612] env[62066]: DEBUG nova.compute.claims [None req-a4d93091-69b6-4b35-887c-e7b927c2300a tempest-ServerMetadataTestJSON-1590103242 tempest-ServerMetadataTestJSON-1590103242-project-member] [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] Aborting claim: {{(pid=62066) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 733.082612] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a4d93091-69b6-4b35-887c-e7b927c2300a tempest-ServerMetadataTestJSON-1590103242 tempest-ServerMetadataTestJSON-1590103242-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 733.130845] env[62066]: DEBUG nova.network.neutron [-] [instance: 0ff2f9f4-6d85-4465-8e0a-fb39f2a25c9b] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 733.279596] env[62066]: DEBUG oslo_concurrency.lockutils [None req-8d368cc7-6ac0-44cd-b71e-9136d8337cb4 tempest-AttachInterfacesUnderV243Test-1832696447 tempest-AttachInterfacesUnderV243Test-1832696447-project-member] Releasing lock "refresh_cache-f91a90f7-be73-424c-966c-1be6f37a0864" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 733.279945] env[62066]: DEBUG nova.compute.manager [None req-8d368cc7-6ac0-44cd-b71e-9136d8337cb4 tempest-AttachInterfacesUnderV243Test-1832696447 tempest-AttachInterfacesUnderV243Test-1832696447-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 733.280210] env[62066]: DEBUG nova.compute.manager [None req-8d368cc7-6ac0-44cd-b71e-9136d8337cb4 tempest-AttachInterfacesUnderV243Test-1832696447 tempest-AttachInterfacesUnderV243Test-1832696447-project-member] [instance: f91a90f7-be73-424c-966c-1be6f37a0864] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 733.280443] env[62066]: DEBUG nova.network.neutron [None req-8d368cc7-6ac0-44cd-b71e-9136d8337cb4 tempest-AttachInterfacesUnderV243Test-1832696447 tempest-AttachInterfacesUnderV243Test-1832696447-project-member] [instance: f91a90f7-be73-424c-966c-1be6f37a0864] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 733.296313] env[62066]: DEBUG nova.network.neutron [None req-8d368cc7-6ac0-44cd-b71e-9136d8337cb4 tempest-AttachInterfacesUnderV243Test-1832696447 tempest-AttachInterfacesUnderV243Test-1832696447-project-member] [instance: f91a90f7-be73-424c-966c-1be6f37a0864] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 733.480505] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55b78f5d-43a7-4145-bba2-5e61315f5f37 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.488128] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4138c2a-1bb4-4267-a194-878329087be0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.517042] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e13d721b-19ae-49bb-9ac1-9337dc675112 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.523972] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c70e4c5b-ff38-42c5-90c3-47ab51a56416 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.539431] env[62066]: DEBUG nova.compute.provider_tree [None req-478d4566-e74c-4f71-b94d-419dc0a31446 tempest-ServerPasswordTestJSON-849130511 tempest-ServerPasswordTestJSON-849130511-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 733.633346] env[62066]: INFO nova.compute.manager [-] [instance: 0ff2f9f4-6d85-4465-8e0a-fb39f2a25c9b] Took 1.04 seconds to deallocate network for instance. [ 733.797619] env[62066]: DEBUG nova.compute.manager [req-6778a3ed-b2a4-406d-b333-c5790420e85c req-9bbd6053-601c-4098-b674-e45d9ca1f944 service nova] [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] Received event network-changed-c44654b4-f098-40c1-93d9-b524a3b29c8c {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 733.797814] env[62066]: DEBUG nova.compute.manager [req-6778a3ed-b2a4-406d-b333-c5790420e85c req-9bbd6053-601c-4098-b674-e45d9ca1f944 service nova] [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] Refreshing instance network info cache due to event network-changed-c44654b4-f098-40c1-93d9-b524a3b29c8c. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 733.798243] env[62066]: DEBUG oslo_concurrency.lockutils [req-6778a3ed-b2a4-406d-b333-c5790420e85c req-9bbd6053-601c-4098-b674-e45d9ca1f944 service nova] Acquiring lock "refresh_cache-2b81d060-ecb7-4ae5-b51d-33197937a76f" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 733.798485] env[62066]: DEBUG oslo_concurrency.lockutils [req-6778a3ed-b2a4-406d-b333-c5790420e85c req-9bbd6053-601c-4098-b674-e45d9ca1f944 service nova] Acquired lock "refresh_cache-2b81d060-ecb7-4ae5-b51d-33197937a76f" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 733.798690] env[62066]: DEBUG nova.network.neutron [req-6778a3ed-b2a4-406d-b333-c5790420e85c req-9bbd6053-601c-4098-b674-e45d9ca1f944 service nova] [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] Refreshing network info cache for port c44654b4-f098-40c1-93d9-b524a3b29c8c {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 733.804154] env[62066]: DEBUG nova.network.neutron [None req-8d368cc7-6ac0-44cd-b71e-9136d8337cb4 tempest-AttachInterfacesUnderV243Test-1832696447 tempest-AttachInterfacesUnderV243Test-1832696447-project-member] [instance: f91a90f7-be73-424c-966c-1be6f37a0864] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 734.033043] env[62066]: DEBUG nova.compute.manager [None req-e9793203-98b3-4ca0-8a49-79767b1f52f7 tempest-ServersTestBootFromVolume-1535558002 tempest-ServersTestBootFromVolume-1535558002-project-member] [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 734.033137] env[62066]: DEBUG nova.virt.hardware [None req-e9793203-98b3-4ca0-8a49-79767b1f52f7 tempest-ServersTestBootFromVolume-1535558002 tempest-ServersTestBootFromVolume-1535558002-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-23T13:40:41Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 734.033357] env[62066]: DEBUG nova.virt.hardware [None req-e9793203-98b3-4ca0-8a49-79767b1f52f7 tempest-ServersTestBootFromVolume-1535558002 tempest-ServersTestBootFromVolume-1535558002-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 734.033512] env[62066]: DEBUG nova.virt.hardware [None req-e9793203-98b3-4ca0-8a49-79767b1f52f7 tempest-ServersTestBootFromVolume-1535558002 tempest-ServersTestBootFromVolume-1535558002-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 734.033693] env[62066]: DEBUG nova.virt.hardware [None req-e9793203-98b3-4ca0-8a49-79767b1f52f7 tempest-ServersTestBootFromVolume-1535558002 tempest-ServersTestBootFromVolume-1535558002-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 734.033841] env[62066]: DEBUG nova.virt.hardware [None req-e9793203-98b3-4ca0-8a49-79767b1f52f7 tempest-ServersTestBootFromVolume-1535558002 tempest-ServersTestBootFromVolume-1535558002-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 734.033986] env[62066]: DEBUG nova.virt.hardware [None req-e9793203-98b3-4ca0-8a49-79767b1f52f7 tempest-ServersTestBootFromVolume-1535558002 tempest-ServersTestBootFromVolume-1535558002-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 734.034209] env[62066]: DEBUG nova.virt.hardware [None req-e9793203-98b3-4ca0-8a49-79767b1f52f7 tempest-ServersTestBootFromVolume-1535558002 tempest-ServersTestBootFromVolume-1535558002-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 734.034367] env[62066]: DEBUG nova.virt.hardware [None req-e9793203-98b3-4ca0-8a49-79767b1f52f7 tempest-ServersTestBootFromVolume-1535558002 tempest-ServersTestBootFromVolume-1535558002-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 734.034535] env[62066]: DEBUG nova.virt.hardware [None req-e9793203-98b3-4ca0-8a49-79767b1f52f7 tempest-ServersTestBootFromVolume-1535558002 tempest-ServersTestBootFromVolume-1535558002-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 734.034694] env[62066]: DEBUG nova.virt.hardware [None req-e9793203-98b3-4ca0-8a49-79767b1f52f7 tempest-ServersTestBootFromVolume-1535558002 tempest-ServersTestBootFromVolume-1535558002-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 734.034884] env[62066]: DEBUG nova.virt.hardware [None req-e9793203-98b3-4ca0-8a49-79767b1f52f7 tempest-ServersTestBootFromVolume-1535558002 tempest-ServersTestBootFromVolume-1535558002-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 734.035746] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcb225c8-38d5-4da2-a46f-9a5113c0b513 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.043658] env[62066]: DEBUG nova.scheduler.client.report [None req-478d4566-e74c-4f71-b94d-419dc0a31446 tempest-ServerPasswordTestJSON-849130511 tempest-ServerPasswordTestJSON-849130511-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 734.047983] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-971c3f36-4239-42dd-b25c-ab93457f7e4c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.062291] env[62066]: ERROR nova.compute.manager [None req-e9793203-98b3-4ca0-8a49-79767b1f52f7 tempest-ServersTestBootFromVolume-1535558002 tempest-ServersTestBootFromVolume-1535558002-project-member] [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port c44654b4-f098-40c1-93d9-b524a3b29c8c, please check neutron logs for more information. [ 734.062291] env[62066]: ERROR nova.compute.manager [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] Traceback (most recent call last): [ 734.062291] env[62066]: ERROR nova.compute.manager [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 734.062291] env[62066]: ERROR nova.compute.manager [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] yield resources [ 734.062291] env[62066]: ERROR nova.compute.manager [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 734.062291] env[62066]: ERROR nova.compute.manager [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] self.driver.spawn(context, instance, image_meta, [ 734.062291] env[62066]: ERROR nova.compute.manager [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 734.062291] env[62066]: ERROR nova.compute.manager [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 734.062291] env[62066]: ERROR nova.compute.manager [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 734.062291] env[62066]: ERROR nova.compute.manager [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] vm_ref = self.build_virtual_machine(instance, [ 734.062291] env[62066]: ERROR nova.compute.manager [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 734.062604] env[62066]: ERROR nova.compute.manager [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] vif_infos = vmwarevif.get_vif_info(self._session, [ 734.062604] env[62066]: ERROR nova.compute.manager [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 734.062604] env[62066]: ERROR nova.compute.manager [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] for vif in network_info: [ 734.062604] env[62066]: ERROR nova.compute.manager [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 734.062604] env[62066]: ERROR nova.compute.manager [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] return self._sync_wrapper(fn, *args, **kwargs) [ 734.062604] env[62066]: ERROR nova.compute.manager [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 734.062604] env[62066]: ERROR nova.compute.manager [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] self.wait() [ 734.062604] env[62066]: ERROR nova.compute.manager [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 734.062604] env[62066]: ERROR nova.compute.manager [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] self[:] = self._gt.wait() [ 734.062604] env[62066]: ERROR nova.compute.manager [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 734.062604] env[62066]: ERROR nova.compute.manager [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] return self._exit_event.wait() [ 734.062604] env[62066]: ERROR nova.compute.manager [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 734.062604] env[62066]: ERROR nova.compute.manager [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] current.throw(*self._exc) [ 734.062860] env[62066]: ERROR nova.compute.manager [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 734.062860] env[62066]: ERROR nova.compute.manager [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] result = function(*args, **kwargs) [ 734.062860] env[62066]: ERROR nova.compute.manager [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 734.062860] env[62066]: ERROR nova.compute.manager [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] return func(*args, **kwargs) [ 734.062860] env[62066]: ERROR nova.compute.manager [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 734.062860] env[62066]: ERROR nova.compute.manager [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] raise e [ 734.062860] env[62066]: ERROR nova.compute.manager [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 734.062860] env[62066]: ERROR nova.compute.manager [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] nwinfo = self.network_api.allocate_for_instance( [ 734.062860] env[62066]: ERROR nova.compute.manager [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 734.062860] env[62066]: ERROR nova.compute.manager [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] created_port_ids = self._update_ports_for_instance( [ 734.062860] env[62066]: ERROR nova.compute.manager [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 734.062860] env[62066]: ERROR nova.compute.manager [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] with excutils.save_and_reraise_exception(): [ 734.062860] env[62066]: ERROR nova.compute.manager [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 734.063128] env[62066]: ERROR nova.compute.manager [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] self.force_reraise() [ 734.063128] env[62066]: ERROR nova.compute.manager [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 734.063128] env[62066]: ERROR nova.compute.manager [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] raise self.value [ 734.063128] env[62066]: ERROR nova.compute.manager [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 734.063128] env[62066]: ERROR nova.compute.manager [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] updated_port = self._update_port( [ 734.063128] env[62066]: ERROR nova.compute.manager [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 734.063128] env[62066]: ERROR nova.compute.manager [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] _ensure_no_port_binding_failure(port) [ 734.063128] env[62066]: ERROR nova.compute.manager [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 734.063128] env[62066]: ERROR nova.compute.manager [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] raise exception.PortBindingFailed(port_id=port['id']) [ 734.063128] env[62066]: ERROR nova.compute.manager [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] nova.exception.PortBindingFailed: Binding failed for port c44654b4-f098-40c1-93d9-b524a3b29c8c, please check neutron logs for more information. [ 734.063128] env[62066]: ERROR nova.compute.manager [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] [ 734.063128] env[62066]: INFO nova.compute.manager [None req-e9793203-98b3-4ca0-8a49-79767b1f52f7 tempest-ServersTestBootFromVolume-1535558002 tempest-ServersTestBootFromVolume-1535558002-project-member] [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] Terminating instance [ 734.064687] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e9793203-98b3-4ca0-8a49-79767b1f52f7 tempest-ServersTestBootFromVolume-1535558002 tempest-ServersTestBootFromVolume-1535558002-project-member] Acquiring lock "refresh_cache-2b81d060-ecb7-4ae5-b51d-33197937a76f" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 734.139611] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a2d866c9-e011-47d1-a21e-f9569fa299fe tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 734.303834] env[62066]: INFO nova.compute.manager [None req-8d368cc7-6ac0-44cd-b71e-9136d8337cb4 tempest-AttachInterfacesUnderV243Test-1832696447 tempest-AttachInterfacesUnderV243Test-1832696447-project-member] [instance: f91a90f7-be73-424c-966c-1be6f37a0864] Took 1.02 seconds to deallocate network for instance. [ 734.317927] env[62066]: DEBUG nova.network.neutron [req-6778a3ed-b2a4-406d-b333-c5790420e85c req-9bbd6053-601c-4098-b674-e45d9ca1f944 service nova] [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 734.401083] env[62066]: DEBUG nova.network.neutron [req-6778a3ed-b2a4-406d-b333-c5790420e85c req-9bbd6053-601c-4098-b674-e45d9ca1f944 service nova] [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 734.553488] env[62066]: DEBUG oslo_concurrency.lockutils [None req-478d4566-e74c-4f71-b94d-419dc0a31446 tempest-ServerPasswordTestJSON-849130511 tempest-ServerPasswordTestJSON-849130511-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.443s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 734.554016] env[62066]: DEBUG nova.compute.manager [None req-478d4566-e74c-4f71-b94d-419dc0a31446 tempest-ServerPasswordTestJSON-849130511 tempest-ServerPasswordTestJSON-849130511-project-member] [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 734.556576] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5b5d7867-70eb-472d-8d1a-c2bb035ec6db tempest-InstanceActionsV221TestJSON-1098427857 tempest-InstanceActionsV221TestJSON-1098427857-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 12.621s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 734.904647] env[62066]: DEBUG oslo_concurrency.lockutils [req-6778a3ed-b2a4-406d-b333-c5790420e85c req-9bbd6053-601c-4098-b674-e45d9ca1f944 service nova] Releasing lock "refresh_cache-2b81d060-ecb7-4ae5-b51d-33197937a76f" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 734.904906] env[62066]: DEBUG nova.compute.manager [req-6778a3ed-b2a4-406d-b333-c5790420e85c req-9bbd6053-601c-4098-b674-e45d9ca1f944 service nova] [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] Received event network-vif-deleted-c44654b4-f098-40c1-93d9-b524a3b29c8c {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 734.905068] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e9793203-98b3-4ca0-8a49-79767b1f52f7 tempest-ServersTestBootFromVolume-1535558002 tempest-ServersTestBootFromVolume-1535558002-project-member] Acquired lock "refresh_cache-2b81d060-ecb7-4ae5-b51d-33197937a76f" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 734.905284] env[62066]: DEBUG nova.network.neutron [None req-e9793203-98b3-4ca0-8a49-79767b1f52f7 tempest-ServersTestBootFromVolume-1535558002 tempest-ServersTestBootFromVolume-1535558002-project-member] [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 735.064774] env[62066]: DEBUG nova.compute.utils [None req-478d4566-e74c-4f71-b94d-419dc0a31446 tempest-ServerPasswordTestJSON-849130511 tempest-ServerPasswordTestJSON-849130511-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 735.066282] env[62066]: DEBUG nova.compute.manager [None req-478d4566-e74c-4f71-b94d-419dc0a31446 tempest-ServerPasswordTestJSON-849130511 tempest-ServerPasswordTestJSON-849130511-project-member] [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 735.066428] env[62066]: DEBUG nova.network.neutron [None req-478d4566-e74c-4f71-b94d-419dc0a31446 tempest-ServerPasswordTestJSON-849130511 tempest-ServerPasswordTestJSON-849130511-project-member] [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 735.123807] env[62066]: DEBUG nova.policy [None req-478d4566-e74c-4f71-b94d-419dc0a31446 tempest-ServerPasswordTestJSON-849130511 tempest-ServerPasswordTestJSON-849130511-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '81ff1c4c24924a94893c5b7700591f14', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f66655e3c27a4d528aed6f00f3dff4d4', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 735.336131] env[62066]: INFO nova.scheduler.client.report [None req-8d368cc7-6ac0-44cd-b71e-9136d8337cb4 tempest-AttachInterfacesUnderV243Test-1832696447 tempest-AttachInterfacesUnderV243Test-1832696447-project-member] Deleted allocations for instance f91a90f7-be73-424c-966c-1be6f37a0864 [ 735.381550] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63dea487-45b5-4928-8d2b-3de67483319c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.391577] env[62066]: DEBUG nova.network.neutron [None req-478d4566-e74c-4f71-b94d-419dc0a31446 tempest-ServerPasswordTestJSON-849130511 tempest-ServerPasswordTestJSON-849130511-project-member] [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] Successfully created port: 43f773e7-1b4c-4da1-ab93-a42f9d34f998 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 735.394440] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-baa5bdcd-40fe-44d3-a1d1-ec5678eea955 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.427760] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-299684bd-90a3-4d66-8806-c09a722be23e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.436338] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8543a8e1-1b3f-4054-b7de-7a41a110b6e1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.450523] env[62066]: DEBUG nova.compute.provider_tree [None req-5b5d7867-70eb-472d-8d1a-c2bb035ec6db tempest-InstanceActionsV221TestJSON-1098427857 tempest-InstanceActionsV221TestJSON-1098427857-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 735.454464] env[62066]: DEBUG nova.network.neutron [None req-e9793203-98b3-4ca0-8a49-79767b1f52f7 tempest-ServersTestBootFromVolume-1535558002 tempest-ServersTestBootFromVolume-1535558002-project-member] [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 735.572111] env[62066]: DEBUG nova.compute.manager [None req-478d4566-e74c-4f71-b94d-419dc0a31446 tempest-ServerPasswordTestJSON-849130511 tempest-ServerPasswordTestJSON-849130511-project-member] [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 735.587722] env[62066]: DEBUG nova.network.neutron [None req-e9793203-98b3-4ca0-8a49-79767b1f52f7 tempest-ServersTestBootFromVolume-1535558002 tempest-ServersTestBootFromVolume-1535558002-project-member] [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 735.846679] env[62066]: DEBUG oslo_concurrency.lockutils [None req-8d368cc7-6ac0-44cd-b71e-9136d8337cb4 tempest-AttachInterfacesUnderV243Test-1832696447 tempest-AttachInterfacesUnderV243Test-1832696447-project-member] Lock "f91a90f7-be73-424c-966c-1be6f37a0864" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 170.510s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 735.956144] env[62066]: DEBUG nova.scheduler.client.report [None req-5b5d7867-70eb-472d-8d1a-c2bb035ec6db tempest-InstanceActionsV221TestJSON-1098427857 tempest-InstanceActionsV221TestJSON-1098427857-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 736.091595] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e9793203-98b3-4ca0-8a49-79767b1f52f7 tempest-ServersTestBootFromVolume-1535558002 tempest-ServersTestBootFromVolume-1535558002-project-member] Releasing lock "refresh_cache-2b81d060-ecb7-4ae5-b51d-33197937a76f" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 736.092186] env[62066]: DEBUG nova.compute.manager [None req-e9793203-98b3-4ca0-8a49-79767b1f52f7 tempest-ServersTestBootFromVolume-1535558002 tempest-ServersTestBootFromVolume-1535558002-project-member] [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 736.092503] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f02b141f-4767-467d-9193-05f6a2dd59bc {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.109161] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a80da3e-9eda-4eb8-a25c-77832aacd7c3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.143237] env[62066]: WARNING nova.virt.vmwareapi.driver [None req-e9793203-98b3-4ca0-8a49-79767b1f52f7 tempest-ServersTestBootFromVolume-1535558002 tempest-ServersTestBootFromVolume-1535558002-project-member] [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] Instance does not exists. Proceeding to delete instance properties on datastore: nova.exception.InstanceNotFound: Instance 2b81d060-ecb7-4ae5-b51d-33197937a76f could not be found. [ 736.143475] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-e9793203-98b3-4ca0-8a49-79767b1f52f7 tempest-ServersTestBootFromVolume-1535558002 tempest-ServersTestBootFromVolume-1535558002-project-member] [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 736.143753] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8ae7172a-14c2-463a-84a9-6ae97055810b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.153620] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-725ed389-6122-42ac-92c8-d350b2f3b526 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.175507] env[62066]: WARNING nova.virt.vmwareapi.vmops [None req-e9793203-98b3-4ca0-8a49-79767b1f52f7 tempest-ServersTestBootFromVolume-1535558002 tempest-ServersTestBootFromVolume-1535558002-project-member] [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 2b81d060-ecb7-4ae5-b51d-33197937a76f could not be found. [ 736.175822] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-e9793203-98b3-4ca0-8a49-79767b1f52f7 tempest-ServersTestBootFromVolume-1535558002 tempest-ServersTestBootFromVolume-1535558002-project-member] [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 736.176026] env[62066]: INFO nova.compute.manager [None req-e9793203-98b3-4ca0-8a49-79767b1f52f7 tempest-ServersTestBootFromVolume-1535558002 tempest-ServersTestBootFromVolume-1535558002-project-member] [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] Took 0.08 seconds to destroy the instance on the hypervisor. [ 736.176273] env[62066]: DEBUG oslo.service.loopingcall [None req-e9793203-98b3-4ca0-8a49-79767b1f52f7 tempest-ServersTestBootFromVolume-1535558002 tempest-ServersTestBootFromVolume-1535558002-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 736.176500] env[62066]: DEBUG nova.compute.manager [-] [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 736.176599] env[62066]: DEBUG nova.network.neutron [-] [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 736.214178] env[62066]: DEBUG nova.network.neutron [-] [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 736.285832] env[62066]: DEBUG nova.compute.manager [req-134ce58b-11cd-40a6-8168-5922c6884f3a req-964f84cd-82a8-43bc-9db2-9458f9e8099b service nova] [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] Received event network-changed-43f773e7-1b4c-4da1-ab93-a42f9d34f998 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 736.286048] env[62066]: DEBUG nova.compute.manager [req-134ce58b-11cd-40a6-8168-5922c6884f3a req-964f84cd-82a8-43bc-9db2-9458f9e8099b service nova] [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] Refreshing instance network info cache due to event network-changed-43f773e7-1b4c-4da1-ab93-a42f9d34f998. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 736.286269] env[62066]: DEBUG oslo_concurrency.lockutils [req-134ce58b-11cd-40a6-8168-5922c6884f3a req-964f84cd-82a8-43bc-9db2-9458f9e8099b service nova] Acquiring lock "refresh_cache-04d25af4-0e73-4650-9c3d-85817754bac9" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 736.286481] env[62066]: DEBUG oslo_concurrency.lockutils [req-134ce58b-11cd-40a6-8168-5922c6884f3a req-964f84cd-82a8-43bc-9db2-9458f9e8099b service nova] Acquired lock "refresh_cache-04d25af4-0e73-4650-9c3d-85817754bac9" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 736.286601] env[62066]: DEBUG nova.network.neutron [req-134ce58b-11cd-40a6-8168-5922c6884f3a req-964f84cd-82a8-43bc-9db2-9458f9e8099b service nova] [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] Refreshing network info cache for port 43f773e7-1b4c-4da1-ab93-a42f9d34f998 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 736.349335] env[62066]: DEBUG nova.compute.manager [None req-7fc7916e-6e82-4e11-a2d0-03a259fc6439 tempest-ServerAddressesNegativeTestJSON-1804754632 tempest-ServerAddressesNegativeTestJSON-1804754632-project-member] [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 736.425551] env[62066]: ERROR nova.compute.manager [None req-478d4566-e74c-4f71-b94d-419dc0a31446 tempest-ServerPasswordTestJSON-849130511 tempest-ServerPasswordTestJSON-849130511-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 43f773e7-1b4c-4da1-ab93-a42f9d34f998, please check neutron logs for more information. [ 736.425551] env[62066]: ERROR nova.compute.manager Traceback (most recent call last): [ 736.425551] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 736.425551] env[62066]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 736.425551] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 736.425551] env[62066]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 736.425551] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 736.425551] env[62066]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 736.425551] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 736.425551] env[62066]: ERROR nova.compute.manager self.force_reraise() [ 736.425551] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 736.425551] env[62066]: ERROR nova.compute.manager raise self.value [ 736.425551] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 736.425551] env[62066]: ERROR nova.compute.manager updated_port = self._update_port( [ 736.425551] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 736.425551] env[62066]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 736.426014] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 736.426014] env[62066]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 736.426014] env[62066]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 43f773e7-1b4c-4da1-ab93-a42f9d34f998, please check neutron logs for more information. [ 736.426014] env[62066]: ERROR nova.compute.manager [ 736.426014] env[62066]: Traceback (most recent call last): [ 736.426014] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 736.426014] env[62066]: listener.cb(fileno) [ 736.426014] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 736.426014] env[62066]: result = function(*args, **kwargs) [ 736.426014] env[62066]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 736.426014] env[62066]: return func(*args, **kwargs) [ 736.426014] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 736.426014] env[62066]: raise e [ 736.426014] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 736.426014] env[62066]: nwinfo = self.network_api.allocate_for_instance( [ 736.426014] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 736.426014] env[62066]: created_port_ids = self._update_ports_for_instance( [ 736.426014] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 736.426014] env[62066]: with excutils.save_and_reraise_exception(): [ 736.426014] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 736.426014] env[62066]: self.force_reraise() [ 736.426014] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 736.426014] env[62066]: raise self.value [ 736.426014] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 736.426014] env[62066]: updated_port = self._update_port( [ 736.426014] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 736.426014] env[62066]: _ensure_no_port_binding_failure(port) [ 736.426014] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 736.426014] env[62066]: raise exception.PortBindingFailed(port_id=port['id']) [ 736.426645] env[62066]: nova.exception.PortBindingFailed: Binding failed for port 43f773e7-1b4c-4da1-ab93-a42f9d34f998, please check neutron logs for more information. [ 736.426645] env[62066]: Removing descriptor: 15 [ 736.463791] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5b5d7867-70eb-472d-8d1a-c2bb035ec6db tempest-InstanceActionsV221TestJSON-1098427857 tempest-InstanceActionsV221TestJSON-1098427857-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.907s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 736.464458] env[62066]: ERROR nova.compute.manager [None req-5b5d7867-70eb-472d-8d1a-c2bb035ec6db tempest-InstanceActionsV221TestJSON-1098427857 tempest-InstanceActionsV221TestJSON-1098427857-project-member] [instance: d7de90c0-95e9-4899-92a0-81658831713a] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 389df780-4dc9-49a5-b7d6-c0cbbcd46c35, please check neutron logs for more information. [ 736.464458] env[62066]: ERROR nova.compute.manager [instance: d7de90c0-95e9-4899-92a0-81658831713a] Traceback (most recent call last): [ 736.464458] env[62066]: ERROR nova.compute.manager [instance: d7de90c0-95e9-4899-92a0-81658831713a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 736.464458] env[62066]: ERROR nova.compute.manager [instance: d7de90c0-95e9-4899-92a0-81658831713a] self.driver.spawn(context, instance, image_meta, [ 736.464458] env[62066]: ERROR nova.compute.manager [instance: d7de90c0-95e9-4899-92a0-81658831713a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 736.464458] env[62066]: ERROR nova.compute.manager [instance: d7de90c0-95e9-4899-92a0-81658831713a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 736.464458] env[62066]: ERROR nova.compute.manager [instance: d7de90c0-95e9-4899-92a0-81658831713a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 736.464458] env[62066]: ERROR nova.compute.manager [instance: d7de90c0-95e9-4899-92a0-81658831713a] vm_ref = self.build_virtual_machine(instance, [ 736.464458] env[62066]: ERROR nova.compute.manager [instance: d7de90c0-95e9-4899-92a0-81658831713a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 736.464458] env[62066]: ERROR nova.compute.manager [instance: d7de90c0-95e9-4899-92a0-81658831713a] vif_infos = vmwarevif.get_vif_info(self._session, [ 736.464458] env[62066]: ERROR nova.compute.manager [instance: d7de90c0-95e9-4899-92a0-81658831713a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 736.464832] env[62066]: ERROR nova.compute.manager [instance: d7de90c0-95e9-4899-92a0-81658831713a] for vif in network_info: [ 736.464832] env[62066]: ERROR nova.compute.manager [instance: d7de90c0-95e9-4899-92a0-81658831713a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 736.464832] env[62066]: ERROR nova.compute.manager [instance: d7de90c0-95e9-4899-92a0-81658831713a] return self._sync_wrapper(fn, *args, **kwargs) [ 736.464832] env[62066]: ERROR nova.compute.manager [instance: d7de90c0-95e9-4899-92a0-81658831713a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 736.464832] env[62066]: ERROR nova.compute.manager [instance: d7de90c0-95e9-4899-92a0-81658831713a] self.wait() [ 736.464832] env[62066]: ERROR nova.compute.manager [instance: d7de90c0-95e9-4899-92a0-81658831713a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 736.464832] env[62066]: ERROR nova.compute.manager [instance: d7de90c0-95e9-4899-92a0-81658831713a] self[:] = self._gt.wait() [ 736.464832] env[62066]: ERROR nova.compute.manager [instance: d7de90c0-95e9-4899-92a0-81658831713a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 736.464832] env[62066]: ERROR nova.compute.manager [instance: d7de90c0-95e9-4899-92a0-81658831713a] return self._exit_event.wait() [ 736.464832] env[62066]: ERROR nova.compute.manager [instance: d7de90c0-95e9-4899-92a0-81658831713a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 736.464832] env[62066]: ERROR nova.compute.manager [instance: d7de90c0-95e9-4899-92a0-81658831713a] result = hub.switch() [ 736.464832] env[62066]: ERROR nova.compute.manager [instance: d7de90c0-95e9-4899-92a0-81658831713a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 736.464832] env[62066]: ERROR nova.compute.manager [instance: d7de90c0-95e9-4899-92a0-81658831713a] return self.greenlet.switch() [ 736.465146] env[62066]: ERROR nova.compute.manager [instance: d7de90c0-95e9-4899-92a0-81658831713a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 736.465146] env[62066]: ERROR nova.compute.manager [instance: d7de90c0-95e9-4899-92a0-81658831713a] result = function(*args, **kwargs) [ 736.465146] env[62066]: ERROR nova.compute.manager [instance: d7de90c0-95e9-4899-92a0-81658831713a] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 736.465146] env[62066]: ERROR nova.compute.manager [instance: d7de90c0-95e9-4899-92a0-81658831713a] return func(*args, **kwargs) [ 736.465146] env[62066]: ERROR nova.compute.manager [instance: d7de90c0-95e9-4899-92a0-81658831713a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 736.465146] env[62066]: ERROR nova.compute.manager [instance: d7de90c0-95e9-4899-92a0-81658831713a] raise e [ 736.465146] env[62066]: ERROR nova.compute.manager [instance: d7de90c0-95e9-4899-92a0-81658831713a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 736.465146] env[62066]: ERROR nova.compute.manager [instance: d7de90c0-95e9-4899-92a0-81658831713a] nwinfo = self.network_api.allocate_for_instance( [ 736.465146] env[62066]: ERROR nova.compute.manager [instance: d7de90c0-95e9-4899-92a0-81658831713a] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 736.465146] env[62066]: ERROR nova.compute.manager [instance: d7de90c0-95e9-4899-92a0-81658831713a] created_port_ids = self._update_ports_for_instance( [ 736.465146] env[62066]: ERROR nova.compute.manager [instance: d7de90c0-95e9-4899-92a0-81658831713a] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 736.465146] env[62066]: ERROR nova.compute.manager [instance: d7de90c0-95e9-4899-92a0-81658831713a] with excutils.save_and_reraise_exception(): [ 736.465146] env[62066]: ERROR nova.compute.manager [instance: d7de90c0-95e9-4899-92a0-81658831713a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 736.465458] env[62066]: ERROR nova.compute.manager [instance: d7de90c0-95e9-4899-92a0-81658831713a] self.force_reraise() [ 736.465458] env[62066]: ERROR nova.compute.manager [instance: d7de90c0-95e9-4899-92a0-81658831713a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 736.465458] env[62066]: ERROR nova.compute.manager [instance: d7de90c0-95e9-4899-92a0-81658831713a] raise self.value [ 736.465458] env[62066]: ERROR nova.compute.manager [instance: d7de90c0-95e9-4899-92a0-81658831713a] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 736.465458] env[62066]: ERROR nova.compute.manager [instance: d7de90c0-95e9-4899-92a0-81658831713a] updated_port = self._update_port( [ 736.465458] env[62066]: ERROR nova.compute.manager [instance: d7de90c0-95e9-4899-92a0-81658831713a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 736.465458] env[62066]: ERROR nova.compute.manager [instance: d7de90c0-95e9-4899-92a0-81658831713a] _ensure_no_port_binding_failure(port) [ 736.465458] env[62066]: ERROR nova.compute.manager [instance: d7de90c0-95e9-4899-92a0-81658831713a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 736.465458] env[62066]: ERROR nova.compute.manager [instance: d7de90c0-95e9-4899-92a0-81658831713a] raise exception.PortBindingFailed(port_id=port['id']) [ 736.465458] env[62066]: ERROR nova.compute.manager [instance: d7de90c0-95e9-4899-92a0-81658831713a] nova.exception.PortBindingFailed: Binding failed for port 389df780-4dc9-49a5-b7d6-c0cbbcd46c35, please check neutron logs for more information. [ 736.465458] env[62066]: ERROR nova.compute.manager [instance: d7de90c0-95e9-4899-92a0-81658831713a] [ 736.465742] env[62066]: DEBUG nova.compute.utils [None req-5b5d7867-70eb-472d-8d1a-c2bb035ec6db tempest-InstanceActionsV221TestJSON-1098427857 tempest-InstanceActionsV221TestJSON-1098427857-project-member] [instance: d7de90c0-95e9-4899-92a0-81658831713a] Binding failed for port 389df780-4dc9-49a5-b7d6-c0cbbcd46c35, please check neutron logs for more information. {{(pid=62066) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 736.466560] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.439s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 736.468018] env[62066]: INFO nova.compute.claims [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] [instance: d973b09f-7727-4055-b05e-123079df9ce4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 736.470849] env[62066]: DEBUG nova.compute.manager [None req-5b5d7867-70eb-472d-8d1a-c2bb035ec6db tempest-InstanceActionsV221TestJSON-1098427857 tempest-InstanceActionsV221TestJSON-1098427857-project-member] [instance: d7de90c0-95e9-4899-92a0-81658831713a] Build of instance d7de90c0-95e9-4899-92a0-81658831713a was re-scheduled: Binding failed for port 389df780-4dc9-49a5-b7d6-c0cbbcd46c35, please check neutron logs for more information. {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 736.471301] env[62066]: DEBUG nova.compute.manager [None req-5b5d7867-70eb-472d-8d1a-c2bb035ec6db tempest-InstanceActionsV221TestJSON-1098427857 tempest-InstanceActionsV221TestJSON-1098427857-project-member] [instance: d7de90c0-95e9-4899-92a0-81658831713a] Unplugging VIFs for instance {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 736.471549] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5b5d7867-70eb-472d-8d1a-c2bb035ec6db tempest-InstanceActionsV221TestJSON-1098427857 tempest-InstanceActionsV221TestJSON-1098427857-project-member] Acquiring lock "refresh_cache-d7de90c0-95e9-4899-92a0-81658831713a" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 736.471696] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5b5d7867-70eb-472d-8d1a-c2bb035ec6db tempest-InstanceActionsV221TestJSON-1098427857 tempest-InstanceActionsV221TestJSON-1098427857-project-member] Acquired lock "refresh_cache-d7de90c0-95e9-4899-92a0-81658831713a" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 736.471857] env[62066]: DEBUG nova.network.neutron [None req-5b5d7867-70eb-472d-8d1a-c2bb035ec6db tempest-InstanceActionsV221TestJSON-1098427857 tempest-InstanceActionsV221TestJSON-1098427857-project-member] [instance: d7de90c0-95e9-4899-92a0-81658831713a] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 736.584185] env[62066]: DEBUG nova.compute.manager [None req-478d4566-e74c-4f71-b94d-419dc0a31446 tempest-ServerPasswordTestJSON-849130511 tempest-ServerPasswordTestJSON-849130511-project-member] [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 736.608941] env[62066]: DEBUG nova.virt.hardware [None req-478d4566-e74c-4f71-b94d-419dc0a31446 tempest-ServerPasswordTestJSON-849130511 tempest-ServerPasswordTestJSON-849130511-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-23T13:40:41Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-23T13:40:23Z,direct_url=,disk_format='vmdk',id=50ff584c-3b50-4395-af07-3e66769bc9f7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='eb52f7069a374c61ae946f052007c6d9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-23T13:40:24Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 736.609131] env[62066]: DEBUG nova.virt.hardware [None req-478d4566-e74c-4f71-b94d-419dc0a31446 tempest-ServerPasswordTestJSON-849130511 tempest-ServerPasswordTestJSON-849130511-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 736.609294] env[62066]: DEBUG nova.virt.hardware [None req-478d4566-e74c-4f71-b94d-419dc0a31446 tempest-ServerPasswordTestJSON-849130511 tempest-ServerPasswordTestJSON-849130511-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 736.609477] env[62066]: DEBUG nova.virt.hardware [None req-478d4566-e74c-4f71-b94d-419dc0a31446 tempest-ServerPasswordTestJSON-849130511 tempest-ServerPasswordTestJSON-849130511-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 736.609638] env[62066]: DEBUG nova.virt.hardware [None req-478d4566-e74c-4f71-b94d-419dc0a31446 tempest-ServerPasswordTestJSON-849130511 tempest-ServerPasswordTestJSON-849130511-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 736.609809] env[62066]: DEBUG nova.virt.hardware [None req-478d4566-e74c-4f71-b94d-419dc0a31446 tempest-ServerPasswordTestJSON-849130511 tempest-ServerPasswordTestJSON-849130511-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 736.610173] env[62066]: DEBUG nova.virt.hardware [None req-478d4566-e74c-4f71-b94d-419dc0a31446 tempest-ServerPasswordTestJSON-849130511 tempest-ServerPasswordTestJSON-849130511-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 736.610387] env[62066]: DEBUG nova.virt.hardware [None req-478d4566-e74c-4f71-b94d-419dc0a31446 tempest-ServerPasswordTestJSON-849130511 tempest-ServerPasswordTestJSON-849130511-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 736.610565] env[62066]: DEBUG nova.virt.hardware [None req-478d4566-e74c-4f71-b94d-419dc0a31446 tempest-ServerPasswordTestJSON-849130511 tempest-ServerPasswordTestJSON-849130511-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 736.610729] env[62066]: DEBUG nova.virt.hardware [None req-478d4566-e74c-4f71-b94d-419dc0a31446 tempest-ServerPasswordTestJSON-849130511 tempest-ServerPasswordTestJSON-849130511-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 736.610906] env[62066]: DEBUG nova.virt.hardware [None req-478d4566-e74c-4f71-b94d-419dc0a31446 tempest-ServerPasswordTestJSON-849130511 tempest-ServerPasswordTestJSON-849130511-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 736.611783] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b79367c-0a93-46ff-b41a-d668af4a8f1e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.619509] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d15e7214-a1e5-4c68-a846-b0bec243e6a0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.635775] env[62066]: ERROR nova.compute.manager [None req-478d4566-e74c-4f71-b94d-419dc0a31446 tempest-ServerPasswordTestJSON-849130511 tempest-ServerPasswordTestJSON-849130511-project-member] [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 43f773e7-1b4c-4da1-ab93-a42f9d34f998, please check neutron logs for more information. [ 736.635775] env[62066]: ERROR nova.compute.manager [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] Traceback (most recent call last): [ 736.635775] env[62066]: ERROR nova.compute.manager [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 736.635775] env[62066]: ERROR nova.compute.manager [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] yield resources [ 736.635775] env[62066]: ERROR nova.compute.manager [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 736.635775] env[62066]: ERROR nova.compute.manager [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] self.driver.spawn(context, instance, image_meta, [ 736.635775] env[62066]: ERROR nova.compute.manager [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 736.635775] env[62066]: ERROR nova.compute.manager [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 736.635775] env[62066]: ERROR nova.compute.manager [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 736.635775] env[62066]: ERROR nova.compute.manager [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] vm_ref = self.build_virtual_machine(instance, [ 736.635775] env[62066]: ERROR nova.compute.manager [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 736.636168] env[62066]: ERROR nova.compute.manager [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] vif_infos = vmwarevif.get_vif_info(self._session, [ 736.636168] env[62066]: ERROR nova.compute.manager [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 736.636168] env[62066]: ERROR nova.compute.manager [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] for vif in network_info: [ 736.636168] env[62066]: ERROR nova.compute.manager [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 736.636168] env[62066]: ERROR nova.compute.manager [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] return self._sync_wrapper(fn, *args, **kwargs) [ 736.636168] env[62066]: ERROR nova.compute.manager [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 736.636168] env[62066]: ERROR nova.compute.manager [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] self.wait() [ 736.636168] env[62066]: ERROR nova.compute.manager [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 736.636168] env[62066]: ERROR nova.compute.manager [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] self[:] = self._gt.wait() [ 736.636168] env[62066]: ERROR nova.compute.manager [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 736.636168] env[62066]: ERROR nova.compute.manager [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] return self._exit_event.wait() [ 736.636168] env[62066]: ERROR nova.compute.manager [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 736.636168] env[62066]: ERROR nova.compute.manager [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] current.throw(*self._exc) [ 736.636513] env[62066]: ERROR nova.compute.manager [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 736.636513] env[62066]: ERROR nova.compute.manager [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] result = function(*args, **kwargs) [ 736.636513] env[62066]: ERROR nova.compute.manager [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 736.636513] env[62066]: ERROR nova.compute.manager [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] return func(*args, **kwargs) [ 736.636513] env[62066]: ERROR nova.compute.manager [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 736.636513] env[62066]: ERROR nova.compute.manager [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] raise e [ 736.636513] env[62066]: ERROR nova.compute.manager [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 736.636513] env[62066]: ERROR nova.compute.manager [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] nwinfo = self.network_api.allocate_for_instance( [ 736.636513] env[62066]: ERROR nova.compute.manager [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 736.636513] env[62066]: ERROR nova.compute.manager [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] created_port_ids = self._update_ports_for_instance( [ 736.636513] env[62066]: ERROR nova.compute.manager [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 736.636513] env[62066]: ERROR nova.compute.manager [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] with excutils.save_and_reraise_exception(): [ 736.636513] env[62066]: ERROR nova.compute.manager [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 736.636796] env[62066]: ERROR nova.compute.manager [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] self.force_reraise() [ 736.636796] env[62066]: ERROR nova.compute.manager [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 736.636796] env[62066]: ERROR nova.compute.manager [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] raise self.value [ 736.636796] env[62066]: ERROR nova.compute.manager [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 736.636796] env[62066]: ERROR nova.compute.manager [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] updated_port = self._update_port( [ 736.636796] env[62066]: ERROR nova.compute.manager [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 736.636796] env[62066]: ERROR nova.compute.manager [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] _ensure_no_port_binding_failure(port) [ 736.636796] env[62066]: ERROR nova.compute.manager [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 736.636796] env[62066]: ERROR nova.compute.manager [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] raise exception.PortBindingFailed(port_id=port['id']) [ 736.636796] env[62066]: ERROR nova.compute.manager [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] nova.exception.PortBindingFailed: Binding failed for port 43f773e7-1b4c-4da1-ab93-a42f9d34f998, please check neutron logs for more information. [ 736.636796] env[62066]: ERROR nova.compute.manager [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] [ 736.636796] env[62066]: INFO nova.compute.manager [None req-478d4566-e74c-4f71-b94d-419dc0a31446 tempest-ServerPasswordTestJSON-849130511 tempest-ServerPasswordTestJSON-849130511-project-member] [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] Terminating instance [ 736.637099] env[62066]: DEBUG oslo_concurrency.lockutils [None req-478d4566-e74c-4f71-b94d-419dc0a31446 tempest-ServerPasswordTestJSON-849130511 tempest-ServerPasswordTestJSON-849130511-project-member] Acquiring lock "refresh_cache-04d25af4-0e73-4650-9c3d-85817754bac9" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 736.717307] env[62066]: DEBUG nova.network.neutron [-] [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 736.810552] env[62066]: DEBUG nova.network.neutron [req-134ce58b-11cd-40a6-8168-5922c6884f3a req-964f84cd-82a8-43bc-9db2-9458f9e8099b service nova] [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 736.879195] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7fc7916e-6e82-4e11-a2d0-03a259fc6439 tempest-ServerAddressesNegativeTestJSON-1804754632 tempest-ServerAddressesNegativeTestJSON-1804754632-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 736.943447] env[62066]: DEBUG nova.network.neutron [req-134ce58b-11cd-40a6-8168-5922c6884f3a req-964f84cd-82a8-43bc-9db2-9458f9e8099b service nova] [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 737.000163] env[62066]: DEBUG nova.network.neutron [None req-5b5d7867-70eb-472d-8d1a-c2bb035ec6db tempest-InstanceActionsV221TestJSON-1098427857 tempest-InstanceActionsV221TestJSON-1098427857-project-member] [instance: d7de90c0-95e9-4899-92a0-81658831713a] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 737.130155] env[62066]: DEBUG nova.network.neutron [None req-5b5d7867-70eb-472d-8d1a-c2bb035ec6db tempest-InstanceActionsV221TestJSON-1098427857 tempest-InstanceActionsV221TestJSON-1098427857-project-member] [instance: d7de90c0-95e9-4899-92a0-81658831713a] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 737.220271] env[62066]: INFO nova.compute.manager [-] [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] Took 1.04 seconds to deallocate network for instance. [ 737.451182] env[62066]: DEBUG oslo_concurrency.lockutils [req-134ce58b-11cd-40a6-8168-5922c6884f3a req-964f84cd-82a8-43bc-9db2-9458f9e8099b service nova] Releasing lock "refresh_cache-04d25af4-0e73-4650-9c3d-85817754bac9" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 737.451182] env[62066]: DEBUG oslo_concurrency.lockutils [None req-478d4566-e74c-4f71-b94d-419dc0a31446 tempest-ServerPasswordTestJSON-849130511 tempest-ServerPasswordTestJSON-849130511-project-member] Acquired lock "refresh_cache-04d25af4-0e73-4650-9c3d-85817754bac9" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 737.451182] env[62066]: DEBUG nova.network.neutron [None req-478d4566-e74c-4f71-b94d-419dc0a31446 tempest-ServerPasswordTestJSON-849130511 tempest-ServerPasswordTestJSON-849130511-project-member] [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 737.635500] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5b5d7867-70eb-472d-8d1a-c2bb035ec6db tempest-InstanceActionsV221TestJSON-1098427857 tempest-InstanceActionsV221TestJSON-1098427857-project-member] Releasing lock "refresh_cache-d7de90c0-95e9-4899-92a0-81658831713a" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 737.635756] env[62066]: DEBUG nova.compute.manager [None req-5b5d7867-70eb-472d-8d1a-c2bb035ec6db tempest-InstanceActionsV221TestJSON-1098427857 tempest-InstanceActionsV221TestJSON-1098427857-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 737.635905] env[62066]: DEBUG nova.compute.manager [None req-5b5d7867-70eb-472d-8d1a-c2bb035ec6db tempest-InstanceActionsV221TestJSON-1098427857 tempest-InstanceActionsV221TestJSON-1098427857-project-member] [instance: d7de90c0-95e9-4899-92a0-81658831713a] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 737.636819] env[62066]: DEBUG nova.network.neutron [None req-5b5d7867-70eb-472d-8d1a-c2bb035ec6db tempest-InstanceActionsV221TestJSON-1098427857 tempest-InstanceActionsV221TestJSON-1098427857-project-member] [instance: d7de90c0-95e9-4899-92a0-81658831713a] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 737.658606] env[62066]: DEBUG nova.network.neutron [None req-5b5d7867-70eb-472d-8d1a-c2bb035ec6db tempest-InstanceActionsV221TestJSON-1098427857 tempest-InstanceActionsV221TestJSON-1098427857-project-member] [instance: d7de90c0-95e9-4899-92a0-81658831713a] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 737.793977] env[62066]: INFO nova.compute.manager [None req-e9793203-98b3-4ca0-8a49-79767b1f52f7 tempest-ServersTestBootFromVolume-1535558002 tempest-ServersTestBootFromVolume-1535558002-project-member] [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] Took 0.57 seconds to detach 1 volumes for instance. [ 737.796497] env[62066]: DEBUG nova.compute.claims [None req-e9793203-98b3-4ca0-8a49-79767b1f52f7 tempest-ServersTestBootFromVolume-1535558002 tempest-ServersTestBootFromVolume-1535558002-project-member] [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] Aborting claim: {{(pid=62066) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 737.796683] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e9793203-98b3-4ca0-8a49-79767b1f52f7 tempest-ServersTestBootFromVolume-1535558002 tempest-ServersTestBootFromVolume-1535558002-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 737.866371] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbe7b85b-d48c-4654-b50a-171b9ca0d6e7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.873468] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-557b0ebb-9328-42bb-bf86-8a8aff62fc7e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.911106] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d0ee7fe-bf7c-4b82-8d5b-0ff107167e34 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.918833] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ca5d504-9d09-4b06-8263-3d4e121b2fb3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.932604] env[62066]: DEBUG nova.compute.provider_tree [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 737.968598] env[62066]: DEBUG nova.network.neutron [None req-478d4566-e74c-4f71-b94d-419dc0a31446 tempest-ServerPasswordTestJSON-849130511 tempest-ServerPasswordTestJSON-849130511-project-member] [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 738.047040] env[62066]: DEBUG nova.network.neutron [None req-478d4566-e74c-4f71-b94d-419dc0a31446 tempest-ServerPasswordTestJSON-849130511 tempest-ServerPasswordTestJSON-849130511-project-member] [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 738.164640] env[62066]: DEBUG nova.network.neutron [None req-5b5d7867-70eb-472d-8d1a-c2bb035ec6db tempest-InstanceActionsV221TestJSON-1098427857 tempest-InstanceActionsV221TestJSON-1098427857-project-member] [instance: d7de90c0-95e9-4899-92a0-81658831713a] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 738.321936] env[62066]: DEBUG nova.compute.manager [req-48c1e4fc-8e4e-424f-92f6-573416d3e00c req-c7a1759b-0230-49e5-9180-4d49f7a57393 service nova] [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] Received event network-vif-deleted-43f773e7-1b4c-4da1-ab93-a42f9d34f998 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 738.435841] env[62066]: DEBUG nova.scheduler.client.report [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 738.555830] env[62066]: DEBUG oslo_concurrency.lockutils [None req-478d4566-e74c-4f71-b94d-419dc0a31446 tempest-ServerPasswordTestJSON-849130511 tempest-ServerPasswordTestJSON-849130511-project-member] Releasing lock "refresh_cache-04d25af4-0e73-4650-9c3d-85817754bac9" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 738.558461] env[62066]: DEBUG nova.compute.manager [None req-478d4566-e74c-4f71-b94d-419dc0a31446 tempest-ServerPasswordTestJSON-849130511 tempest-ServerPasswordTestJSON-849130511-project-member] [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 738.558461] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-478d4566-e74c-4f71-b94d-419dc0a31446 tempest-ServerPasswordTestJSON-849130511 tempest-ServerPasswordTestJSON-849130511-project-member] [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 738.558461] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4a886c04-f17d-48a5-ad06-4b13eb05161f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.567719] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e671dd6-0ee3-43df-b89c-205dc9c05842 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.591992] env[62066]: WARNING nova.virt.vmwareapi.vmops [None req-478d4566-e74c-4f71-b94d-419dc0a31446 tempest-ServerPasswordTestJSON-849130511 tempest-ServerPasswordTestJSON-849130511-project-member] [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 04d25af4-0e73-4650-9c3d-85817754bac9 could not be found. [ 738.591992] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-478d4566-e74c-4f71-b94d-419dc0a31446 tempest-ServerPasswordTestJSON-849130511 tempest-ServerPasswordTestJSON-849130511-project-member] [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 738.592184] env[62066]: INFO nova.compute.manager [None req-478d4566-e74c-4f71-b94d-419dc0a31446 tempest-ServerPasswordTestJSON-849130511 tempest-ServerPasswordTestJSON-849130511-project-member] [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] Took 0.03 seconds to destroy the instance on the hypervisor. [ 738.592392] env[62066]: DEBUG oslo.service.loopingcall [None req-478d4566-e74c-4f71-b94d-419dc0a31446 tempest-ServerPasswordTestJSON-849130511 tempest-ServerPasswordTestJSON-849130511-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 738.592638] env[62066]: DEBUG nova.compute.manager [-] [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 738.592845] env[62066]: DEBUG nova.network.neutron [-] [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 738.608201] env[62066]: DEBUG nova.network.neutron [-] [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 738.667599] env[62066]: INFO nova.compute.manager [None req-5b5d7867-70eb-472d-8d1a-c2bb035ec6db tempest-InstanceActionsV221TestJSON-1098427857 tempest-InstanceActionsV221TestJSON-1098427857-project-member] [instance: d7de90c0-95e9-4899-92a0-81658831713a] Took 1.03 seconds to deallocate network for instance. [ 738.944023] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.475s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 738.944023] env[62066]: DEBUG nova.compute.manager [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] [instance: d973b09f-7727-4055-b05e-123079df9ce4] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 738.945183] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.822s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 738.946622] env[62066]: INFO nova.compute.claims [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 739.111047] env[62066]: DEBUG nova.network.neutron [-] [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 739.451575] env[62066]: DEBUG nova.compute.utils [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 739.455730] env[62066]: DEBUG nova.compute.manager [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] [instance: d973b09f-7727-4055-b05e-123079df9ce4] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 739.455730] env[62066]: DEBUG nova.network.neutron [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] [instance: d973b09f-7727-4055-b05e-123079df9ce4] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 739.499977] env[62066]: DEBUG nova.policy [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5ae0f813ad034e8988e31d93da7894cc', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0545f648e24c455ab6fd7ae1ff1a0214', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 739.614728] env[62066]: INFO nova.compute.manager [-] [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] Took 1.02 seconds to deallocate network for instance. [ 739.617161] env[62066]: DEBUG nova.compute.claims [None req-478d4566-e74c-4f71-b94d-419dc0a31446 tempest-ServerPasswordTestJSON-849130511 tempest-ServerPasswordTestJSON-849130511-project-member] [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] Aborting claim: {{(pid=62066) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 739.617426] env[62066]: DEBUG oslo_concurrency.lockutils [None req-478d4566-e74c-4f71-b94d-419dc0a31446 tempest-ServerPasswordTestJSON-849130511 tempest-ServerPasswordTestJSON-849130511-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 739.705256] env[62066]: INFO nova.scheduler.client.report [None req-5b5d7867-70eb-472d-8d1a-c2bb035ec6db tempest-InstanceActionsV221TestJSON-1098427857 tempest-InstanceActionsV221TestJSON-1098427857-project-member] Deleted allocations for instance d7de90c0-95e9-4899-92a0-81658831713a [ 739.885571] env[62066]: DEBUG nova.network.neutron [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] [instance: d973b09f-7727-4055-b05e-123079df9ce4] Successfully created port: 72daca54-f990-4f16-9362-cccaded8dba7 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 739.956890] env[62066]: DEBUG nova.compute.manager [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] [instance: d973b09f-7727-4055-b05e-123079df9ce4] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 740.223768] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5b5d7867-70eb-472d-8d1a-c2bb035ec6db tempest-InstanceActionsV221TestJSON-1098427857 tempest-InstanceActionsV221TestJSON-1098427857-project-member] Lock "d7de90c0-95e9-4899-92a0-81658831713a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 167.594s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 740.357021] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95a80de5-3d40-4998-892e-1e98def50f0c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.367036] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a676a59-5253-451b-9bd4-108c7ea680be {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.402034] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-140975c8-4ab4-4647-99c4-5d3b815d1ce8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.410499] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fdf910cc-9067-48ac-9239-5682466eb7fb {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.424769] env[62066]: DEBUG nova.compute.provider_tree [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 740.429904] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3d67ea1c-5f75-4620-b677-6f1e0faf1bd6 tempest-ServersV294TestFqdnHostnames-843300434 tempest-ServersV294TestFqdnHostnames-843300434-project-member] Acquiring lock "2709d0ad-5dd8-4e3c-b1e6-3d1e2e242380" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 740.430142] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3d67ea1c-5f75-4620-b677-6f1e0faf1bd6 tempest-ServersV294TestFqdnHostnames-843300434 tempest-ServersV294TestFqdnHostnames-843300434-project-member] Lock "2709d0ad-5dd8-4e3c-b1e6-3d1e2e242380" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 740.730100] env[62066]: DEBUG nova.compute.manager [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 0ebba3b2-f82d-4c1d-b01a-f75c3559c117] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 740.931453] env[62066]: DEBUG nova.scheduler.client.report [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 740.976112] env[62066]: DEBUG nova.compute.manager [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] [instance: d973b09f-7727-4055-b05e-123079df9ce4] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 741.005856] env[62066]: DEBUG nova.virt.hardware [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-23T13:40:41Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-23T13:40:23Z,direct_url=,disk_format='vmdk',id=50ff584c-3b50-4395-af07-3e66769bc9f7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='eb52f7069a374c61ae946f052007c6d9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-23T13:40:24Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 741.006144] env[62066]: DEBUG nova.virt.hardware [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 741.006351] env[62066]: DEBUG nova.virt.hardware [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 741.006545] env[62066]: DEBUG nova.virt.hardware [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 741.006692] env[62066]: DEBUG nova.virt.hardware [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 741.006835] env[62066]: DEBUG nova.virt.hardware [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 741.007396] env[62066]: DEBUG nova.virt.hardware [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 741.007837] env[62066]: DEBUG nova.virt.hardware [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 741.007837] env[62066]: DEBUG nova.virt.hardware [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 741.008140] env[62066]: DEBUG nova.virt.hardware [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 741.008393] env[62066]: DEBUG nova.virt.hardware [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 741.009742] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da84e7b2-13e3-4968-abca-a580164adad2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.020035] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dac18c80-01c6-4974-8693-ef9c25be259e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.256836] env[62066]: DEBUG oslo_concurrency.lockutils [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 741.439171] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.492s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 741.439171] env[62066]: DEBUG nova.compute.manager [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 741.440552] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5bf8bcd1-f041-4648-9b3c-4febff568fa9 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.668s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 741.491397] env[62066]: DEBUG nova.compute.manager [req-35f3bf7a-84d8-4b7c-ad4b-76aefa90020e req-0d54b4df-f016-4ff6-b858-29397dff847a service nova] [instance: d973b09f-7727-4055-b05e-123079df9ce4] Received event network-changed-72daca54-f990-4f16-9362-cccaded8dba7 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 741.491602] env[62066]: DEBUG nova.compute.manager [req-35f3bf7a-84d8-4b7c-ad4b-76aefa90020e req-0d54b4df-f016-4ff6-b858-29397dff847a service nova] [instance: d973b09f-7727-4055-b05e-123079df9ce4] Refreshing instance network info cache due to event network-changed-72daca54-f990-4f16-9362-cccaded8dba7. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 741.491812] env[62066]: DEBUG oslo_concurrency.lockutils [req-35f3bf7a-84d8-4b7c-ad4b-76aefa90020e req-0d54b4df-f016-4ff6-b858-29397dff847a service nova] Acquiring lock "refresh_cache-d973b09f-7727-4055-b05e-123079df9ce4" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 741.491958] env[62066]: DEBUG oslo_concurrency.lockutils [req-35f3bf7a-84d8-4b7c-ad4b-76aefa90020e req-0d54b4df-f016-4ff6-b858-29397dff847a service nova] Acquired lock "refresh_cache-d973b09f-7727-4055-b05e-123079df9ce4" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 741.492696] env[62066]: DEBUG nova.network.neutron [req-35f3bf7a-84d8-4b7c-ad4b-76aefa90020e req-0d54b4df-f016-4ff6-b858-29397dff847a service nova] [instance: d973b09f-7727-4055-b05e-123079df9ce4] Refreshing network info cache for port 72daca54-f990-4f16-9362-cccaded8dba7 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 741.707173] env[62066]: ERROR nova.compute.manager [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 72daca54-f990-4f16-9362-cccaded8dba7, please check neutron logs for more information. [ 741.707173] env[62066]: ERROR nova.compute.manager Traceback (most recent call last): [ 741.707173] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 741.707173] env[62066]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 741.707173] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 741.707173] env[62066]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 741.707173] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 741.707173] env[62066]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 741.707173] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 741.707173] env[62066]: ERROR nova.compute.manager self.force_reraise() [ 741.707173] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 741.707173] env[62066]: ERROR nova.compute.manager raise self.value [ 741.707173] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 741.707173] env[62066]: ERROR nova.compute.manager updated_port = self._update_port( [ 741.707173] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 741.707173] env[62066]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 741.707562] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 741.707562] env[62066]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 741.707562] env[62066]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 72daca54-f990-4f16-9362-cccaded8dba7, please check neutron logs for more information. [ 741.707562] env[62066]: ERROR nova.compute.manager [ 741.707562] env[62066]: Traceback (most recent call last): [ 741.707562] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 741.707562] env[62066]: listener.cb(fileno) [ 741.707562] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 741.707562] env[62066]: result = function(*args, **kwargs) [ 741.707562] env[62066]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 741.707562] env[62066]: return func(*args, **kwargs) [ 741.707562] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 741.707562] env[62066]: raise e [ 741.707562] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 741.707562] env[62066]: nwinfo = self.network_api.allocate_for_instance( [ 741.707562] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 741.707562] env[62066]: created_port_ids = self._update_ports_for_instance( [ 741.707562] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 741.707562] env[62066]: with excutils.save_and_reraise_exception(): [ 741.707562] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 741.707562] env[62066]: self.force_reraise() [ 741.707562] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 741.707562] env[62066]: raise self.value [ 741.707562] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 741.707562] env[62066]: updated_port = self._update_port( [ 741.707562] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 741.707562] env[62066]: _ensure_no_port_binding_failure(port) [ 741.707562] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 741.707562] env[62066]: raise exception.PortBindingFailed(port_id=port['id']) [ 741.708300] env[62066]: nova.exception.PortBindingFailed: Binding failed for port 72daca54-f990-4f16-9362-cccaded8dba7, please check neutron logs for more information. [ 741.708300] env[62066]: Removing descriptor: 15 [ 741.708300] env[62066]: ERROR nova.compute.manager [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] [instance: d973b09f-7727-4055-b05e-123079df9ce4] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 72daca54-f990-4f16-9362-cccaded8dba7, please check neutron logs for more information. [ 741.708300] env[62066]: ERROR nova.compute.manager [instance: d973b09f-7727-4055-b05e-123079df9ce4] Traceback (most recent call last): [ 741.708300] env[62066]: ERROR nova.compute.manager [instance: d973b09f-7727-4055-b05e-123079df9ce4] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 741.708300] env[62066]: ERROR nova.compute.manager [instance: d973b09f-7727-4055-b05e-123079df9ce4] yield resources [ 741.708300] env[62066]: ERROR nova.compute.manager [instance: d973b09f-7727-4055-b05e-123079df9ce4] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 741.708300] env[62066]: ERROR nova.compute.manager [instance: d973b09f-7727-4055-b05e-123079df9ce4] self.driver.spawn(context, instance, image_meta, [ 741.708300] env[62066]: ERROR nova.compute.manager [instance: d973b09f-7727-4055-b05e-123079df9ce4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 741.708300] env[62066]: ERROR nova.compute.manager [instance: d973b09f-7727-4055-b05e-123079df9ce4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 741.708300] env[62066]: ERROR nova.compute.manager [instance: d973b09f-7727-4055-b05e-123079df9ce4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 741.708300] env[62066]: ERROR nova.compute.manager [instance: d973b09f-7727-4055-b05e-123079df9ce4] vm_ref = self.build_virtual_machine(instance, [ 741.708689] env[62066]: ERROR nova.compute.manager [instance: d973b09f-7727-4055-b05e-123079df9ce4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 741.708689] env[62066]: ERROR nova.compute.manager [instance: d973b09f-7727-4055-b05e-123079df9ce4] vif_infos = vmwarevif.get_vif_info(self._session, [ 741.708689] env[62066]: ERROR nova.compute.manager [instance: d973b09f-7727-4055-b05e-123079df9ce4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 741.708689] env[62066]: ERROR nova.compute.manager [instance: d973b09f-7727-4055-b05e-123079df9ce4] for vif in network_info: [ 741.708689] env[62066]: ERROR nova.compute.manager [instance: d973b09f-7727-4055-b05e-123079df9ce4] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 741.708689] env[62066]: ERROR nova.compute.manager [instance: d973b09f-7727-4055-b05e-123079df9ce4] return self._sync_wrapper(fn, *args, **kwargs) [ 741.708689] env[62066]: ERROR nova.compute.manager [instance: d973b09f-7727-4055-b05e-123079df9ce4] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 741.708689] env[62066]: ERROR nova.compute.manager [instance: d973b09f-7727-4055-b05e-123079df9ce4] self.wait() [ 741.708689] env[62066]: ERROR nova.compute.manager [instance: d973b09f-7727-4055-b05e-123079df9ce4] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 741.708689] env[62066]: ERROR nova.compute.manager [instance: d973b09f-7727-4055-b05e-123079df9ce4] self[:] = self._gt.wait() [ 741.708689] env[62066]: ERROR nova.compute.manager [instance: d973b09f-7727-4055-b05e-123079df9ce4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 741.708689] env[62066]: ERROR nova.compute.manager [instance: d973b09f-7727-4055-b05e-123079df9ce4] return self._exit_event.wait() [ 741.708689] env[62066]: ERROR nova.compute.manager [instance: d973b09f-7727-4055-b05e-123079df9ce4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 741.709047] env[62066]: ERROR nova.compute.manager [instance: d973b09f-7727-4055-b05e-123079df9ce4] result = hub.switch() [ 741.709047] env[62066]: ERROR nova.compute.manager [instance: d973b09f-7727-4055-b05e-123079df9ce4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 741.709047] env[62066]: ERROR nova.compute.manager [instance: d973b09f-7727-4055-b05e-123079df9ce4] return self.greenlet.switch() [ 741.709047] env[62066]: ERROR nova.compute.manager [instance: d973b09f-7727-4055-b05e-123079df9ce4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 741.709047] env[62066]: ERROR nova.compute.manager [instance: d973b09f-7727-4055-b05e-123079df9ce4] result = function(*args, **kwargs) [ 741.709047] env[62066]: ERROR nova.compute.manager [instance: d973b09f-7727-4055-b05e-123079df9ce4] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 741.709047] env[62066]: ERROR nova.compute.manager [instance: d973b09f-7727-4055-b05e-123079df9ce4] return func(*args, **kwargs) [ 741.709047] env[62066]: ERROR nova.compute.manager [instance: d973b09f-7727-4055-b05e-123079df9ce4] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 741.709047] env[62066]: ERROR nova.compute.manager [instance: d973b09f-7727-4055-b05e-123079df9ce4] raise e [ 741.709047] env[62066]: ERROR nova.compute.manager [instance: d973b09f-7727-4055-b05e-123079df9ce4] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 741.709047] env[62066]: ERROR nova.compute.manager [instance: d973b09f-7727-4055-b05e-123079df9ce4] nwinfo = self.network_api.allocate_for_instance( [ 741.709047] env[62066]: ERROR nova.compute.manager [instance: d973b09f-7727-4055-b05e-123079df9ce4] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 741.709047] env[62066]: ERROR nova.compute.manager [instance: d973b09f-7727-4055-b05e-123079df9ce4] created_port_ids = self._update_ports_for_instance( [ 741.709372] env[62066]: ERROR nova.compute.manager [instance: d973b09f-7727-4055-b05e-123079df9ce4] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 741.709372] env[62066]: ERROR nova.compute.manager [instance: d973b09f-7727-4055-b05e-123079df9ce4] with excutils.save_and_reraise_exception(): [ 741.709372] env[62066]: ERROR nova.compute.manager [instance: d973b09f-7727-4055-b05e-123079df9ce4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 741.709372] env[62066]: ERROR nova.compute.manager [instance: d973b09f-7727-4055-b05e-123079df9ce4] self.force_reraise() [ 741.709372] env[62066]: ERROR nova.compute.manager [instance: d973b09f-7727-4055-b05e-123079df9ce4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 741.709372] env[62066]: ERROR nova.compute.manager [instance: d973b09f-7727-4055-b05e-123079df9ce4] raise self.value [ 741.709372] env[62066]: ERROR nova.compute.manager [instance: d973b09f-7727-4055-b05e-123079df9ce4] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 741.709372] env[62066]: ERROR nova.compute.manager [instance: d973b09f-7727-4055-b05e-123079df9ce4] updated_port = self._update_port( [ 741.709372] env[62066]: ERROR nova.compute.manager [instance: d973b09f-7727-4055-b05e-123079df9ce4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 741.709372] env[62066]: ERROR nova.compute.manager [instance: d973b09f-7727-4055-b05e-123079df9ce4] _ensure_no_port_binding_failure(port) [ 741.709372] env[62066]: ERROR nova.compute.manager [instance: d973b09f-7727-4055-b05e-123079df9ce4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 741.709372] env[62066]: ERROR nova.compute.manager [instance: d973b09f-7727-4055-b05e-123079df9ce4] raise exception.PortBindingFailed(port_id=port['id']) [ 741.709722] env[62066]: ERROR nova.compute.manager [instance: d973b09f-7727-4055-b05e-123079df9ce4] nova.exception.PortBindingFailed: Binding failed for port 72daca54-f990-4f16-9362-cccaded8dba7, please check neutron logs for more information. [ 741.709722] env[62066]: ERROR nova.compute.manager [instance: d973b09f-7727-4055-b05e-123079df9ce4] [ 741.709722] env[62066]: INFO nova.compute.manager [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] [instance: d973b09f-7727-4055-b05e-123079df9ce4] Terminating instance [ 741.710791] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] Acquiring lock "refresh_cache-d973b09f-7727-4055-b05e-123079df9ce4" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 741.945703] env[62066]: DEBUG nova.compute.utils [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 741.950021] env[62066]: DEBUG nova.compute.manager [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 741.950021] env[62066]: DEBUG nova.network.neutron [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 742.002308] env[62066]: DEBUG nova.policy [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5ae0f813ad034e8988e31d93da7894cc', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0545f648e24c455ab6fd7ae1ff1a0214', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 742.026019] env[62066]: DEBUG nova.network.neutron [req-35f3bf7a-84d8-4b7c-ad4b-76aefa90020e req-0d54b4df-f016-4ff6-b858-29397dff847a service nova] [instance: d973b09f-7727-4055-b05e-123079df9ce4] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 742.164144] env[62066]: DEBUG nova.network.neutron [req-35f3bf7a-84d8-4b7c-ad4b-76aefa90020e req-0d54b4df-f016-4ff6-b858-29397dff847a service nova] [instance: d973b09f-7727-4055-b05e-123079df9ce4] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 742.357994] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae12efc3-fefc-4c91-97fa-2d2250106f28 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.364898] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58b34bca-3cf2-43cb-8c92-35c32b725c36 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.401419] env[62066]: DEBUG nova.network.neutron [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] Successfully created port: a7a012fe-04f6-4d88-b008-27cf1ef05505 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 742.403932] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-daf80575-05fe-4428-a98d-264b1acb40a9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.413991] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5de831f5-923e-4601-8e60-cc1c946be61a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.431026] env[62066]: DEBUG nova.compute.provider_tree [None req-5bf8bcd1-f041-4648-9b3c-4febff568fa9 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 742.453915] env[62066]: DEBUG nova.compute.manager [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 742.667131] env[62066]: DEBUG oslo_concurrency.lockutils [req-35f3bf7a-84d8-4b7c-ad4b-76aefa90020e req-0d54b4df-f016-4ff6-b858-29397dff847a service nova] Releasing lock "refresh_cache-d973b09f-7727-4055-b05e-123079df9ce4" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 742.668032] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] Acquired lock "refresh_cache-d973b09f-7727-4055-b05e-123079df9ce4" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 742.668032] env[62066]: DEBUG nova.network.neutron [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] [instance: d973b09f-7727-4055-b05e-123079df9ce4] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 742.934041] env[62066]: DEBUG nova.scheduler.client.report [None req-5bf8bcd1-f041-4648-9b3c-4febff568fa9 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 743.176746] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b8836db2-293e-442c-80f2-42d8bc4c2918 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Acquiring lock "88daaaa6-f385-4161-bc74-9fffabb1145d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 743.176993] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b8836db2-293e-442c-80f2-42d8bc4c2918 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Lock "88daaaa6-f385-4161-bc74-9fffabb1145d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 743.193099] env[62066]: DEBUG nova.network.neutron [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] [instance: d973b09f-7727-4055-b05e-123079df9ce4] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 743.288418] env[62066]: DEBUG nova.network.neutron [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] [instance: d973b09f-7727-4055-b05e-123079df9ce4] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 743.440786] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5bf8bcd1-f041-4648-9b3c-4febff568fa9 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 743.441455] env[62066]: ERROR nova.compute.manager [None req-5bf8bcd1-f041-4648-9b3c-4febff568fa9 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] [instance: 3624ca43-9ee3-418a-9205-debb699295df] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port bf6b4191-9489-4988-8fd5-320a5a2607a1, please check neutron logs for more information. [ 743.441455] env[62066]: ERROR nova.compute.manager [instance: 3624ca43-9ee3-418a-9205-debb699295df] Traceback (most recent call last): [ 743.441455] env[62066]: ERROR nova.compute.manager [instance: 3624ca43-9ee3-418a-9205-debb699295df] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 743.441455] env[62066]: ERROR nova.compute.manager [instance: 3624ca43-9ee3-418a-9205-debb699295df] self.driver.spawn(context, instance, image_meta, [ 743.441455] env[62066]: ERROR nova.compute.manager [instance: 3624ca43-9ee3-418a-9205-debb699295df] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 743.441455] env[62066]: ERROR nova.compute.manager [instance: 3624ca43-9ee3-418a-9205-debb699295df] self._vmops.spawn(context, instance, image_meta, injected_files, [ 743.441455] env[62066]: ERROR nova.compute.manager [instance: 3624ca43-9ee3-418a-9205-debb699295df] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 743.441455] env[62066]: ERROR nova.compute.manager [instance: 3624ca43-9ee3-418a-9205-debb699295df] vm_ref = self.build_virtual_machine(instance, [ 743.441455] env[62066]: ERROR nova.compute.manager [instance: 3624ca43-9ee3-418a-9205-debb699295df] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 743.441455] env[62066]: ERROR nova.compute.manager [instance: 3624ca43-9ee3-418a-9205-debb699295df] vif_infos = vmwarevif.get_vif_info(self._session, [ 743.441455] env[62066]: ERROR nova.compute.manager [instance: 3624ca43-9ee3-418a-9205-debb699295df] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 743.441764] env[62066]: ERROR nova.compute.manager [instance: 3624ca43-9ee3-418a-9205-debb699295df] for vif in network_info: [ 743.441764] env[62066]: ERROR nova.compute.manager [instance: 3624ca43-9ee3-418a-9205-debb699295df] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 743.441764] env[62066]: ERROR nova.compute.manager [instance: 3624ca43-9ee3-418a-9205-debb699295df] return self._sync_wrapper(fn, *args, **kwargs) [ 743.441764] env[62066]: ERROR nova.compute.manager [instance: 3624ca43-9ee3-418a-9205-debb699295df] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 743.441764] env[62066]: ERROR nova.compute.manager [instance: 3624ca43-9ee3-418a-9205-debb699295df] self.wait() [ 743.441764] env[62066]: ERROR nova.compute.manager [instance: 3624ca43-9ee3-418a-9205-debb699295df] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 743.441764] env[62066]: ERROR nova.compute.manager [instance: 3624ca43-9ee3-418a-9205-debb699295df] self[:] = self._gt.wait() [ 743.441764] env[62066]: ERROR nova.compute.manager [instance: 3624ca43-9ee3-418a-9205-debb699295df] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 743.441764] env[62066]: ERROR nova.compute.manager [instance: 3624ca43-9ee3-418a-9205-debb699295df] return self._exit_event.wait() [ 743.441764] env[62066]: ERROR nova.compute.manager [instance: 3624ca43-9ee3-418a-9205-debb699295df] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 743.441764] env[62066]: ERROR nova.compute.manager [instance: 3624ca43-9ee3-418a-9205-debb699295df] current.throw(*self._exc) [ 743.441764] env[62066]: ERROR nova.compute.manager [instance: 3624ca43-9ee3-418a-9205-debb699295df] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 743.441764] env[62066]: ERROR nova.compute.manager [instance: 3624ca43-9ee3-418a-9205-debb699295df] result = function(*args, **kwargs) [ 743.442116] env[62066]: ERROR nova.compute.manager [instance: 3624ca43-9ee3-418a-9205-debb699295df] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 743.442116] env[62066]: ERROR nova.compute.manager [instance: 3624ca43-9ee3-418a-9205-debb699295df] return func(*args, **kwargs) [ 743.442116] env[62066]: ERROR nova.compute.manager [instance: 3624ca43-9ee3-418a-9205-debb699295df] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 743.442116] env[62066]: ERROR nova.compute.manager [instance: 3624ca43-9ee3-418a-9205-debb699295df] raise e [ 743.442116] env[62066]: ERROR nova.compute.manager [instance: 3624ca43-9ee3-418a-9205-debb699295df] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 743.442116] env[62066]: ERROR nova.compute.manager [instance: 3624ca43-9ee3-418a-9205-debb699295df] nwinfo = self.network_api.allocate_for_instance( [ 743.442116] env[62066]: ERROR nova.compute.manager [instance: 3624ca43-9ee3-418a-9205-debb699295df] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 743.442116] env[62066]: ERROR nova.compute.manager [instance: 3624ca43-9ee3-418a-9205-debb699295df] created_port_ids = self._update_ports_for_instance( [ 743.442116] env[62066]: ERROR nova.compute.manager [instance: 3624ca43-9ee3-418a-9205-debb699295df] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 743.442116] env[62066]: ERROR nova.compute.manager [instance: 3624ca43-9ee3-418a-9205-debb699295df] with excutils.save_and_reraise_exception(): [ 743.442116] env[62066]: ERROR nova.compute.manager [instance: 3624ca43-9ee3-418a-9205-debb699295df] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 743.442116] env[62066]: ERROR nova.compute.manager [instance: 3624ca43-9ee3-418a-9205-debb699295df] self.force_reraise() [ 743.442116] env[62066]: ERROR nova.compute.manager [instance: 3624ca43-9ee3-418a-9205-debb699295df] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 743.442461] env[62066]: ERROR nova.compute.manager [instance: 3624ca43-9ee3-418a-9205-debb699295df] raise self.value [ 743.442461] env[62066]: ERROR nova.compute.manager [instance: 3624ca43-9ee3-418a-9205-debb699295df] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 743.442461] env[62066]: ERROR nova.compute.manager [instance: 3624ca43-9ee3-418a-9205-debb699295df] updated_port = self._update_port( [ 743.442461] env[62066]: ERROR nova.compute.manager [instance: 3624ca43-9ee3-418a-9205-debb699295df] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 743.442461] env[62066]: ERROR nova.compute.manager [instance: 3624ca43-9ee3-418a-9205-debb699295df] _ensure_no_port_binding_failure(port) [ 743.442461] env[62066]: ERROR nova.compute.manager [instance: 3624ca43-9ee3-418a-9205-debb699295df] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 743.442461] env[62066]: ERROR nova.compute.manager [instance: 3624ca43-9ee3-418a-9205-debb699295df] raise exception.PortBindingFailed(port_id=port['id']) [ 743.442461] env[62066]: ERROR nova.compute.manager [instance: 3624ca43-9ee3-418a-9205-debb699295df] nova.exception.PortBindingFailed: Binding failed for port bf6b4191-9489-4988-8fd5-320a5a2607a1, please check neutron logs for more information. [ 743.442461] env[62066]: ERROR nova.compute.manager [instance: 3624ca43-9ee3-418a-9205-debb699295df] [ 743.442461] env[62066]: DEBUG nova.compute.utils [None req-5bf8bcd1-f041-4648-9b3c-4febff568fa9 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] [instance: 3624ca43-9ee3-418a-9205-debb699295df] Binding failed for port bf6b4191-9489-4988-8fd5-320a5a2607a1, please check neutron logs for more information. {{(pid=62066) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 743.444312] env[62066]: ERROR nova.compute.manager [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port a7a012fe-04f6-4d88-b008-27cf1ef05505, please check neutron logs for more information. [ 743.444312] env[62066]: ERROR nova.compute.manager Traceback (most recent call last): [ 743.444312] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 743.444312] env[62066]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 743.444312] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 743.444312] env[62066]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 743.444312] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 743.444312] env[62066]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 743.444312] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 743.444312] env[62066]: ERROR nova.compute.manager self.force_reraise() [ 743.444312] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 743.444312] env[62066]: ERROR nova.compute.manager raise self.value [ 743.444312] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 743.444312] env[62066]: ERROR nova.compute.manager updated_port = self._update_port( [ 743.444312] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 743.444312] env[62066]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 743.444733] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 743.444733] env[62066]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 743.444733] env[62066]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port a7a012fe-04f6-4d88-b008-27cf1ef05505, please check neutron logs for more information. [ 743.444733] env[62066]: ERROR nova.compute.manager [ 743.444733] env[62066]: Traceback (most recent call last): [ 743.444733] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 743.444733] env[62066]: listener.cb(fileno) [ 743.444733] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 743.444733] env[62066]: result = function(*args, **kwargs) [ 743.444733] env[62066]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 743.444733] env[62066]: return func(*args, **kwargs) [ 743.444733] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 743.444733] env[62066]: raise e [ 743.444733] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 743.444733] env[62066]: nwinfo = self.network_api.allocate_for_instance( [ 743.444733] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 743.444733] env[62066]: created_port_ids = self._update_ports_for_instance( [ 743.444733] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 743.444733] env[62066]: with excutils.save_and_reraise_exception(): [ 743.444733] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 743.444733] env[62066]: self.force_reraise() [ 743.444733] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 743.444733] env[62066]: raise self.value [ 743.444733] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 743.444733] env[62066]: updated_port = self._update_port( [ 743.444733] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 743.444733] env[62066]: _ensure_no_port_binding_failure(port) [ 743.444733] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 743.444733] env[62066]: raise exception.PortBindingFailed(port_id=port['id']) [ 743.445494] env[62066]: nova.exception.PortBindingFailed: Binding failed for port a7a012fe-04f6-4d88-b008-27cf1ef05505, please check neutron logs for more information. [ 743.445494] env[62066]: Removing descriptor: 15 [ 743.445494] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7d1d419b-6e61-4f85-9698-fe0c6e4f3988 tempest-AttachInterfacesV270Test-236309384 tempest-AttachInterfacesV270Test-236309384-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.866s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 743.448316] env[62066]: DEBUG nova.compute.manager [None req-5bf8bcd1-f041-4648-9b3c-4febff568fa9 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] [instance: 3624ca43-9ee3-418a-9205-debb699295df] Build of instance 3624ca43-9ee3-418a-9205-debb699295df was re-scheduled: Binding failed for port bf6b4191-9489-4988-8fd5-320a5a2607a1, please check neutron logs for more information. {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 743.448552] env[62066]: DEBUG nova.compute.manager [None req-5bf8bcd1-f041-4648-9b3c-4febff568fa9 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] [instance: 3624ca43-9ee3-418a-9205-debb699295df] Unplugging VIFs for instance {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 743.448923] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5bf8bcd1-f041-4648-9b3c-4febff568fa9 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Acquiring lock "refresh_cache-3624ca43-9ee3-418a-9205-debb699295df" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 743.449125] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5bf8bcd1-f041-4648-9b3c-4febff568fa9 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Acquired lock "refresh_cache-3624ca43-9ee3-418a-9205-debb699295df" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 743.449299] env[62066]: DEBUG nova.network.neutron [None req-5bf8bcd1-f041-4648-9b3c-4febff568fa9 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] [instance: 3624ca43-9ee3-418a-9205-debb699295df] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 743.468425] env[62066]: DEBUG nova.compute.manager [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 743.493828] env[62066]: DEBUG nova.virt.hardware [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-23T13:40:41Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-23T13:40:23Z,direct_url=,disk_format='vmdk',id=50ff584c-3b50-4395-af07-3e66769bc9f7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='eb52f7069a374c61ae946f052007c6d9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-23T13:40:24Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 743.494088] env[62066]: DEBUG nova.virt.hardware [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 743.494248] env[62066]: DEBUG nova.virt.hardware [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 743.494433] env[62066]: DEBUG nova.virt.hardware [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 743.494579] env[62066]: DEBUG nova.virt.hardware [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 743.494725] env[62066]: DEBUG nova.virt.hardware [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 743.494933] env[62066]: DEBUG nova.virt.hardware [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 743.495120] env[62066]: DEBUG nova.virt.hardware [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 743.495292] env[62066]: DEBUG nova.virt.hardware [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 743.495452] env[62066]: DEBUG nova.virt.hardware [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 743.495639] env[62066]: DEBUG nova.virt.hardware [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 743.496740] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c47b72c5-8ad0-4b56-8885-e03237acda63 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.504582] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45fa4508-9b58-42eb-bdde-4450b898c92b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.519312] env[62066]: ERROR nova.compute.manager [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port a7a012fe-04f6-4d88-b008-27cf1ef05505, please check neutron logs for more information. [ 743.519312] env[62066]: ERROR nova.compute.manager [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] Traceback (most recent call last): [ 743.519312] env[62066]: ERROR nova.compute.manager [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 743.519312] env[62066]: ERROR nova.compute.manager [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] yield resources [ 743.519312] env[62066]: ERROR nova.compute.manager [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 743.519312] env[62066]: ERROR nova.compute.manager [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] self.driver.spawn(context, instance, image_meta, [ 743.519312] env[62066]: ERROR nova.compute.manager [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 743.519312] env[62066]: ERROR nova.compute.manager [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 743.519312] env[62066]: ERROR nova.compute.manager [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 743.519312] env[62066]: ERROR nova.compute.manager [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] vm_ref = self.build_virtual_machine(instance, [ 743.519312] env[62066]: ERROR nova.compute.manager [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 743.519675] env[62066]: ERROR nova.compute.manager [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] vif_infos = vmwarevif.get_vif_info(self._session, [ 743.519675] env[62066]: ERROR nova.compute.manager [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 743.519675] env[62066]: ERROR nova.compute.manager [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] for vif in network_info: [ 743.519675] env[62066]: ERROR nova.compute.manager [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 743.519675] env[62066]: ERROR nova.compute.manager [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] return self._sync_wrapper(fn, *args, **kwargs) [ 743.519675] env[62066]: ERROR nova.compute.manager [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 743.519675] env[62066]: ERROR nova.compute.manager [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] self.wait() [ 743.519675] env[62066]: ERROR nova.compute.manager [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 743.519675] env[62066]: ERROR nova.compute.manager [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] self[:] = self._gt.wait() [ 743.519675] env[62066]: ERROR nova.compute.manager [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 743.519675] env[62066]: ERROR nova.compute.manager [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] return self._exit_event.wait() [ 743.519675] env[62066]: ERROR nova.compute.manager [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 743.519675] env[62066]: ERROR nova.compute.manager [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] current.throw(*self._exc) [ 743.520148] env[62066]: ERROR nova.compute.manager [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 743.520148] env[62066]: ERROR nova.compute.manager [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] result = function(*args, **kwargs) [ 743.520148] env[62066]: ERROR nova.compute.manager [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 743.520148] env[62066]: ERROR nova.compute.manager [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] return func(*args, **kwargs) [ 743.520148] env[62066]: ERROR nova.compute.manager [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 743.520148] env[62066]: ERROR nova.compute.manager [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] raise e [ 743.520148] env[62066]: ERROR nova.compute.manager [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 743.520148] env[62066]: ERROR nova.compute.manager [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] nwinfo = self.network_api.allocate_for_instance( [ 743.520148] env[62066]: ERROR nova.compute.manager [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 743.520148] env[62066]: ERROR nova.compute.manager [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] created_port_ids = self._update_ports_for_instance( [ 743.520148] env[62066]: ERROR nova.compute.manager [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 743.520148] env[62066]: ERROR nova.compute.manager [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] with excutils.save_and_reraise_exception(): [ 743.520148] env[62066]: ERROR nova.compute.manager [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 743.520638] env[62066]: ERROR nova.compute.manager [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] self.force_reraise() [ 743.520638] env[62066]: ERROR nova.compute.manager [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 743.520638] env[62066]: ERROR nova.compute.manager [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] raise self.value [ 743.520638] env[62066]: ERROR nova.compute.manager [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 743.520638] env[62066]: ERROR nova.compute.manager [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] updated_port = self._update_port( [ 743.520638] env[62066]: ERROR nova.compute.manager [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 743.520638] env[62066]: ERROR nova.compute.manager [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] _ensure_no_port_binding_failure(port) [ 743.520638] env[62066]: ERROR nova.compute.manager [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 743.520638] env[62066]: ERROR nova.compute.manager [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] raise exception.PortBindingFailed(port_id=port['id']) [ 743.520638] env[62066]: ERROR nova.compute.manager [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] nova.exception.PortBindingFailed: Binding failed for port a7a012fe-04f6-4d88-b008-27cf1ef05505, please check neutron logs for more information. [ 743.520638] env[62066]: ERROR nova.compute.manager [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] [ 743.520638] env[62066]: INFO nova.compute.manager [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] Terminating instance [ 743.522980] env[62066]: DEBUG nova.compute.manager [req-00ce9b58-f854-46b4-8e11-cfc5bde60048 req-67edf850-9238-4bf1-ac2a-b9aa88b1b2ca service nova] [instance: d973b09f-7727-4055-b05e-123079df9ce4] Received event network-vif-deleted-72daca54-f990-4f16-9362-cccaded8dba7 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 743.523180] env[62066]: DEBUG nova.compute.manager [req-00ce9b58-f854-46b4-8e11-cfc5bde60048 req-67edf850-9238-4bf1-ac2a-b9aa88b1b2ca service nova] [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] Received event network-changed-a7a012fe-04f6-4d88-b008-27cf1ef05505 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 743.524440] env[62066]: DEBUG nova.compute.manager [req-00ce9b58-f854-46b4-8e11-cfc5bde60048 req-67edf850-9238-4bf1-ac2a-b9aa88b1b2ca service nova] [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] Refreshing instance network info cache due to event network-changed-a7a012fe-04f6-4d88-b008-27cf1ef05505. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 743.524440] env[62066]: DEBUG oslo_concurrency.lockutils [req-00ce9b58-f854-46b4-8e11-cfc5bde60048 req-67edf850-9238-4bf1-ac2a-b9aa88b1b2ca service nova] Acquiring lock "refresh_cache-71c44a07-fdf0-4847-868a-47f16c033c3d" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 743.524440] env[62066]: DEBUG oslo_concurrency.lockutils [req-00ce9b58-f854-46b4-8e11-cfc5bde60048 req-67edf850-9238-4bf1-ac2a-b9aa88b1b2ca service nova] Acquired lock "refresh_cache-71c44a07-fdf0-4847-868a-47f16c033c3d" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 743.524440] env[62066]: DEBUG nova.network.neutron [req-00ce9b58-f854-46b4-8e11-cfc5bde60048 req-67edf850-9238-4bf1-ac2a-b9aa88b1b2ca service nova] [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] Refreshing network info cache for port a7a012fe-04f6-4d88-b008-27cf1ef05505 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 743.524851] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] Acquiring lock "refresh_cache-71c44a07-fdf0-4847-868a-47f16c033c3d" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 743.791023] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] Releasing lock "refresh_cache-d973b09f-7727-4055-b05e-123079df9ce4" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 743.791327] env[62066]: DEBUG nova.compute.manager [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] [instance: d973b09f-7727-4055-b05e-123079df9ce4] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 743.791531] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] [instance: d973b09f-7727-4055-b05e-123079df9ce4] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 743.791844] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-995c448b-5228-4888-b31f-2aa9ce5a3cfc {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.801650] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f664ae17-57cc-4c31-ac0e-8e7d80b78674 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.822648] env[62066]: WARNING nova.virt.vmwareapi.vmops [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] [instance: d973b09f-7727-4055-b05e-123079df9ce4] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance d973b09f-7727-4055-b05e-123079df9ce4 could not be found. [ 743.822881] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] [instance: d973b09f-7727-4055-b05e-123079df9ce4] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 743.823070] env[62066]: INFO nova.compute.manager [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] [instance: d973b09f-7727-4055-b05e-123079df9ce4] Took 0.03 seconds to destroy the instance on the hypervisor. [ 743.823319] env[62066]: DEBUG oslo.service.loopingcall [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 743.823537] env[62066]: DEBUG nova.compute.manager [-] [instance: d973b09f-7727-4055-b05e-123079df9ce4] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 743.823630] env[62066]: DEBUG nova.network.neutron [-] [instance: d973b09f-7727-4055-b05e-123079df9ce4] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 743.841679] env[62066]: DEBUG nova.network.neutron [-] [instance: d973b09f-7727-4055-b05e-123079df9ce4] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 743.971530] env[62066]: DEBUG nova.network.neutron [None req-5bf8bcd1-f041-4648-9b3c-4febff568fa9 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] [instance: 3624ca43-9ee3-418a-9205-debb699295df] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 744.040908] env[62066]: DEBUG nova.network.neutron [req-00ce9b58-f854-46b4-8e11-cfc5bde60048 req-67edf850-9238-4bf1-ac2a-b9aa88b1b2ca service nova] [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 744.068578] env[62066]: DEBUG nova.network.neutron [None req-5bf8bcd1-f041-4648-9b3c-4febff568fa9 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] [instance: 3624ca43-9ee3-418a-9205-debb699295df] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 744.111333] env[62066]: DEBUG nova.network.neutron [req-00ce9b58-f854-46b4-8e11-cfc5bde60048 req-67edf850-9238-4bf1-ac2a-b9aa88b1b2ca service nova] [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 744.339214] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5424f253-a766-4c3d-8b01-2e7c9e811242 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.343815] env[62066]: DEBUG nova.network.neutron [-] [instance: d973b09f-7727-4055-b05e-123079df9ce4] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 744.347937] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-749a3408-86bf-4acc-ae2b-8fe226f50b5d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.382243] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c307e4d8-005e-45aa-be40-0adcf6cfd3a5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.390071] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05a78d1f-864b-421c-9a0d-83d6e0e379a7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.403975] env[62066]: DEBUG nova.compute.provider_tree [None req-7d1d419b-6e61-4f85-9698-fe0c6e4f3988 tempest-AttachInterfacesV270Test-236309384 tempest-AttachInterfacesV270Test-236309384-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 744.573543] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5bf8bcd1-f041-4648-9b3c-4febff568fa9 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Releasing lock "refresh_cache-3624ca43-9ee3-418a-9205-debb699295df" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 744.573805] env[62066]: DEBUG nova.compute.manager [None req-5bf8bcd1-f041-4648-9b3c-4febff568fa9 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 744.573968] env[62066]: DEBUG nova.compute.manager [None req-5bf8bcd1-f041-4648-9b3c-4febff568fa9 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] [instance: 3624ca43-9ee3-418a-9205-debb699295df] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 744.574156] env[62066]: DEBUG nova.network.neutron [None req-5bf8bcd1-f041-4648-9b3c-4febff568fa9 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] [instance: 3624ca43-9ee3-418a-9205-debb699295df] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 744.589325] env[62066]: DEBUG nova.network.neutron [None req-5bf8bcd1-f041-4648-9b3c-4febff568fa9 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] [instance: 3624ca43-9ee3-418a-9205-debb699295df] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 744.613898] env[62066]: DEBUG oslo_concurrency.lockutils [req-00ce9b58-f854-46b4-8e11-cfc5bde60048 req-67edf850-9238-4bf1-ac2a-b9aa88b1b2ca service nova] Releasing lock "refresh_cache-71c44a07-fdf0-4847-868a-47f16c033c3d" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 744.614169] env[62066]: DEBUG nova.compute.manager [req-00ce9b58-f854-46b4-8e11-cfc5bde60048 req-67edf850-9238-4bf1-ac2a-b9aa88b1b2ca service nova] [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] Received event network-vif-deleted-a7a012fe-04f6-4d88-b008-27cf1ef05505 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 744.614516] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] Acquired lock "refresh_cache-71c44a07-fdf0-4847-868a-47f16c033c3d" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 744.614688] env[62066]: DEBUG nova.network.neutron [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 744.852426] env[62066]: INFO nova.compute.manager [-] [instance: d973b09f-7727-4055-b05e-123079df9ce4] Took 1.03 seconds to deallocate network for instance. [ 744.854565] env[62066]: DEBUG nova.compute.claims [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] [instance: d973b09f-7727-4055-b05e-123079df9ce4] Aborting claim: {{(pid=62066) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 744.854736] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 744.908058] env[62066]: DEBUG nova.scheduler.client.report [None req-7d1d419b-6e61-4f85-9698-fe0c6e4f3988 tempest-AttachInterfacesV270Test-236309384 tempest-AttachInterfacesV270Test-236309384-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 745.091625] env[62066]: DEBUG nova.network.neutron [None req-5bf8bcd1-f041-4648-9b3c-4febff568fa9 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] [instance: 3624ca43-9ee3-418a-9205-debb699295df] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 745.131592] env[62066]: DEBUG nova.network.neutron [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 745.211301] env[62066]: DEBUG nova.network.neutron [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 745.413607] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7d1d419b-6e61-4f85-9698-fe0c6e4f3988 tempest-AttachInterfacesV270Test-236309384 tempest-AttachInterfacesV270Test-236309384-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.969s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 745.414236] env[62066]: ERROR nova.compute.manager [None req-7d1d419b-6e61-4f85-9698-fe0c6e4f3988 tempest-AttachInterfacesV270Test-236309384 tempest-AttachInterfacesV270Test-236309384-project-member] [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port a3e86a4a-9e56-428c-a598-e42c247ab0e2, please check neutron logs for more information. [ 745.414236] env[62066]: ERROR nova.compute.manager [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] Traceback (most recent call last): [ 745.414236] env[62066]: ERROR nova.compute.manager [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 745.414236] env[62066]: ERROR nova.compute.manager [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] self.driver.spawn(context, instance, image_meta, [ 745.414236] env[62066]: ERROR nova.compute.manager [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 745.414236] env[62066]: ERROR nova.compute.manager [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 745.414236] env[62066]: ERROR nova.compute.manager [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 745.414236] env[62066]: ERROR nova.compute.manager [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] vm_ref = self.build_virtual_machine(instance, [ 745.414236] env[62066]: ERROR nova.compute.manager [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 745.414236] env[62066]: ERROR nova.compute.manager [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] vif_infos = vmwarevif.get_vif_info(self._session, [ 745.414236] env[62066]: ERROR nova.compute.manager [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 745.414541] env[62066]: ERROR nova.compute.manager [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] for vif in network_info: [ 745.414541] env[62066]: ERROR nova.compute.manager [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 745.414541] env[62066]: ERROR nova.compute.manager [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] return self._sync_wrapper(fn, *args, **kwargs) [ 745.414541] env[62066]: ERROR nova.compute.manager [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 745.414541] env[62066]: ERROR nova.compute.manager [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] self.wait() [ 745.414541] env[62066]: ERROR nova.compute.manager [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 745.414541] env[62066]: ERROR nova.compute.manager [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] self[:] = self._gt.wait() [ 745.414541] env[62066]: ERROR nova.compute.manager [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 745.414541] env[62066]: ERROR nova.compute.manager [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] return self._exit_event.wait() [ 745.414541] env[62066]: ERROR nova.compute.manager [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 745.414541] env[62066]: ERROR nova.compute.manager [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] current.throw(*self._exc) [ 745.414541] env[62066]: ERROR nova.compute.manager [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 745.414541] env[62066]: ERROR nova.compute.manager [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] result = function(*args, **kwargs) [ 745.414887] env[62066]: ERROR nova.compute.manager [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 745.414887] env[62066]: ERROR nova.compute.manager [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] return func(*args, **kwargs) [ 745.414887] env[62066]: ERROR nova.compute.manager [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 745.414887] env[62066]: ERROR nova.compute.manager [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] raise e [ 745.414887] env[62066]: ERROR nova.compute.manager [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 745.414887] env[62066]: ERROR nova.compute.manager [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] nwinfo = self.network_api.allocate_for_instance( [ 745.414887] env[62066]: ERROR nova.compute.manager [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 745.414887] env[62066]: ERROR nova.compute.manager [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] created_port_ids = self._update_ports_for_instance( [ 745.414887] env[62066]: ERROR nova.compute.manager [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 745.414887] env[62066]: ERROR nova.compute.manager [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] with excutils.save_and_reraise_exception(): [ 745.414887] env[62066]: ERROR nova.compute.manager [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 745.414887] env[62066]: ERROR nova.compute.manager [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] self.force_reraise() [ 745.414887] env[62066]: ERROR nova.compute.manager [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 745.415189] env[62066]: ERROR nova.compute.manager [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] raise self.value [ 745.415189] env[62066]: ERROR nova.compute.manager [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 745.415189] env[62066]: ERROR nova.compute.manager [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] updated_port = self._update_port( [ 745.415189] env[62066]: ERROR nova.compute.manager [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 745.415189] env[62066]: ERROR nova.compute.manager [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] _ensure_no_port_binding_failure(port) [ 745.415189] env[62066]: ERROR nova.compute.manager [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 745.415189] env[62066]: ERROR nova.compute.manager [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] raise exception.PortBindingFailed(port_id=port['id']) [ 745.415189] env[62066]: ERROR nova.compute.manager [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] nova.exception.PortBindingFailed: Binding failed for port a3e86a4a-9e56-428c-a598-e42c247ab0e2, please check neutron logs for more information. [ 745.415189] env[62066]: ERROR nova.compute.manager [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] [ 745.415189] env[62066]: DEBUG nova.compute.utils [None req-7d1d419b-6e61-4f85-9698-fe0c6e4f3988 tempest-AttachInterfacesV270Test-236309384 tempest-AttachInterfacesV270Test-236309384-project-member] [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] Binding failed for port a3e86a4a-9e56-428c-a598-e42c247ab0e2, please check neutron logs for more information. {{(pid=62066) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 745.416121] env[62066]: DEBUG oslo_concurrency.lockutils [None req-fbd85018-c9a9-4c9f-87ff-96cad2b995f9 tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 15.719s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 745.416299] env[62066]: DEBUG nova.objects.instance [None req-fbd85018-c9a9-4c9f-87ff-96cad2b995f9 tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] [instance: 0ff2f9f4-6d85-4465-8e0a-fb39f2a25c9b] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62066) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 745.419016] env[62066]: DEBUG nova.compute.manager [None req-7d1d419b-6e61-4f85-9698-fe0c6e4f3988 tempest-AttachInterfacesV270Test-236309384 tempest-AttachInterfacesV270Test-236309384-project-member] [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] Build of instance c908a3ed-88a6-4e79-8c27-2e16bf2bdca4 was re-scheduled: Binding failed for port a3e86a4a-9e56-428c-a598-e42c247ab0e2, please check neutron logs for more information. {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 745.419432] env[62066]: DEBUG nova.compute.manager [None req-7d1d419b-6e61-4f85-9698-fe0c6e4f3988 tempest-AttachInterfacesV270Test-236309384 tempest-AttachInterfacesV270Test-236309384-project-member] [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] Unplugging VIFs for instance {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 745.419696] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7d1d419b-6e61-4f85-9698-fe0c6e4f3988 tempest-AttachInterfacesV270Test-236309384 tempest-AttachInterfacesV270Test-236309384-project-member] Acquiring lock "refresh_cache-c908a3ed-88a6-4e79-8c27-2e16bf2bdca4" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 745.419822] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7d1d419b-6e61-4f85-9698-fe0c6e4f3988 tempest-AttachInterfacesV270Test-236309384 tempest-AttachInterfacesV270Test-236309384-project-member] Acquired lock "refresh_cache-c908a3ed-88a6-4e79-8c27-2e16bf2bdca4" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 745.419987] env[62066]: DEBUG nova.network.neutron [None req-7d1d419b-6e61-4f85-9698-fe0c6e4f3988 tempest-AttachInterfacesV270Test-236309384 tempest-AttachInterfacesV270Test-236309384-project-member] [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 745.594261] env[62066]: INFO nova.compute.manager [None req-5bf8bcd1-f041-4648-9b3c-4febff568fa9 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] [instance: 3624ca43-9ee3-418a-9205-debb699295df] Took 1.02 seconds to deallocate network for instance. [ 745.714043] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] Releasing lock "refresh_cache-71c44a07-fdf0-4847-868a-47f16c033c3d" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 745.714388] env[62066]: DEBUG nova.compute.manager [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 745.714582] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 745.714872] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e55b03b7-2d7c-4b82-9a85-823ab13312da {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.723719] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83f64ead-ab8e-4dc4-99cc-268fe70ad902 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.744442] env[62066]: WARNING nova.virt.vmwareapi.vmops [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 71c44a07-fdf0-4847-868a-47f16c033c3d could not be found. [ 745.744652] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 745.744829] env[62066]: INFO nova.compute.manager [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] Took 0.03 seconds to destroy the instance on the hypervisor. [ 745.745078] env[62066]: DEBUG oslo.service.loopingcall [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 745.745293] env[62066]: DEBUG nova.compute.manager [-] [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 745.745386] env[62066]: DEBUG nova.network.neutron [-] [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 745.759831] env[62066]: DEBUG nova.network.neutron [-] [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 745.945795] env[62066]: DEBUG nova.network.neutron [None req-7d1d419b-6e61-4f85-9698-fe0c6e4f3988 tempest-AttachInterfacesV270Test-236309384 tempest-AttachInterfacesV270Test-236309384-project-member] [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 746.048223] env[62066]: DEBUG nova.network.neutron [None req-7d1d419b-6e61-4f85-9698-fe0c6e4f3988 tempest-AttachInterfacesV270Test-236309384 tempest-AttachInterfacesV270Test-236309384-project-member] [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 746.262367] env[62066]: DEBUG nova.network.neutron [-] [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 746.427823] env[62066]: DEBUG oslo_concurrency.lockutils [None req-fbd85018-c9a9-4c9f-87ff-96cad2b995f9 tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.012s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 746.428999] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.411s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 746.430459] env[62066]: INFO nova.compute.claims [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 746.550731] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7d1d419b-6e61-4f85-9698-fe0c6e4f3988 tempest-AttachInterfacesV270Test-236309384 tempest-AttachInterfacesV270Test-236309384-project-member] Releasing lock "refresh_cache-c908a3ed-88a6-4e79-8c27-2e16bf2bdca4" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 746.550965] env[62066]: DEBUG nova.compute.manager [None req-7d1d419b-6e61-4f85-9698-fe0c6e4f3988 tempest-AttachInterfacesV270Test-236309384 tempest-AttachInterfacesV270Test-236309384-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 746.551198] env[62066]: DEBUG nova.compute.manager [None req-7d1d419b-6e61-4f85-9698-fe0c6e4f3988 tempest-AttachInterfacesV270Test-236309384 tempest-AttachInterfacesV270Test-236309384-project-member] [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 746.551366] env[62066]: DEBUG nova.network.neutron [None req-7d1d419b-6e61-4f85-9698-fe0c6e4f3988 tempest-AttachInterfacesV270Test-236309384 tempest-AttachInterfacesV270Test-236309384-project-member] [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 746.565997] env[62066]: DEBUG nova.network.neutron [None req-7d1d419b-6e61-4f85-9698-fe0c6e4f3988 tempest-AttachInterfacesV270Test-236309384 tempest-AttachInterfacesV270Test-236309384-project-member] [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 746.623262] env[62066]: INFO nova.scheduler.client.report [None req-5bf8bcd1-f041-4648-9b3c-4febff568fa9 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Deleted allocations for instance 3624ca43-9ee3-418a-9205-debb699295df [ 746.767578] env[62066]: INFO nova.compute.manager [-] [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] Took 1.02 seconds to deallocate network for instance. [ 746.770086] env[62066]: DEBUG nova.compute.claims [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] Aborting claim: {{(pid=62066) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 746.770268] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 747.068461] env[62066]: DEBUG nova.network.neutron [None req-7d1d419b-6e61-4f85-9698-fe0c6e4f3988 tempest-AttachInterfacesV270Test-236309384 tempest-AttachInterfacesV270Test-236309384-project-member] [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 747.131599] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5bf8bcd1-f041-4648-9b3c-4febff568fa9 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Lock "3624ca43-9ee3-418a-9205-debb699295df" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 164.743s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 747.571590] env[62066]: INFO nova.compute.manager [None req-7d1d419b-6e61-4f85-9698-fe0c6e4f3988 tempest-AttachInterfacesV270Test-236309384 tempest-AttachInterfacesV270Test-236309384-project-member] [instance: c908a3ed-88a6-4e79-8c27-2e16bf2bdca4] Took 1.02 seconds to deallocate network for instance. [ 747.635947] env[62066]: DEBUG nova.compute.manager [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 98fb270c-f3f8-4375-8b5c-c2279305d476] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 747.756974] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f96d0bb6-d55a-40c1-a89b-cc600df2e30d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.765679] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ac64ba7-cf68-4c16-aa7f-3cd0ccbf5347 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.802240] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b132b599-e91f-4aaf-842e-1dbad3d1517c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.809905] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e6a5a03-29da-4a48-b1f3-f82bc2cd6d71 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.824497] env[62066]: DEBUG nova.compute.provider_tree [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 748.159873] env[62066]: DEBUG oslo_concurrency.lockutils [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 748.326868] env[62066]: DEBUG nova.scheduler.client.report [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 748.612111] env[62066]: INFO nova.scheduler.client.report [None req-7d1d419b-6e61-4f85-9698-fe0c6e4f3988 tempest-AttachInterfacesV270Test-236309384 tempest-AttachInterfacesV270Test-236309384-project-member] Deleted allocations for instance c908a3ed-88a6-4e79-8c27-2e16bf2bdca4 [ 748.833102] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.404s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 748.835631] env[62066]: DEBUG nova.compute.manager [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 748.836558] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 18.611s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 749.117809] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7d1d419b-6e61-4f85-9698-fe0c6e4f3988 tempest-AttachInterfacesV270Test-236309384 tempest-AttachInterfacesV270Test-236309384-project-member] Lock "c908a3ed-88a6-4e79-8c27-2e16bf2bdca4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 147.934s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 749.341178] env[62066]: DEBUG nova.compute.utils [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 749.346385] env[62066]: DEBUG nova.compute.manager [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 749.346385] env[62066]: DEBUG nova.network.neutron [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 749.395781] env[62066]: DEBUG nova.policy [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5ae0f813ad034e8988e31d93da7894cc', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0545f648e24c455ab6fd7ae1ff1a0214', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 749.620724] env[62066]: DEBUG nova.compute.manager [None req-064a3fac-0a45-4f27-8eed-cb07d9679634 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] [instance: 1a4b0637-1a56-41ef-b89b-6b56d24ed206] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 749.719708] env[62066]: DEBUG nova.network.neutron [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] Successfully created port: 85acf49f-2497-490a-8686-8131ebcad644 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 749.762198] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d2edd17d-ac3d-4b81-90f1-278c62ed416c tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Acquiring lock "de2e2be7-efdb-45a8-842a-640ab9deb1d9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 749.762438] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d2edd17d-ac3d-4b81-90f1-278c62ed416c tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Lock "de2e2be7-efdb-45a8-842a-640ab9deb1d9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 749.849888] env[62066]: DEBUG nova.compute.manager [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 749.963706] env[62066]: WARNING nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Instance 0ff2f9f4-6d85-4465-8e0a-fb39f2a25c9b is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 749.963706] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Instance 0ed07533-9e79-416d-a495-a8e20b97e108 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 749.963706] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Instance 2b81d060-ecb7-4ae5-b51d-33197937a76f actively managed on this compute host and has allocations in placement: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 749.963706] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Instance 04d25af4-0e73-4650-9c3d-85817754bac9 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 749.963854] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Instance d973b09f-7727-4055-b05e-123079df9ce4 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 749.963854] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Instance 71c44a07-fdf0-4847-868a-47f16c033c3d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 749.963854] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Instance f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 750.147893] env[62066]: DEBUG oslo_concurrency.lockutils [None req-064a3fac-0a45-4f27-8eed-cb07d9679634 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 750.465851] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Instance 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 750.694073] env[62066]: DEBUG nova.compute.manager [req-84767154-9e06-4bb2-9e27-f9a2b0eb54f7 req-c6fbb230-2bce-478d-8cca-d26c32c30c3f service nova] [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] Received event network-changed-85acf49f-2497-490a-8686-8131ebcad644 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 750.694073] env[62066]: DEBUG nova.compute.manager [req-84767154-9e06-4bb2-9e27-f9a2b0eb54f7 req-c6fbb230-2bce-478d-8cca-d26c32c30c3f service nova] [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] Refreshing instance network info cache due to event network-changed-85acf49f-2497-490a-8686-8131ebcad644. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 750.694073] env[62066]: DEBUG oslo_concurrency.lockutils [req-84767154-9e06-4bb2-9e27-f9a2b0eb54f7 req-c6fbb230-2bce-478d-8cca-d26c32c30c3f service nova] Acquiring lock "refresh_cache-f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 750.694073] env[62066]: DEBUG oslo_concurrency.lockutils [req-84767154-9e06-4bb2-9e27-f9a2b0eb54f7 req-c6fbb230-2bce-478d-8cca-d26c32c30c3f service nova] Acquired lock "refresh_cache-f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 750.694073] env[62066]: DEBUG nova.network.neutron [req-84767154-9e06-4bb2-9e27-f9a2b0eb54f7 req-c6fbb230-2bce-478d-8cca-d26c32c30c3f service nova] [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] Refreshing network info cache for port 85acf49f-2497-490a-8686-8131ebcad644 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 750.863966] env[62066]: DEBUG nova.compute.manager [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 750.904405] env[62066]: DEBUG nova.virt.hardware [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-23T13:40:41Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-23T13:40:23Z,direct_url=,disk_format='vmdk',id=50ff584c-3b50-4395-af07-3e66769bc9f7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='eb52f7069a374c61ae946f052007c6d9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-23T13:40:24Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 750.904649] env[62066]: DEBUG nova.virt.hardware [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 750.904803] env[62066]: DEBUG nova.virt.hardware [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 750.904985] env[62066]: DEBUG nova.virt.hardware [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 750.905143] env[62066]: DEBUG nova.virt.hardware [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 750.905291] env[62066]: DEBUG nova.virt.hardware [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 750.905496] env[62066]: DEBUG nova.virt.hardware [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 750.905652] env[62066]: DEBUG nova.virt.hardware [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 750.905816] env[62066]: DEBUG nova.virt.hardware [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 750.905976] env[62066]: DEBUG nova.virt.hardware [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 750.909962] env[62066]: DEBUG nova.virt.hardware [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 750.911528] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe0e973f-9cc9-4b2e-9fc0-861be753968e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.918241] env[62066]: ERROR nova.compute.manager [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 85acf49f-2497-490a-8686-8131ebcad644, please check neutron logs for more information. [ 750.918241] env[62066]: ERROR nova.compute.manager Traceback (most recent call last): [ 750.918241] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 750.918241] env[62066]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 750.918241] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 750.918241] env[62066]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 750.918241] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 750.918241] env[62066]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 750.918241] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 750.918241] env[62066]: ERROR nova.compute.manager self.force_reraise() [ 750.918241] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 750.918241] env[62066]: ERROR nova.compute.manager raise self.value [ 750.918241] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 750.918241] env[62066]: ERROR nova.compute.manager updated_port = self._update_port( [ 750.918241] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 750.918241] env[62066]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 750.918658] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 750.918658] env[62066]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 750.918658] env[62066]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 85acf49f-2497-490a-8686-8131ebcad644, please check neutron logs for more information. [ 750.918658] env[62066]: ERROR nova.compute.manager [ 750.918658] env[62066]: Traceback (most recent call last): [ 750.918658] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 750.918658] env[62066]: listener.cb(fileno) [ 750.918658] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 750.918658] env[62066]: result = function(*args, **kwargs) [ 750.918658] env[62066]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 750.918658] env[62066]: return func(*args, **kwargs) [ 750.918658] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 750.918658] env[62066]: raise e [ 750.918658] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 750.918658] env[62066]: nwinfo = self.network_api.allocate_for_instance( [ 750.918658] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 750.918658] env[62066]: created_port_ids = self._update_ports_for_instance( [ 750.918658] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 750.918658] env[62066]: with excutils.save_and_reraise_exception(): [ 750.918658] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 750.918658] env[62066]: self.force_reraise() [ 750.918658] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 750.918658] env[62066]: raise self.value [ 750.918658] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 750.918658] env[62066]: updated_port = self._update_port( [ 750.918658] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 750.918658] env[62066]: _ensure_no_port_binding_failure(port) [ 750.918658] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 750.918658] env[62066]: raise exception.PortBindingFailed(port_id=port['id']) [ 750.920658] env[62066]: nova.exception.PortBindingFailed: Binding failed for port 85acf49f-2497-490a-8686-8131ebcad644, please check neutron logs for more information. [ 750.920658] env[62066]: Removing descriptor: 20 [ 750.922548] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d060ec5a-c3ee-4744-af3f-8694419d4077 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.938373] env[62066]: ERROR nova.compute.manager [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 85acf49f-2497-490a-8686-8131ebcad644, please check neutron logs for more information. [ 750.938373] env[62066]: ERROR nova.compute.manager [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] Traceback (most recent call last): [ 750.938373] env[62066]: ERROR nova.compute.manager [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 750.938373] env[62066]: ERROR nova.compute.manager [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] yield resources [ 750.938373] env[62066]: ERROR nova.compute.manager [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 750.938373] env[62066]: ERROR nova.compute.manager [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] self.driver.spawn(context, instance, image_meta, [ 750.938373] env[62066]: ERROR nova.compute.manager [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 750.938373] env[62066]: ERROR nova.compute.manager [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] self._vmops.spawn(context, instance, image_meta, injected_files, [ 750.938373] env[62066]: ERROR nova.compute.manager [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 750.938373] env[62066]: ERROR nova.compute.manager [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] vm_ref = self.build_virtual_machine(instance, [ 750.938373] env[62066]: ERROR nova.compute.manager [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 750.938713] env[62066]: ERROR nova.compute.manager [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] vif_infos = vmwarevif.get_vif_info(self._session, [ 750.938713] env[62066]: ERROR nova.compute.manager [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 750.938713] env[62066]: ERROR nova.compute.manager [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] for vif in network_info: [ 750.938713] env[62066]: ERROR nova.compute.manager [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 750.938713] env[62066]: ERROR nova.compute.manager [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] return self._sync_wrapper(fn, *args, **kwargs) [ 750.938713] env[62066]: ERROR nova.compute.manager [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 750.938713] env[62066]: ERROR nova.compute.manager [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] self.wait() [ 750.938713] env[62066]: ERROR nova.compute.manager [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 750.938713] env[62066]: ERROR nova.compute.manager [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] self[:] = self._gt.wait() [ 750.938713] env[62066]: ERROR nova.compute.manager [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 750.938713] env[62066]: ERROR nova.compute.manager [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] return self._exit_event.wait() [ 750.938713] env[62066]: ERROR nova.compute.manager [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 750.938713] env[62066]: ERROR nova.compute.manager [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] current.throw(*self._exc) [ 750.939018] env[62066]: ERROR nova.compute.manager [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 750.939018] env[62066]: ERROR nova.compute.manager [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] result = function(*args, **kwargs) [ 750.939018] env[62066]: ERROR nova.compute.manager [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 750.939018] env[62066]: ERROR nova.compute.manager [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] return func(*args, **kwargs) [ 750.939018] env[62066]: ERROR nova.compute.manager [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 750.939018] env[62066]: ERROR nova.compute.manager [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] raise e [ 750.939018] env[62066]: ERROR nova.compute.manager [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 750.939018] env[62066]: ERROR nova.compute.manager [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] nwinfo = self.network_api.allocate_for_instance( [ 750.939018] env[62066]: ERROR nova.compute.manager [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 750.939018] env[62066]: ERROR nova.compute.manager [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] created_port_ids = self._update_ports_for_instance( [ 750.939018] env[62066]: ERROR nova.compute.manager [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 750.939018] env[62066]: ERROR nova.compute.manager [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] with excutils.save_and_reraise_exception(): [ 750.939018] env[62066]: ERROR nova.compute.manager [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 750.939303] env[62066]: ERROR nova.compute.manager [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] self.force_reraise() [ 750.939303] env[62066]: ERROR nova.compute.manager [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 750.939303] env[62066]: ERROR nova.compute.manager [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] raise self.value [ 750.939303] env[62066]: ERROR nova.compute.manager [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 750.939303] env[62066]: ERROR nova.compute.manager [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] updated_port = self._update_port( [ 750.939303] env[62066]: ERROR nova.compute.manager [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 750.939303] env[62066]: ERROR nova.compute.manager [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] _ensure_no_port_binding_failure(port) [ 750.939303] env[62066]: ERROR nova.compute.manager [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 750.939303] env[62066]: ERROR nova.compute.manager [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] raise exception.PortBindingFailed(port_id=port['id']) [ 750.939303] env[62066]: ERROR nova.compute.manager [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] nova.exception.PortBindingFailed: Binding failed for port 85acf49f-2497-490a-8686-8131ebcad644, please check neutron logs for more information. [ 750.939303] env[62066]: ERROR nova.compute.manager [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] [ 750.939303] env[62066]: INFO nova.compute.manager [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] Terminating instance [ 750.940943] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] Acquiring lock "refresh_cache-f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 750.969019] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Instance 0ebba3b2-f82d-4c1d-b01a-f75c3559c117 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 751.211617] env[62066]: DEBUG nova.network.neutron [req-84767154-9e06-4bb2-9e27-f9a2b0eb54f7 req-c6fbb230-2bce-478d-8cca-d26c32c30c3f service nova] [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 751.296900] env[62066]: DEBUG nova.network.neutron [req-84767154-9e06-4bb2-9e27-f9a2b0eb54f7 req-c6fbb230-2bce-478d-8cca-d26c32c30c3f service nova] [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 751.479037] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Instance 98fb270c-f3f8-4375-8b5c-c2279305d476 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 751.800093] env[62066]: DEBUG oslo_concurrency.lockutils [req-84767154-9e06-4bb2-9e27-f9a2b0eb54f7 req-c6fbb230-2bce-478d-8cca-d26c32c30c3f service nova] Releasing lock "refresh_cache-f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 751.800560] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] Acquired lock "refresh_cache-f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 751.800692] env[62066]: DEBUG nova.network.neutron [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 751.981848] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Instance 1a4b0637-1a56-41ef-b89b-6b56d24ed206 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 752.317380] env[62066]: DEBUG nova.network.neutron [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 752.413409] env[62066]: DEBUG nova.network.neutron [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 752.484618] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Instance 1e1a2ede-4ad8-4600-851e-6f2046b3f919 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 752.717380] env[62066]: DEBUG nova.compute.manager [req-ad0dc634-cb48-4409-a849-d3b05dcd6bc8 req-348f4322-9da1-4b51-aad5-ee3074e71c6d service nova] [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] Received event network-vif-deleted-85acf49f-2497-490a-8686-8131ebcad644 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 752.916928] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] Releasing lock "refresh_cache-f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 752.916928] env[62066]: DEBUG nova.compute.manager [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 752.917146] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 752.917285] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-63da21f7-af70-457e-bac0-9e395786112f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.926666] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a93dda41-85c2-47fb-8b48-7b6eaf121677 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.947175] env[62066]: WARNING nova.virt.vmwareapi.vmops [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5 could not be found. [ 752.947383] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 752.947562] env[62066]: INFO nova.compute.manager [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] Took 0.03 seconds to destroy the instance on the hypervisor. [ 752.947790] env[62066]: DEBUG oslo.service.loopingcall [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 752.948044] env[62066]: DEBUG nova.compute.manager [-] [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 752.948145] env[62066]: DEBUG nova.network.neutron [-] [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 752.962109] env[62066]: DEBUG nova.network.neutron [-] [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 752.987983] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Instance daffaf51-4c45-44aa-8fc2-4db066a09971 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 753.464195] env[62066]: DEBUG nova.network.neutron [-] [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 753.491045] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Instance 02fb3e92-5dd6-4b1e-a6e4-d60d3fc85b07 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 753.967188] env[62066]: INFO nova.compute.manager [-] [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] Took 1.02 seconds to deallocate network for instance. [ 753.969581] env[62066]: DEBUG nova.compute.claims [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] Aborting claim: {{(pid=62066) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 753.969774] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 753.993642] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Instance 459f432b-e817-405f-9c3e-dfa5d957788c has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 754.496709] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Instance 31d47299-83a3-4f27-aeb7-95c4cd36c5ac has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 754.999485] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Instance 49f46244-34fa-48a1-95a2-8e95850f345d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 755.503027] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Instance 738ebb2f-7cc6-4d0f-871c-1428c9c21384 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 756.004959] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Instance 5f645eaa-6edc-4362-82ad-38c4c57b2be2 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 756.508227] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Instance 050040cb-6ee3-4ad2-960a-fcebb53ac394 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 757.011063] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Instance 49bc1350-0095-406b-bc68-005eb6b681a8 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 757.514175] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Instance 8ff6fbb9-c90f-498d-9a85-d220a8c2f794 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 758.016912] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Instance 2709d0ad-5dd8-4e3c-b1e6-3d1e2e242380 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 758.523930] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Instance 88daaaa6-f385-4161-bc74-9fffabb1145d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 759.028333] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Instance de2e2be7-efdb-45a8-842a-640ab9deb1d9 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 759.028630] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Total usable vcpus: 48, total allocated vcpus: 6 {{(pid=62066) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 759.028845] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1664MB phys_disk=200GB used_disk=5GB total_vcpus=48 used_vcpus=6 pci_stats=[] {{(pid=62066) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 759.241135] env[62066]: DEBUG nova.scheduler.client.report [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Refreshing inventories for resource provider 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 759.255675] env[62066]: DEBUG nova.scheduler.client.report [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Updating ProviderTree inventory for provider 8e4f7194-1498-4f08-8723-ab7260524bcb from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 759.255852] env[62066]: DEBUG nova.compute.provider_tree [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Updating inventory in ProviderTree for provider 8e4f7194-1498-4f08-8723-ab7260524bcb with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 759.266220] env[62066]: DEBUG nova.scheduler.client.report [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Refreshing aggregate associations for resource provider 8e4f7194-1498-4f08-8723-ab7260524bcb, aggregates: None {{(pid=62066) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 759.284298] env[62066]: DEBUG nova.scheduler.client.report [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Refreshing trait associations for resource provider 8e4f7194-1498-4f08-8723-ab7260524bcb, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,HW_ARCH_X86_64,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=62066) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 759.572791] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-748aa141-835f-4baa-b1ce-d18fd96b5329 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.580235] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5606c1f5-c0ab-4e6a-833b-47e93dfdf946 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.608970] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a672ef76-41c1-4801-9e27-6c1e3e5bf61b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.616119] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fbe8c1d-3193-42bc-b108-2476f4c3a313 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.630080] env[62066]: DEBUG nova.compute.provider_tree [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 760.133638] env[62066]: DEBUG nova.scheduler.client.report [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 760.640932] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62066) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 760.641221] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 11.805s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 760.641500] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a4d93091-69b6-4b35-887c-e7b927c2300a tempest-ServerMetadataTestJSON-1590103242 tempest-ServerMetadataTestJSON-1590103242-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 27.559s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 760.644288] env[62066]: DEBUG oslo_service.periodic_task [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 760.645496] env[62066]: DEBUG nova.compute.manager [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Cleaning up deleted instances {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11228}} [ 761.151603] env[62066]: DEBUG nova.compute.manager [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] There are 3 instances to clean {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11237}} [ 761.151917] env[62066]: DEBUG nova.compute.manager [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] [instance: 5d54c3fd-b239-4965-b7a3-5909e8de8bc0] Instance has had 0 of 5 cleanup attempts {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 761.442598] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-644a2ecc-6376-44c6-a351-b02e64f0296b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.451107] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d05999f3-0d7e-40e9-9742-f7cd8c12914b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.481265] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c87b876-d89a-4d32-999d-144c491a12bd {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.488526] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-945e4e49-3a06-40c6-9255-522a49497200 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.501489] env[62066]: DEBUG nova.compute.provider_tree [None req-a4d93091-69b6-4b35-887c-e7b927c2300a tempest-ServerMetadataTestJSON-1590103242 tempest-ServerMetadataTestJSON-1590103242-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 761.659021] env[62066]: DEBUG nova.compute.manager [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] [instance: 56eee633-2c01-4df2-8d2e-58ad712942f6] Instance has had 0 of 5 cleanup attempts {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 762.004785] env[62066]: DEBUG nova.scheduler.client.report [None req-a4d93091-69b6-4b35-887c-e7b927c2300a tempest-ServerMetadataTestJSON-1590103242 tempest-ServerMetadataTestJSON-1590103242-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 762.162254] env[62066]: DEBUG nova.compute.manager [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] [instance: 4b34ff57-e137-4abc-8aed-76dd2ec8d313] Instance has had 0 of 5 cleanup attempts {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 762.517936] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a4d93091-69b6-4b35-887c-e7b927c2300a tempest-ServerMetadataTestJSON-1590103242 tempest-ServerMetadataTestJSON-1590103242-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.876s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 762.518291] env[62066]: ERROR nova.compute.manager [None req-a4d93091-69b6-4b35-887c-e7b927c2300a tempest-ServerMetadataTestJSON-1590103242 tempest-ServerMetadataTestJSON-1590103242-project-member] [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 91b620d9-65dc-4e47-87b9-3747a6897b6f, please check neutron logs for more information. [ 762.518291] env[62066]: ERROR nova.compute.manager [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] Traceback (most recent call last): [ 762.518291] env[62066]: ERROR nova.compute.manager [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 762.518291] env[62066]: ERROR nova.compute.manager [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] self.driver.spawn(context, instance, image_meta, [ 762.518291] env[62066]: ERROR nova.compute.manager [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 762.518291] env[62066]: ERROR nova.compute.manager [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] self._vmops.spawn(context, instance, image_meta, injected_files, [ 762.518291] env[62066]: ERROR nova.compute.manager [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 762.518291] env[62066]: ERROR nova.compute.manager [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] vm_ref = self.build_virtual_machine(instance, [ 762.518291] env[62066]: ERROR nova.compute.manager [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 762.518291] env[62066]: ERROR nova.compute.manager [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] vif_infos = vmwarevif.get_vif_info(self._session, [ 762.518291] env[62066]: ERROR nova.compute.manager [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 762.518632] env[62066]: ERROR nova.compute.manager [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] for vif in network_info: [ 762.518632] env[62066]: ERROR nova.compute.manager [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 762.518632] env[62066]: ERROR nova.compute.manager [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] return self._sync_wrapper(fn, *args, **kwargs) [ 762.518632] env[62066]: ERROR nova.compute.manager [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 762.518632] env[62066]: ERROR nova.compute.manager [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] self.wait() [ 762.518632] env[62066]: ERROR nova.compute.manager [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 762.518632] env[62066]: ERROR nova.compute.manager [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] self[:] = self._gt.wait() [ 762.518632] env[62066]: ERROR nova.compute.manager [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 762.518632] env[62066]: ERROR nova.compute.manager [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] return self._exit_event.wait() [ 762.518632] env[62066]: ERROR nova.compute.manager [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 762.518632] env[62066]: ERROR nova.compute.manager [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] current.throw(*self._exc) [ 762.518632] env[62066]: ERROR nova.compute.manager [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 762.518632] env[62066]: ERROR nova.compute.manager [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] result = function(*args, **kwargs) [ 762.519153] env[62066]: ERROR nova.compute.manager [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 762.519153] env[62066]: ERROR nova.compute.manager [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] return func(*args, **kwargs) [ 762.519153] env[62066]: ERROR nova.compute.manager [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 762.519153] env[62066]: ERROR nova.compute.manager [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] raise e [ 762.519153] env[62066]: ERROR nova.compute.manager [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 762.519153] env[62066]: ERROR nova.compute.manager [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] nwinfo = self.network_api.allocate_for_instance( [ 762.519153] env[62066]: ERROR nova.compute.manager [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 762.519153] env[62066]: ERROR nova.compute.manager [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] created_port_ids = self._update_ports_for_instance( [ 762.519153] env[62066]: ERROR nova.compute.manager [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 762.519153] env[62066]: ERROR nova.compute.manager [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] with excutils.save_and_reraise_exception(): [ 762.519153] env[62066]: ERROR nova.compute.manager [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 762.519153] env[62066]: ERROR nova.compute.manager [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] self.force_reraise() [ 762.519153] env[62066]: ERROR nova.compute.manager [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 762.519510] env[62066]: ERROR nova.compute.manager [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] raise self.value [ 762.519510] env[62066]: ERROR nova.compute.manager [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 762.519510] env[62066]: ERROR nova.compute.manager [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] updated_port = self._update_port( [ 762.519510] env[62066]: ERROR nova.compute.manager [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 762.519510] env[62066]: ERROR nova.compute.manager [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] _ensure_no_port_binding_failure(port) [ 762.519510] env[62066]: ERROR nova.compute.manager [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 762.519510] env[62066]: ERROR nova.compute.manager [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] raise exception.PortBindingFailed(port_id=port['id']) [ 762.519510] env[62066]: ERROR nova.compute.manager [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] nova.exception.PortBindingFailed: Binding failed for port 91b620d9-65dc-4e47-87b9-3747a6897b6f, please check neutron logs for more information. [ 762.519510] env[62066]: ERROR nova.compute.manager [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] [ 762.519510] env[62066]: DEBUG nova.compute.utils [None req-a4d93091-69b6-4b35-887c-e7b927c2300a tempest-ServerMetadataTestJSON-1590103242 tempest-ServerMetadataTestJSON-1590103242-project-member] [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] Binding failed for port 91b620d9-65dc-4e47-87b9-3747a6897b6f, please check neutron logs for more information. {{(pid=62066) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 762.520544] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a2d866c9-e011-47d1-a21e-f9569fa299fe tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 28.381s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 762.520750] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a2d866c9-e011-47d1-a21e-f9569fa299fe tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 762.522691] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7fc7916e-6e82-4e11-a2d0-03a259fc6439 tempest-ServerAddressesNegativeTestJSON-1804754632 tempest-ServerAddressesNegativeTestJSON-1804754632-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.644s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 762.524152] env[62066]: INFO nova.compute.claims [None req-7fc7916e-6e82-4e11-a2d0-03a259fc6439 tempest-ServerAddressesNegativeTestJSON-1804754632 tempest-ServerAddressesNegativeTestJSON-1804754632-project-member] [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 762.527032] env[62066]: DEBUG nova.compute.manager [None req-a4d93091-69b6-4b35-887c-e7b927c2300a tempest-ServerMetadataTestJSON-1590103242 tempest-ServerMetadataTestJSON-1590103242-project-member] [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] Build of instance 0ed07533-9e79-416d-a495-a8e20b97e108 was re-scheduled: Binding failed for port 91b620d9-65dc-4e47-87b9-3747a6897b6f, please check neutron logs for more information. {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 762.527402] env[62066]: DEBUG nova.compute.manager [None req-a4d93091-69b6-4b35-887c-e7b927c2300a tempest-ServerMetadataTestJSON-1590103242 tempest-ServerMetadataTestJSON-1590103242-project-member] [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] Unplugging VIFs for instance {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 762.527596] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a4d93091-69b6-4b35-887c-e7b927c2300a tempest-ServerMetadataTestJSON-1590103242 tempest-ServerMetadataTestJSON-1590103242-project-member] Acquiring lock "refresh_cache-0ed07533-9e79-416d-a495-a8e20b97e108" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 762.527745] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a4d93091-69b6-4b35-887c-e7b927c2300a tempest-ServerMetadataTestJSON-1590103242 tempest-ServerMetadataTestJSON-1590103242-project-member] Acquired lock "refresh_cache-0ed07533-9e79-416d-a495-a8e20b97e108" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 762.527902] env[62066]: DEBUG nova.network.neutron [None req-a4d93091-69b6-4b35-887c-e7b927c2300a tempest-ServerMetadataTestJSON-1590103242 tempest-ServerMetadataTestJSON-1590103242-project-member] [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 762.556136] env[62066]: INFO nova.scheduler.client.report [None req-a2d866c9-e011-47d1-a21e-f9569fa299fe tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Deleted allocations for instance 0ff2f9f4-6d85-4465-8e0a-fb39f2a25c9b [ 762.666822] env[62066]: DEBUG oslo_service.periodic_task [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 762.666822] env[62066]: DEBUG nova.compute.manager [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Cleaning up deleted instances with incomplete migration {{(pid=62066) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11266}} [ 763.049157] env[62066]: DEBUG nova.network.neutron [None req-a4d93091-69b6-4b35-887c-e7b927c2300a tempest-ServerMetadataTestJSON-1590103242 tempest-ServerMetadataTestJSON-1590103242-project-member] [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 763.064070] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a2d866c9-e011-47d1-a21e-f9569fa299fe tempest-ServerShowV254Test-48129616 tempest-ServerShowV254Test-48129616-project-member] Lock "0ff2f9f4-6d85-4465-8e0a-fb39f2a25c9b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 32.661s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 763.150260] env[62066]: DEBUG nova.network.neutron [None req-a4d93091-69b6-4b35-887c-e7b927c2300a tempest-ServerMetadataTestJSON-1590103242 tempest-ServerMetadataTestJSON-1590103242-project-member] [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 763.168607] env[62066]: DEBUG oslo_service.periodic_task [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 763.653158] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a4d93091-69b6-4b35-887c-e7b927c2300a tempest-ServerMetadataTestJSON-1590103242 tempest-ServerMetadataTestJSON-1590103242-project-member] Releasing lock "refresh_cache-0ed07533-9e79-416d-a495-a8e20b97e108" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 763.653361] env[62066]: DEBUG nova.compute.manager [None req-a4d93091-69b6-4b35-887c-e7b927c2300a tempest-ServerMetadataTestJSON-1590103242 tempest-ServerMetadataTestJSON-1590103242-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 763.653954] env[62066]: DEBUG nova.compute.manager [None req-a4d93091-69b6-4b35-887c-e7b927c2300a tempest-ServerMetadataTestJSON-1590103242 tempest-ServerMetadataTestJSON-1590103242-project-member] [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 763.653954] env[62066]: DEBUG nova.network.neutron [None req-a4d93091-69b6-4b35-887c-e7b927c2300a tempest-ServerMetadataTestJSON-1590103242 tempest-ServerMetadataTestJSON-1590103242-project-member] [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 763.667369] env[62066]: DEBUG nova.network.neutron [None req-a4d93091-69b6-4b35-887c-e7b927c2300a tempest-ServerMetadataTestJSON-1590103242 tempest-ServerMetadataTestJSON-1590103242-project-member] [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 763.817162] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1c8d65e-7ec5-46ca-82ee-8e9181556b90 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.824748] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acf5ed57-7fce-40d0-ad7d-54271533c971 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.856294] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d7e3bcf-117a-438e-822f-ec6e68435b39 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.864074] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-102a94ca-6196-4d36-ab3a-93ee4ba6c4a3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.877360] env[62066]: DEBUG nova.compute.provider_tree [None req-7fc7916e-6e82-4e11-a2d0-03a259fc6439 tempest-ServerAddressesNegativeTestJSON-1804754632 tempest-ServerAddressesNegativeTestJSON-1804754632-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 764.171862] env[62066]: DEBUG nova.network.neutron [None req-a4d93091-69b6-4b35-887c-e7b927c2300a tempest-ServerMetadataTestJSON-1590103242 tempest-ServerMetadataTestJSON-1590103242-project-member] [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 764.380600] env[62066]: DEBUG nova.scheduler.client.report [None req-7fc7916e-6e82-4e11-a2d0-03a259fc6439 tempest-ServerAddressesNegativeTestJSON-1804754632 tempest-ServerAddressesNegativeTestJSON-1804754632-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 764.674793] env[62066]: INFO nova.compute.manager [None req-a4d93091-69b6-4b35-887c-e7b927c2300a tempest-ServerMetadataTestJSON-1590103242 tempest-ServerMetadataTestJSON-1590103242-project-member] [instance: 0ed07533-9e79-416d-a495-a8e20b97e108] Took 1.02 seconds to deallocate network for instance. [ 764.885512] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7fc7916e-6e82-4e11-a2d0-03a259fc6439 tempest-ServerAddressesNegativeTestJSON-1804754632 tempest-ServerAddressesNegativeTestJSON-1804754632-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.363s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 764.886050] env[62066]: DEBUG nova.compute.manager [None req-7fc7916e-6e82-4e11-a2d0-03a259fc6439 tempest-ServerAddressesNegativeTestJSON-1804754632 tempest-ServerAddressesNegativeTestJSON-1804754632-project-member] [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 764.888539] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e9793203-98b3-4ca0-8a49-79767b1f52f7 tempest-ServersTestBootFromVolume-1535558002 tempest-ServersTestBootFromVolume-1535558002-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 27.092s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 765.393276] env[62066]: DEBUG nova.compute.utils [None req-7fc7916e-6e82-4e11-a2d0-03a259fc6439 tempest-ServerAddressesNegativeTestJSON-1804754632 tempest-ServerAddressesNegativeTestJSON-1804754632-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 765.399290] env[62066]: DEBUG nova.compute.manager [None req-7fc7916e-6e82-4e11-a2d0-03a259fc6439 tempest-ServerAddressesNegativeTestJSON-1804754632 tempest-ServerAddressesNegativeTestJSON-1804754632-project-member] [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 765.399290] env[62066]: DEBUG nova.network.neutron [None req-7fc7916e-6e82-4e11-a2d0-03a259fc6439 tempest-ServerAddressesNegativeTestJSON-1804754632 tempest-ServerAddressesNegativeTestJSON-1804754632-project-member] [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 765.481281] env[62066]: DEBUG nova.policy [None req-7fc7916e-6e82-4e11-a2d0-03a259fc6439 tempest-ServerAddressesNegativeTestJSON-1804754632 tempest-ServerAddressesNegativeTestJSON-1804754632-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '93e21df9e3554acfb7a3809a55aae6ba', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd7288068ddae484b8796d727e46c9bf3', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 765.681251] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce394bff-a264-4f50-a60b-30b6490f0e93 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.690931] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5421eda-946f-404c-a323-dd5934f73e1e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.721676] env[62066]: INFO nova.scheduler.client.report [None req-a4d93091-69b6-4b35-887c-e7b927c2300a tempest-ServerMetadataTestJSON-1590103242 tempest-ServerMetadataTestJSON-1590103242-project-member] Deleted allocations for instance 0ed07533-9e79-416d-a495-a8e20b97e108 [ 765.727317] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e141da8-64d5-4e98-9841-89a7501b49c2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.736056] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc143864-f7c6-49f4-a9f6-731773dd80b0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.749822] env[62066]: DEBUG nova.compute.provider_tree [None req-e9793203-98b3-4ca0-8a49-79767b1f52f7 tempest-ServersTestBootFromVolume-1535558002 tempest-ServersTestBootFromVolume-1535558002-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 765.803694] env[62066]: DEBUG nova.network.neutron [None req-7fc7916e-6e82-4e11-a2d0-03a259fc6439 tempest-ServerAddressesNegativeTestJSON-1804754632 tempest-ServerAddressesNegativeTestJSON-1804754632-project-member] [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] Successfully created port: 17996c34-360e-41b6-9b1e-f194859938a1 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 765.900447] env[62066]: DEBUG nova.compute.manager [None req-7fc7916e-6e82-4e11-a2d0-03a259fc6439 tempest-ServerAddressesNegativeTestJSON-1804754632 tempest-ServerAddressesNegativeTestJSON-1804754632-project-member] [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 766.232428] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a4d93091-69b6-4b35-887c-e7b927c2300a tempest-ServerMetadataTestJSON-1590103242 tempest-ServerMetadataTestJSON-1590103242-project-member] Lock "0ed07533-9e79-416d-a495-a8e20b97e108" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 156.131s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 766.255265] env[62066]: DEBUG nova.scheduler.client.report [None req-e9793203-98b3-4ca0-8a49-79767b1f52f7 tempest-ServersTestBootFromVolume-1535558002 tempest-ServersTestBootFromVolume-1535558002-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 766.735715] env[62066]: DEBUG nova.compute.manager [None req-3baf3239-22c0-4858-82f6-d3b75bd59a7c tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] [instance: 1e1a2ede-4ad8-4600-851e-6f2046b3f919] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 766.762020] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e9793203-98b3-4ca0-8a49-79767b1f52f7 tempest-ServersTestBootFromVolume-1535558002 tempest-ServersTestBootFromVolume-1535558002-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.872s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 766.763963] env[62066]: ERROR nova.compute.manager [None req-e9793203-98b3-4ca0-8a49-79767b1f52f7 tempest-ServersTestBootFromVolume-1535558002 tempest-ServersTestBootFromVolume-1535558002-project-member] [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port c44654b4-f098-40c1-93d9-b524a3b29c8c, please check neutron logs for more information. [ 766.763963] env[62066]: ERROR nova.compute.manager [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] Traceback (most recent call last): [ 766.763963] env[62066]: ERROR nova.compute.manager [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 766.763963] env[62066]: ERROR nova.compute.manager [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] self.driver.spawn(context, instance, image_meta, [ 766.763963] env[62066]: ERROR nova.compute.manager [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 766.763963] env[62066]: ERROR nova.compute.manager [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 766.763963] env[62066]: ERROR nova.compute.manager [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 766.763963] env[62066]: ERROR nova.compute.manager [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] vm_ref = self.build_virtual_machine(instance, [ 766.763963] env[62066]: ERROR nova.compute.manager [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 766.763963] env[62066]: ERROR nova.compute.manager [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] vif_infos = vmwarevif.get_vif_info(self._session, [ 766.763963] env[62066]: ERROR nova.compute.manager [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 766.764607] env[62066]: ERROR nova.compute.manager [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] for vif in network_info: [ 766.764607] env[62066]: ERROR nova.compute.manager [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 766.764607] env[62066]: ERROR nova.compute.manager [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] return self._sync_wrapper(fn, *args, **kwargs) [ 766.764607] env[62066]: ERROR nova.compute.manager [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 766.764607] env[62066]: ERROR nova.compute.manager [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] self.wait() [ 766.764607] env[62066]: ERROR nova.compute.manager [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 766.764607] env[62066]: ERROR nova.compute.manager [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] self[:] = self._gt.wait() [ 766.764607] env[62066]: ERROR nova.compute.manager [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 766.764607] env[62066]: ERROR nova.compute.manager [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] return self._exit_event.wait() [ 766.764607] env[62066]: ERROR nova.compute.manager [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 766.764607] env[62066]: ERROR nova.compute.manager [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] current.throw(*self._exc) [ 766.764607] env[62066]: ERROR nova.compute.manager [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 766.764607] env[62066]: ERROR nova.compute.manager [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] result = function(*args, **kwargs) [ 766.764910] env[62066]: ERROR nova.compute.manager [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 766.764910] env[62066]: ERROR nova.compute.manager [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] return func(*args, **kwargs) [ 766.764910] env[62066]: ERROR nova.compute.manager [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 766.764910] env[62066]: ERROR nova.compute.manager [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] raise e [ 766.764910] env[62066]: ERROR nova.compute.manager [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 766.764910] env[62066]: ERROR nova.compute.manager [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] nwinfo = self.network_api.allocate_for_instance( [ 766.764910] env[62066]: ERROR nova.compute.manager [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 766.764910] env[62066]: ERROR nova.compute.manager [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] created_port_ids = self._update_ports_for_instance( [ 766.764910] env[62066]: ERROR nova.compute.manager [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 766.764910] env[62066]: ERROR nova.compute.manager [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] with excutils.save_and_reraise_exception(): [ 766.764910] env[62066]: ERROR nova.compute.manager [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 766.764910] env[62066]: ERROR nova.compute.manager [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] self.force_reraise() [ 766.764910] env[62066]: ERROR nova.compute.manager [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 766.765253] env[62066]: ERROR nova.compute.manager [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] raise self.value [ 766.765253] env[62066]: ERROR nova.compute.manager [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 766.765253] env[62066]: ERROR nova.compute.manager [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] updated_port = self._update_port( [ 766.765253] env[62066]: ERROR nova.compute.manager [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 766.765253] env[62066]: ERROR nova.compute.manager [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] _ensure_no_port_binding_failure(port) [ 766.765253] env[62066]: ERROR nova.compute.manager [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 766.765253] env[62066]: ERROR nova.compute.manager [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] raise exception.PortBindingFailed(port_id=port['id']) [ 766.765253] env[62066]: ERROR nova.compute.manager [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] nova.exception.PortBindingFailed: Binding failed for port c44654b4-f098-40c1-93d9-b524a3b29c8c, please check neutron logs for more information. [ 766.765253] env[62066]: ERROR nova.compute.manager [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] [ 766.765253] env[62066]: DEBUG nova.compute.utils [None req-e9793203-98b3-4ca0-8a49-79767b1f52f7 tempest-ServersTestBootFromVolume-1535558002 tempest-ServersTestBootFromVolume-1535558002-project-member] [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] Binding failed for port c44654b4-f098-40c1-93d9-b524a3b29c8c, please check neutron logs for more information. {{(pid=62066) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 766.767909] env[62066]: DEBUG oslo_concurrency.lockutils [None req-478d4566-e74c-4f71-b94d-419dc0a31446 tempest-ServerPasswordTestJSON-849130511 tempest-ServerPasswordTestJSON-849130511-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 27.150s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 766.772661] env[62066]: DEBUG nova.compute.manager [req-c9d610cf-2a8a-45c5-88f9-5433d41e1738 req-3d522f2c-a432-41cb-9007-384cba24d36d service nova] [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] Received event network-changed-17996c34-360e-41b6-9b1e-f194859938a1 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 766.772982] env[62066]: DEBUG nova.compute.manager [req-c9d610cf-2a8a-45c5-88f9-5433d41e1738 req-3d522f2c-a432-41cb-9007-384cba24d36d service nova] [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] Refreshing instance network info cache due to event network-changed-17996c34-360e-41b6-9b1e-f194859938a1. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 766.773527] env[62066]: DEBUG oslo_concurrency.lockutils [req-c9d610cf-2a8a-45c5-88f9-5433d41e1738 req-3d522f2c-a432-41cb-9007-384cba24d36d service nova] Acquiring lock "refresh_cache-239f0b4b-32cf-4f9f-a47e-bac4235fb6ef" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 766.773742] env[62066]: DEBUG oslo_concurrency.lockutils [req-c9d610cf-2a8a-45c5-88f9-5433d41e1738 req-3d522f2c-a432-41cb-9007-384cba24d36d service nova] Acquired lock "refresh_cache-239f0b4b-32cf-4f9f-a47e-bac4235fb6ef" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 766.774141] env[62066]: DEBUG nova.network.neutron [req-c9d610cf-2a8a-45c5-88f9-5433d41e1738 req-3d522f2c-a432-41cb-9007-384cba24d36d service nova] [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] Refreshing network info cache for port 17996c34-360e-41b6-9b1e-f194859938a1 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 766.775625] env[62066]: DEBUG nova.compute.manager [None req-e9793203-98b3-4ca0-8a49-79767b1f52f7 tempest-ServersTestBootFromVolume-1535558002 tempest-ServersTestBootFromVolume-1535558002-project-member] [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] Build of instance 2b81d060-ecb7-4ae5-b51d-33197937a76f was re-scheduled: Binding failed for port c44654b4-f098-40c1-93d9-b524a3b29c8c, please check neutron logs for more information. {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 766.776422] env[62066]: DEBUG nova.compute.manager [None req-e9793203-98b3-4ca0-8a49-79767b1f52f7 tempest-ServersTestBootFromVolume-1535558002 tempest-ServersTestBootFromVolume-1535558002-project-member] [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] Unplugging VIFs for instance {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 766.776422] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e9793203-98b3-4ca0-8a49-79767b1f52f7 tempest-ServersTestBootFromVolume-1535558002 tempest-ServersTestBootFromVolume-1535558002-project-member] Acquiring lock "refresh_cache-2b81d060-ecb7-4ae5-b51d-33197937a76f" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 766.776536] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e9793203-98b3-4ca0-8a49-79767b1f52f7 tempest-ServersTestBootFromVolume-1535558002 tempest-ServersTestBootFromVolume-1535558002-project-member] Acquired lock "refresh_cache-2b81d060-ecb7-4ae5-b51d-33197937a76f" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 766.776981] env[62066]: DEBUG nova.network.neutron [None req-e9793203-98b3-4ca0-8a49-79767b1f52f7 tempest-ServersTestBootFromVolume-1535558002 tempest-ServersTestBootFromVolume-1535558002-project-member] [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 766.911050] env[62066]: DEBUG nova.compute.manager [None req-7fc7916e-6e82-4e11-a2d0-03a259fc6439 tempest-ServerAddressesNegativeTestJSON-1804754632 tempest-ServerAddressesNegativeTestJSON-1804754632-project-member] [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 766.936778] env[62066]: DEBUG nova.virt.hardware [None req-7fc7916e-6e82-4e11-a2d0-03a259fc6439 tempest-ServerAddressesNegativeTestJSON-1804754632 tempest-ServerAddressesNegativeTestJSON-1804754632-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-23T13:40:41Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-23T13:40:23Z,direct_url=,disk_format='vmdk',id=50ff584c-3b50-4395-af07-3e66769bc9f7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='eb52f7069a374c61ae946f052007c6d9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-23T13:40:24Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 766.936778] env[62066]: DEBUG nova.virt.hardware [None req-7fc7916e-6e82-4e11-a2d0-03a259fc6439 tempest-ServerAddressesNegativeTestJSON-1804754632 tempest-ServerAddressesNegativeTestJSON-1804754632-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 766.936950] env[62066]: DEBUG nova.virt.hardware [None req-7fc7916e-6e82-4e11-a2d0-03a259fc6439 tempest-ServerAddressesNegativeTestJSON-1804754632 tempest-ServerAddressesNegativeTestJSON-1804754632-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 766.937168] env[62066]: DEBUG nova.virt.hardware [None req-7fc7916e-6e82-4e11-a2d0-03a259fc6439 tempest-ServerAddressesNegativeTestJSON-1804754632 tempest-ServerAddressesNegativeTestJSON-1804754632-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 766.937366] env[62066]: DEBUG nova.virt.hardware [None req-7fc7916e-6e82-4e11-a2d0-03a259fc6439 tempest-ServerAddressesNegativeTestJSON-1804754632 tempest-ServerAddressesNegativeTestJSON-1804754632-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 766.938342] env[62066]: DEBUG nova.virt.hardware [None req-7fc7916e-6e82-4e11-a2d0-03a259fc6439 tempest-ServerAddressesNegativeTestJSON-1804754632 tempest-ServerAddressesNegativeTestJSON-1804754632-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 766.938342] env[62066]: DEBUG nova.virt.hardware [None req-7fc7916e-6e82-4e11-a2d0-03a259fc6439 tempest-ServerAddressesNegativeTestJSON-1804754632 tempest-ServerAddressesNegativeTestJSON-1804754632-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 766.938342] env[62066]: DEBUG nova.virt.hardware [None req-7fc7916e-6e82-4e11-a2d0-03a259fc6439 tempest-ServerAddressesNegativeTestJSON-1804754632 tempest-ServerAddressesNegativeTestJSON-1804754632-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 766.938342] env[62066]: DEBUG nova.virt.hardware [None req-7fc7916e-6e82-4e11-a2d0-03a259fc6439 tempest-ServerAddressesNegativeTestJSON-1804754632 tempest-ServerAddressesNegativeTestJSON-1804754632-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 766.938342] env[62066]: DEBUG nova.virt.hardware [None req-7fc7916e-6e82-4e11-a2d0-03a259fc6439 tempest-ServerAddressesNegativeTestJSON-1804754632 tempest-ServerAddressesNegativeTestJSON-1804754632-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 766.938616] env[62066]: DEBUG nova.virt.hardware [None req-7fc7916e-6e82-4e11-a2d0-03a259fc6439 tempest-ServerAddressesNegativeTestJSON-1804754632 tempest-ServerAddressesNegativeTestJSON-1804754632-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 766.939578] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-768fe6a6-8561-4fb6-a35f-1033d6f7714d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.948569] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f285796b-e026-4e22-a0c1-710bbee68d64 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.965641] env[62066]: ERROR nova.compute.manager [None req-7fc7916e-6e82-4e11-a2d0-03a259fc6439 tempest-ServerAddressesNegativeTestJSON-1804754632 tempest-ServerAddressesNegativeTestJSON-1804754632-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 17996c34-360e-41b6-9b1e-f194859938a1, please check neutron logs for more information. [ 766.965641] env[62066]: ERROR nova.compute.manager Traceback (most recent call last): [ 766.965641] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 766.965641] env[62066]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 766.965641] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 766.965641] env[62066]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 766.965641] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 766.965641] env[62066]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 766.965641] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 766.965641] env[62066]: ERROR nova.compute.manager self.force_reraise() [ 766.965641] env[62066]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 766.965641] env[62066]: ERROR nova.compute.manager raise self.value [ 766.965641] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 766.965641] env[62066]: ERROR nova.compute.manager updated_port = self._update_port( [ 766.965641] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 766.965641] env[62066]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 766.966191] env[62066]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 766.966191] env[62066]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 766.966191] env[62066]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 17996c34-360e-41b6-9b1e-f194859938a1, please check neutron logs for more information. [ 766.966191] env[62066]: ERROR nova.compute.manager [ 766.966191] env[62066]: Traceback (most recent call last): [ 766.966191] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 766.966191] env[62066]: listener.cb(fileno) [ 766.966191] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 766.966191] env[62066]: result = function(*args, **kwargs) [ 766.966191] env[62066]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 766.966191] env[62066]: return func(*args, **kwargs) [ 766.966191] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 766.966191] env[62066]: raise e [ 766.966191] env[62066]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 766.966191] env[62066]: nwinfo = self.network_api.allocate_for_instance( [ 766.966191] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 766.966191] env[62066]: created_port_ids = self._update_ports_for_instance( [ 766.966191] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 766.966191] env[62066]: with excutils.save_and_reraise_exception(): [ 766.966191] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 766.966191] env[62066]: self.force_reraise() [ 766.966191] env[62066]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 766.966191] env[62066]: raise self.value [ 766.966191] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 766.966191] env[62066]: updated_port = self._update_port( [ 766.966191] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 766.966191] env[62066]: _ensure_no_port_binding_failure(port) [ 766.966191] env[62066]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 766.966191] env[62066]: raise exception.PortBindingFailed(port_id=port['id']) [ 766.967268] env[62066]: nova.exception.PortBindingFailed: Binding failed for port 17996c34-360e-41b6-9b1e-f194859938a1, please check neutron logs for more information. [ 766.967268] env[62066]: Removing descriptor: 20 [ 766.967268] env[62066]: ERROR nova.compute.manager [None req-7fc7916e-6e82-4e11-a2d0-03a259fc6439 tempest-ServerAddressesNegativeTestJSON-1804754632 tempest-ServerAddressesNegativeTestJSON-1804754632-project-member] [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 17996c34-360e-41b6-9b1e-f194859938a1, please check neutron logs for more information. [ 766.967268] env[62066]: ERROR nova.compute.manager [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] Traceback (most recent call last): [ 766.967268] env[62066]: ERROR nova.compute.manager [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 766.967268] env[62066]: ERROR nova.compute.manager [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] yield resources [ 766.967268] env[62066]: ERROR nova.compute.manager [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 766.967268] env[62066]: ERROR nova.compute.manager [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] self.driver.spawn(context, instance, image_meta, [ 766.967268] env[62066]: ERROR nova.compute.manager [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 766.967268] env[62066]: ERROR nova.compute.manager [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] self._vmops.spawn(context, instance, image_meta, injected_files, [ 766.967268] env[62066]: ERROR nova.compute.manager [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 766.967268] env[62066]: ERROR nova.compute.manager [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] vm_ref = self.build_virtual_machine(instance, [ 766.967920] env[62066]: ERROR nova.compute.manager [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 766.967920] env[62066]: ERROR nova.compute.manager [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] vif_infos = vmwarevif.get_vif_info(self._session, [ 766.967920] env[62066]: ERROR nova.compute.manager [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 766.967920] env[62066]: ERROR nova.compute.manager [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] for vif in network_info: [ 766.967920] env[62066]: ERROR nova.compute.manager [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 766.967920] env[62066]: ERROR nova.compute.manager [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] return self._sync_wrapper(fn, *args, **kwargs) [ 766.967920] env[62066]: ERROR nova.compute.manager [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 766.967920] env[62066]: ERROR nova.compute.manager [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] self.wait() [ 766.967920] env[62066]: ERROR nova.compute.manager [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 766.967920] env[62066]: ERROR nova.compute.manager [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] self[:] = self._gt.wait() [ 766.967920] env[62066]: ERROR nova.compute.manager [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 766.967920] env[62066]: ERROR nova.compute.manager [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] return self._exit_event.wait() [ 766.967920] env[62066]: ERROR nova.compute.manager [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 766.968292] env[62066]: ERROR nova.compute.manager [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] result = hub.switch() [ 766.968292] env[62066]: ERROR nova.compute.manager [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 766.968292] env[62066]: ERROR nova.compute.manager [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] return self.greenlet.switch() [ 766.968292] env[62066]: ERROR nova.compute.manager [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 766.968292] env[62066]: ERROR nova.compute.manager [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] result = function(*args, **kwargs) [ 766.968292] env[62066]: ERROR nova.compute.manager [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 766.968292] env[62066]: ERROR nova.compute.manager [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] return func(*args, **kwargs) [ 766.968292] env[62066]: ERROR nova.compute.manager [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 766.968292] env[62066]: ERROR nova.compute.manager [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] raise e [ 766.968292] env[62066]: ERROR nova.compute.manager [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 766.968292] env[62066]: ERROR nova.compute.manager [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] nwinfo = self.network_api.allocate_for_instance( [ 766.968292] env[62066]: ERROR nova.compute.manager [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 766.968292] env[62066]: ERROR nova.compute.manager [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] created_port_ids = self._update_ports_for_instance( [ 766.968918] env[62066]: ERROR nova.compute.manager [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 766.968918] env[62066]: ERROR nova.compute.manager [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] with excutils.save_and_reraise_exception(): [ 766.968918] env[62066]: ERROR nova.compute.manager [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 766.968918] env[62066]: ERROR nova.compute.manager [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] self.force_reraise() [ 766.968918] env[62066]: ERROR nova.compute.manager [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 766.968918] env[62066]: ERROR nova.compute.manager [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] raise self.value [ 766.968918] env[62066]: ERROR nova.compute.manager [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 766.968918] env[62066]: ERROR nova.compute.manager [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] updated_port = self._update_port( [ 766.968918] env[62066]: ERROR nova.compute.manager [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 766.968918] env[62066]: ERROR nova.compute.manager [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] _ensure_no_port_binding_failure(port) [ 766.968918] env[62066]: ERROR nova.compute.manager [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 766.968918] env[62066]: ERROR nova.compute.manager [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] raise exception.PortBindingFailed(port_id=port['id']) [ 766.969323] env[62066]: ERROR nova.compute.manager [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] nova.exception.PortBindingFailed: Binding failed for port 17996c34-360e-41b6-9b1e-f194859938a1, please check neutron logs for more information. [ 766.969323] env[62066]: ERROR nova.compute.manager [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] [ 766.969323] env[62066]: INFO nova.compute.manager [None req-7fc7916e-6e82-4e11-a2d0-03a259fc6439 tempest-ServerAddressesNegativeTestJSON-1804754632 tempest-ServerAddressesNegativeTestJSON-1804754632-project-member] [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] Terminating instance [ 766.969796] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7fc7916e-6e82-4e11-a2d0-03a259fc6439 tempest-ServerAddressesNegativeTestJSON-1804754632 tempest-ServerAddressesNegativeTestJSON-1804754632-project-member] Acquiring lock "refresh_cache-239f0b4b-32cf-4f9f-a47e-bac4235fb6ef" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 767.261855] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3baf3239-22c0-4858-82f6-d3b75bd59a7c tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 767.303140] env[62066]: DEBUG nova.network.neutron [req-c9d610cf-2a8a-45c5-88f9-5433d41e1738 req-3d522f2c-a432-41cb-9007-384cba24d36d service nova] [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 767.310201] env[62066]: DEBUG nova.network.neutron [None req-e9793203-98b3-4ca0-8a49-79767b1f52f7 tempest-ServersTestBootFromVolume-1535558002 tempest-ServersTestBootFromVolume-1535558002-project-member] [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 768.228998] env[62066]: DEBUG nova.network.neutron [req-c9d610cf-2a8a-45c5-88f9-5433d41e1738 req-3d522f2c-a432-41cb-9007-384cba24d36d service nova] [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 768.230681] env[62066]: DEBUG nova.network.neutron [None req-e9793203-98b3-4ca0-8a49-79767b1f52f7 tempest-ServersTestBootFromVolume-1535558002 tempest-ServersTestBootFromVolume-1535558002-project-member] [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 768.387034] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-367db220-a7f3-4c4c-b005-b19510e25975 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.393603] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cb4b30f-25d0-490e-ba6f-8d1416028b4e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.422196] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88b69b86-ddd8-4c7d-a914-4d8e04ee3135 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.429711] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97d242df-9cd3-48f7-8f18-404d02019257 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.442356] env[62066]: DEBUG nova.compute.provider_tree [None req-478d4566-e74c-4f71-b94d-419dc0a31446 tempest-ServerPasswordTestJSON-849130511 tempest-ServerPasswordTestJSON-849130511-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 768.733011] env[62066]: DEBUG oslo_concurrency.lockutils [req-c9d610cf-2a8a-45c5-88f9-5433d41e1738 req-3d522f2c-a432-41cb-9007-384cba24d36d service nova] Releasing lock "refresh_cache-239f0b4b-32cf-4f9f-a47e-bac4235fb6ef" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 768.733614] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e9793203-98b3-4ca0-8a49-79767b1f52f7 tempest-ServersTestBootFromVolume-1535558002 tempest-ServersTestBootFromVolume-1535558002-project-member] Releasing lock "refresh_cache-2b81d060-ecb7-4ae5-b51d-33197937a76f" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 768.733861] env[62066]: DEBUG nova.compute.manager [None req-e9793203-98b3-4ca0-8a49-79767b1f52f7 tempest-ServersTestBootFromVolume-1535558002 tempest-ServersTestBootFromVolume-1535558002-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 768.734067] env[62066]: DEBUG nova.compute.manager [None req-e9793203-98b3-4ca0-8a49-79767b1f52f7 tempest-ServersTestBootFromVolume-1535558002 tempest-ServersTestBootFromVolume-1535558002-project-member] [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 768.734222] env[62066]: DEBUG nova.network.neutron [None req-e9793203-98b3-4ca0-8a49-79767b1f52f7 tempest-ServersTestBootFromVolume-1535558002 tempest-ServersTestBootFromVolume-1535558002-project-member] [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 768.735901] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7fc7916e-6e82-4e11-a2d0-03a259fc6439 tempest-ServerAddressesNegativeTestJSON-1804754632 tempest-ServerAddressesNegativeTestJSON-1804754632-project-member] Acquired lock "refresh_cache-239f0b4b-32cf-4f9f-a47e-bac4235fb6ef" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 768.736097] env[62066]: DEBUG nova.network.neutron [None req-7fc7916e-6e82-4e11-a2d0-03a259fc6439 tempest-ServerAddressesNegativeTestJSON-1804754632 tempest-ServerAddressesNegativeTestJSON-1804754632-project-member] [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 768.755992] env[62066]: DEBUG nova.network.neutron [None req-e9793203-98b3-4ca0-8a49-79767b1f52f7 tempest-ServersTestBootFromVolume-1535558002 tempest-ServersTestBootFromVolume-1535558002-project-member] [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 768.786091] env[62066]: DEBUG nova.compute.manager [req-09e4d3e8-3052-4bc8-8fe6-f710e141a38e req-bf89f43a-fd08-4497-b472-8a1c77f4183f service nova] [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] Received event network-vif-deleted-17996c34-360e-41b6-9b1e-f194859938a1 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 768.945901] env[62066]: DEBUG nova.scheduler.client.report [None req-478d4566-e74c-4f71-b94d-419dc0a31446 tempest-ServerPasswordTestJSON-849130511 tempest-ServerPasswordTestJSON-849130511-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 769.255338] env[62066]: DEBUG nova.network.neutron [None req-7fc7916e-6e82-4e11-a2d0-03a259fc6439 tempest-ServerAddressesNegativeTestJSON-1804754632 tempest-ServerAddressesNegativeTestJSON-1804754632-project-member] [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 769.261506] env[62066]: DEBUG nova.network.neutron [None req-e9793203-98b3-4ca0-8a49-79767b1f52f7 tempest-ServersTestBootFromVolume-1535558002 tempest-ServersTestBootFromVolume-1535558002-project-member] [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 769.351600] env[62066]: DEBUG nova.network.neutron [None req-7fc7916e-6e82-4e11-a2d0-03a259fc6439 tempest-ServerAddressesNegativeTestJSON-1804754632 tempest-ServerAddressesNegativeTestJSON-1804754632-project-member] [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 769.451367] env[62066]: DEBUG oslo_concurrency.lockutils [None req-478d4566-e74c-4f71-b94d-419dc0a31446 tempest-ServerPasswordTestJSON-849130511 tempest-ServerPasswordTestJSON-849130511-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.683s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 769.451970] env[62066]: ERROR nova.compute.manager [None req-478d4566-e74c-4f71-b94d-419dc0a31446 tempest-ServerPasswordTestJSON-849130511 tempest-ServerPasswordTestJSON-849130511-project-member] [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 43f773e7-1b4c-4da1-ab93-a42f9d34f998, please check neutron logs for more information. [ 769.451970] env[62066]: ERROR nova.compute.manager [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] Traceback (most recent call last): [ 769.451970] env[62066]: ERROR nova.compute.manager [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 769.451970] env[62066]: ERROR nova.compute.manager [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] self.driver.spawn(context, instance, image_meta, [ 769.451970] env[62066]: ERROR nova.compute.manager [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 769.451970] env[62066]: ERROR nova.compute.manager [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 769.451970] env[62066]: ERROR nova.compute.manager [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 769.451970] env[62066]: ERROR nova.compute.manager [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] vm_ref = self.build_virtual_machine(instance, [ 769.451970] env[62066]: ERROR nova.compute.manager [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 769.451970] env[62066]: ERROR nova.compute.manager [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] vif_infos = vmwarevif.get_vif_info(self._session, [ 769.451970] env[62066]: ERROR nova.compute.manager [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 769.452352] env[62066]: ERROR nova.compute.manager [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] for vif in network_info: [ 769.452352] env[62066]: ERROR nova.compute.manager [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 769.452352] env[62066]: ERROR nova.compute.manager [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] return self._sync_wrapper(fn, *args, **kwargs) [ 769.452352] env[62066]: ERROR nova.compute.manager [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 769.452352] env[62066]: ERROR nova.compute.manager [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] self.wait() [ 769.452352] env[62066]: ERROR nova.compute.manager [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 769.452352] env[62066]: ERROR nova.compute.manager [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] self[:] = self._gt.wait() [ 769.452352] env[62066]: ERROR nova.compute.manager [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 769.452352] env[62066]: ERROR nova.compute.manager [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] return self._exit_event.wait() [ 769.452352] env[62066]: ERROR nova.compute.manager [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 769.452352] env[62066]: ERROR nova.compute.manager [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] current.throw(*self._exc) [ 769.452352] env[62066]: ERROR nova.compute.manager [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 769.452352] env[62066]: ERROR nova.compute.manager [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] result = function(*args, **kwargs) [ 769.452704] env[62066]: ERROR nova.compute.manager [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 769.452704] env[62066]: ERROR nova.compute.manager [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] return func(*args, **kwargs) [ 769.452704] env[62066]: ERROR nova.compute.manager [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 769.452704] env[62066]: ERROR nova.compute.manager [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] raise e [ 769.452704] env[62066]: ERROR nova.compute.manager [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 769.452704] env[62066]: ERROR nova.compute.manager [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] nwinfo = self.network_api.allocate_for_instance( [ 769.452704] env[62066]: ERROR nova.compute.manager [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 769.452704] env[62066]: ERROR nova.compute.manager [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] created_port_ids = self._update_ports_for_instance( [ 769.452704] env[62066]: ERROR nova.compute.manager [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 769.452704] env[62066]: ERROR nova.compute.manager [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] with excutils.save_and_reraise_exception(): [ 769.452704] env[62066]: ERROR nova.compute.manager [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 769.452704] env[62066]: ERROR nova.compute.manager [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] self.force_reraise() [ 769.452704] env[62066]: ERROR nova.compute.manager [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 769.453058] env[62066]: ERROR nova.compute.manager [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] raise self.value [ 769.453058] env[62066]: ERROR nova.compute.manager [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 769.453058] env[62066]: ERROR nova.compute.manager [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] updated_port = self._update_port( [ 769.453058] env[62066]: ERROR nova.compute.manager [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 769.453058] env[62066]: ERROR nova.compute.manager [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] _ensure_no_port_binding_failure(port) [ 769.453058] env[62066]: ERROR nova.compute.manager [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 769.453058] env[62066]: ERROR nova.compute.manager [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] raise exception.PortBindingFailed(port_id=port['id']) [ 769.453058] env[62066]: ERROR nova.compute.manager [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] nova.exception.PortBindingFailed: Binding failed for port 43f773e7-1b4c-4da1-ab93-a42f9d34f998, please check neutron logs for more information. [ 769.453058] env[62066]: ERROR nova.compute.manager [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] [ 769.453058] env[62066]: DEBUG nova.compute.utils [None req-478d4566-e74c-4f71-b94d-419dc0a31446 tempest-ServerPasswordTestJSON-849130511 tempest-ServerPasswordTestJSON-849130511-project-member] [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] Binding failed for port 43f773e7-1b4c-4da1-ab93-a42f9d34f998, please check neutron logs for more information. {{(pid=62066) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 769.453902] env[62066]: DEBUG oslo_concurrency.lockutils [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.197s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 769.455756] env[62066]: INFO nova.compute.claims [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 0ebba3b2-f82d-4c1d-b01a-f75c3559c117] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 769.458684] env[62066]: DEBUG nova.compute.manager [None req-478d4566-e74c-4f71-b94d-419dc0a31446 tempest-ServerPasswordTestJSON-849130511 tempest-ServerPasswordTestJSON-849130511-project-member] [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] Build of instance 04d25af4-0e73-4650-9c3d-85817754bac9 was re-scheduled: Binding failed for port 43f773e7-1b4c-4da1-ab93-a42f9d34f998, please check neutron logs for more information. {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 769.460955] env[62066]: DEBUG nova.compute.manager [None req-478d4566-e74c-4f71-b94d-419dc0a31446 tempest-ServerPasswordTestJSON-849130511 tempest-ServerPasswordTestJSON-849130511-project-member] [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] Unplugging VIFs for instance {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 769.460955] env[62066]: DEBUG oslo_concurrency.lockutils [None req-478d4566-e74c-4f71-b94d-419dc0a31446 tempest-ServerPasswordTestJSON-849130511 tempest-ServerPasswordTestJSON-849130511-project-member] Acquiring lock "refresh_cache-04d25af4-0e73-4650-9c3d-85817754bac9" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 769.460955] env[62066]: DEBUG oslo_concurrency.lockutils [None req-478d4566-e74c-4f71-b94d-419dc0a31446 tempest-ServerPasswordTestJSON-849130511 tempest-ServerPasswordTestJSON-849130511-project-member] Acquired lock "refresh_cache-04d25af4-0e73-4650-9c3d-85817754bac9" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 769.460955] env[62066]: DEBUG nova.network.neutron [None req-478d4566-e74c-4f71-b94d-419dc0a31446 tempest-ServerPasswordTestJSON-849130511 tempest-ServerPasswordTestJSON-849130511-project-member] [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 769.768856] env[62066]: INFO nova.compute.manager [None req-e9793203-98b3-4ca0-8a49-79767b1f52f7 tempest-ServersTestBootFromVolume-1535558002 tempest-ServersTestBootFromVolume-1535558002-project-member] [instance: 2b81d060-ecb7-4ae5-b51d-33197937a76f] Took 1.03 seconds to deallocate network for instance. [ 769.854153] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7fc7916e-6e82-4e11-a2d0-03a259fc6439 tempest-ServerAddressesNegativeTestJSON-1804754632 tempest-ServerAddressesNegativeTestJSON-1804754632-project-member] Releasing lock "refresh_cache-239f0b4b-32cf-4f9f-a47e-bac4235fb6ef" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 769.857088] env[62066]: DEBUG nova.compute.manager [None req-7fc7916e-6e82-4e11-a2d0-03a259fc6439 tempest-ServerAddressesNegativeTestJSON-1804754632 tempest-ServerAddressesNegativeTestJSON-1804754632-project-member] [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 769.857088] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-7fc7916e-6e82-4e11-a2d0-03a259fc6439 tempest-ServerAddressesNegativeTestJSON-1804754632 tempest-ServerAddressesNegativeTestJSON-1804754632-project-member] [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 769.857088] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b85f7152-523e-469b-8c13-74893f9c73c0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.864480] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-954648f9-7545-4efb-95a7-5aa441f6dad5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.888898] env[62066]: WARNING nova.virt.vmwareapi.vmops [None req-7fc7916e-6e82-4e11-a2d0-03a259fc6439 tempest-ServerAddressesNegativeTestJSON-1804754632 tempest-ServerAddressesNegativeTestJSON-1804754632-project-member] [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef could not be found. [ 769.889313] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-7fc7916e-6e82-4e11-a2d0-03a259fc6439 tempest-ServerAddressesNegativeTestJSON-1804754632 tempest-ServerAddressesNegativeTestJSON-1804754632-project-member] [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 769.889658] env[62066]: INFO nova.compute.manager [None req-7fc7916e-6e82-4e11-a2d0-03a259fc6439 tempest-ServerAddressesNegativeTestJSON-1804754632 tempest-ServerAddressesNegativeTestJSON-1804754632-project-member] [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] Took 0.03 seconds to destroy the instance on the hypervisor. [ 769.890034] env[62066]: DEBUG oslo.service.loopingcall [None req-7fc7916e-6e82-4e11-a2d0-03a259fc6439 tempest-ServerAddressesNegativeTestJSON-1804754632 tempest-ServerAddressesNegativeTestJSON-1804754632-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 769.890378] env[62066]: DEBUG nova.compute.manager [-] [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 769.890559] env[62066]: DEBUG nova.network.neutron [-] [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 769.907965] env[62066]: DEBUG nova.network.neutron [-] [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 769.983780] env[62066]: DEBUG nova.network.neutron [None req-478d4566-e74c-4f71-b94d-419dc0a31446 tempest-ServerPasswordTestJSON-849130511 tempest-ServerPasswordTestJSON-849130511-project-member] [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 770.051733] env[62066]: DEBUG nova.network.neutron [None req-478d4566-e74c-4f71-b94d-419dc0a31446 tempest-ServerPasswordTestJSON-849130511 tempest-ServerPasswordTestJSON-849130511-project-member] [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 770.410804] env[62066]: DEBUG nova.network.neutron [-] [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 770.554694] env[62066]: DEBUG oslo_concurrency.lockutils [None req-478d4566-e74c-4f71-b94d-419dc0a31446 tempest-ServerPasswordTestJSON-849130511 tempest-ServerPasswordTestJSON-849130511-project-member] Releasing lock "refresh_cache-04d25af4-0e73-4650-9c3d-85817754bac9" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 770.555109] env[62066]: DEBUG nova.compute.manager [None req-478d4566-e74c-4f71-b94d-419dc0a31446 tempest-ServerPasswordTestJSON-849130511 tempest-ServerPasswordTestJSON-849130511-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 770.555387] env[62066]: DEBUG nova.compute.manager [None req-478d4566-e74c-4f71-b94d-419dc0a31446 tempest-ServerPasswordTestJSON-849130511 tempest-ServerPasswordTestJSON-849130511-project-member] [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 770.555572] env[62066]: DEBUG nova.network.neutron [None req-478d4566-e74c-4f71-b94d-419dc0a31446 tempest-ServerPasswordTestJSON-849130511 tempest-ServerPasswordTestJSON-849130511-project-member] [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 770.572430] env[62066]: DEBUG nova.network.neutron [None req-478d4566-e74c-4f71-b94d-419dc0a31446 tempest-ServerPasswordTestJSON-849130511 tempest-ServerPasswordTestJSON-849130511-project-member] [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 770.715358] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b0ca88c-cb34-4532-a3b2-43b6b6c10a64 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.723094] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-590a3b62-72c4-4730-b915-7ab0599ecc5c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.751693] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf665838-031a-4108-8b2b-1ed26bfd63c9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.758517] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b49c088b-4da3-44db-b702-bbfa45a51717 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.771455] env[62066]: DEBUG nova.compute.provider_tree [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 770.798116] env[62066]: INFO nova.scheduler.client.report [None req-e9793203-98b3-4ca0-8a49-79767b1f52f7 tempest-ServersTestBootFromVolume-1535558002 tempest-ServersTestBootFromVolume-1535558002-project-member] Deleted allocations for instance 2b81d060-ecb7-4ae5-b51d-33197937a76f [ 770.913290] env[62066]: INFO nova.compute.manager [-] [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] Took 1.02 seconds to deallocate network for instance. [ 770.915568] env[62066]: DEBUG nova.compute.claims [None req-7fc7916e-6e82-4e11-a2d0-03a259fc6439 tempest-ServerAddressesNegativeTestJSON-1804754632 tempest-ServerAddressesNegativeTestJSON-1804754632-project-member] [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] Aborting claim: {{(pid=62066) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 770.915748] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7fc7916e-6e82-4e11-a2d0-03a259fc6439 tempest-ServerAddressesNegativeTestJSON-1804754632 tempest-ServerAddressesNegativeTestJSON-1804754632-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 771.076030] env[62066]: DEBUG nova.network.neutron [None req-478d4566-e74c-4f71-b94d-419dc0a31446 tempest-ServerPasswordTestJSON-849130511 tempest-ServerPasswordTestJSON-849130511-project-member] [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 771.274845] env[62066]: DEBUG nova.scheduler.client.report [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 771.305913] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e9793203-98b3-4ca0-8a49-79767b1f52f7 tempest-ServersTestBootFromVolume-1535558002 tempest-ServersTestBootFromVolume-1535558002-project-member] Lock "2b81d060-ecb7-4ae5-b51d-33197937a76f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 157.076s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 771.578795] env[62066]: INFO nova.compute.manager [None req-478d4566-e74c-4f71-b94d-419dc0a31446 tempest-ServerPasswordTestJSON-849130511 tempest-ServerPasswordTestJSON-849130511-project-member] [instance: 04d25af4-0e73-4650-9c3d-85817754bac9] Took 1.02 seconds to deallocate network for instance. [ 771.780405] env[62066]: DEBUG oslo_concurrency.lockutils [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.325s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 771.780860] env[62066]: DEBUG nova.compute.manager [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 0ebba3b2-f82d-4c1d-b01a-f75c3559c117] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 771.783471] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 26.929s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 771.810562] env[62066]: DEBUG nova.compute.manager [None req-86820846-d757-4de7-918c-ad5501b560db tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 772.287805] env[62066]: DEBUG nova.compute.utils [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 772.292900] env[62066]: DEBUG nova.compute.manager [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 0ebba3b2-f82d-4c1d-b01a-f75c3559c117] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 772.293094] env[62066]: DEBUG nova.network.neutron [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 0ebba3b2-f82d-4c1d-b01a-f75c3559c117] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 772.334365] env[62066]: DEBUG oslo_concurrency.lockutils [None req-86820846-d757-4de7-918c-ad5501b560db tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 772.361489] env[62066]: DEBUG nova.policy [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '236eede7716d4664a75f5df76aa403ce', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7664e11610664ce5a85a2ad8758986f8', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 772.591535] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43106397-33ab-47f0-8af0-19f1efdd2403 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.601497] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd0af5a3-98d0-477f-b70a-d190ba3d04d8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.633332] env[62066]: INFO nova.scheduler.client.report [None req-478d4566-e74c-4f71-b94d-419dc0a31446 tempest-ServerPasswordTestJSON-849130511 tempest-ServerPasswordTestJSON-849130511-project-member] Deleted allocations for instance 04d25af4-0e73-4650-9c3d-85817754bac9 [ 772.644147] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c33062d-88ac-4787-98a5-540fe7370ed1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.653272] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-626e433a-f3ce-422a-bee5-1192ae1e81a3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.668485] env[62066]: DEBUG nova.compute.provider_tree [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 772.798074] env[62066]: DEBUG nova.compute.manager [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 0ebba3b2-f82d-4c1d-b01a-f75c3559c117] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 773.000555] env[62066]: DEBUG nova.network.neutron [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 0ebba3b2-f82d-4c1d-b01a-f75c3559c117] Successfully created port: d9ded989-5fd8-4b73-86d4-dd8d8247d504 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 773.149897] env[62066]: DEBUG oslo_concurrency.lockutils [None req-478d4566-e74c-4f71-b94d-419dc0a31446 tempest-ServerPasswordTestJSON-849130511 tempest-ServerPasswordTestJSON-849130511-project-member] Lock "04d25af4-0e73-4650-9c3d-85817754bac9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 153.394s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 773.175825] env[62066]: DEBUG nova.scheduler.client.report [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 773.651351] env[62066]: DEBUG nova.compute.manager [None req-c70bcb7f-78ff-424d-8121-684c45ecd12b tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 02fb3e92-5dd6-4b1e-a6e4-d60d3fc85b07] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 773.679474] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.896s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 773.682614] env[62066]: ERROR nova.compute.manager [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] [instance: d973b09f-7727-4055-b05e-123079df9ce4] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 72daca54-f990-4f16-9362-cccaded8dba7, please check neutron logs for more information. [ 773.682614] env[62066]: ERROR nova.compute.manager [instance: d973b09f-7727-4055-b05e-123079df9ce4] Traceback (most recent call last): [ 773.682614] env[62066]: ERROR nova.compute.manager [instance: d973b09f-7727-4055-b05e-123079df9ce4] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 773.682614] env[62066]: ERROR nova.compute.manager [instance: d973b09f-7727-4055-b05e-123079df9ce4] self.driver.spawn(context, instance, image_meta, [ 773.682614] env[62066]: ERROR nova.compute.manager [instance: d973b09f-7727-4055-b05e-123079df9ce4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 773.682614] env[62066]: ERROR nova.compute.manager [instance: d973b09f-7727-4055-b05e-123079df9ce4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 773.682614] env[62066]: ERROR nova.compute.manager [instance: d973b09f-7727-4055-b05e-123079df9ce4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 773.682614] env[62066]: ERROR nova.compute.manager [instance: d973b09f-7727-4055-b05e-123079df9ce4] vm_ref = self.build_virtual_machine(instance, [ 773.682614] env[62066]: ERROR nova.compute.manager [instance: d973b09f-7727-4055-b05e-123079df9ce4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 773.682614] env[62066]: ERROR nova.compute.manager [instance: d973b09f-7727-4055-b05e-123079df9ce4] vif_infos = vmwarevif.get_vif_info(self._session, [ 773.682614] env[62066]: ERROR nova.compute.manager [instance: d973b09f-7727-4055-b05e-123079df9ce4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 773.683109] env[62066]: ERROR nova.compute.manager [instance: d973b09f-7727-4055-b05e-123079df9ce4] for vif in network_info: [ 773.683109] env[62066]: ERROR nova.compute.manager [instance: d973b09f-7727-4055-b05e-123079df9ce4] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 773.683109] env[62066]: ERROR nova.compute.manager [instance: d973b09f-7727-4055-b05e-123079df9ce4] return self._sync_wrapper(fn, *args, **kwargs) [ 773.683109] env[62066]: ERROR nova.compute.manager [instance: d973b09f-7727-4055-b05e-123079df9ce4] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 773.683109] env[62066]: ERROR nova.compute.manager [instance: d973b09f-7727-4055-b05e-123079df9ce4] self.wait() [ 773.683109] env[62066]: ERROR nova.compute.manager [instance: d973b09f-7727-4055-b05e-123079df9ce4] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 773.683109] env[62066]: ERROR nova.compute.manager [instance: d973b09f-7727-4055-b05e-123079df9ce4] self[:] = self._gt.wait() [ 773.683109] env[62066]: ERROR nova.compute.manager [instance: d973b09f-7727-4055-b05e-123079df9ce4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 773.683109] env[62066]: ERROR nova.compute.manager [instance: d973b09f-7727-4055-b05e-123079df9ce4] return self._exit_event.wait() [ 773.683109] env[62066]: ERROR nova.compute.manager [instance: d973b09f-7727-4055-b05e-123079df9ce4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 773.683109] env[62066]: ERROR nova.compute.manager [instance: d973b09f-7727-4055-b05e-123079df9ce4] result = hub.switch() [ 773.683109] env[62066]: ERROR nova.compute.manager [instance: d973b09f-7727-4055-b05e-123079df9ce4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 773.683109] env[62066]: ERROR nova.compute.manager [instance: d973b09f-7727-4055-b05e-123079df9ce4] return self.greenlet.switch() [ 773.683491] env[62066]: ERROR nova.compute.manager [instance: d973b09f-7727-4055-b05e-123079df9ce4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 773.683491] env[62066]: ERROR nova.compute.manager [instance: d973b09f-7727-4055-b05e-123079df9ce4] result = function(*args, **kwargs) [ 773.683491] env[62066]: ERROR nova.compute.manager [instance: d973b09f-7727-4055-b05e-123079df9ce4] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 773.683491] env[62066]: ERROR nova.compute.manager [instance: d973b09f-7727-4055-b05e-123079df9ce4] return func(*args, **kwargs) [ 773.683491] env[62066]: ERROR nova.compute.manager [instance: d973b09f-7727-4055-b05e-123079df9ce4] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 773.683491] env[62066]: ERROR nova.compute.manager [instance: d973b09f-7727-4055-b05e-123079df9ce4] raise e [ 773.683491] env[62066]: ERROR nova.compute.manager [instance: d973b09f-7727-4055-b05e-123079df9ce4] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 773.683491] env[62066]: ERROR nova.compute.manager [instance: d973b09f-7727-4055-b05e-123079df9ce4] nwinfo = self.network_api.allocate_for_instance( [ 773.683491] env[62066]: ERROR nova.compute.manager [instance: d973b09f-7727-4055-b05e-123079df9ce4] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 773.683491] env[62066]: ERROR nova.compute.manager [instance: d973b09f-7727-4055-b05e-123079df9ce4] created_port_ids = self._update_ports_for_instance( [ 773.683491] env[62066]: ERROR nova.compute.manager [instance: d973b09f-7727-4055-b05e-123079df9ce4] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 773.683491] env[62066]: ERROR nova.compute.manager [instance: d973b09f-7727-4055-b05e-123079df9ce4] with excutils.save_and_reraise_exception(): [ 773.683491] env[62066]: ERROR nova.compute.manager [instance: d973b09f-7727-4055-b05e-123079df9ce4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 773.684419] env[62066]: ERROR nova.compute.manager [instance: d973b09f-7727-4055-b05e-123079df9ce4] self.force_reraise() [ 773.684419] env[62066]: ERROR nova.compute.manager [instance: d973b09f-7727-4055-b05e-123079df9ce4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 773.684419] env[62066]: ERROR nova.compute.manager [instance: d973b09f-7727-4055-b05e-123079df9ce4] raise self.value [ 773.684419] env[62066]: ERROR nova.compute.manager [instance: d973b09f-7727-4055-b05e-123079df9ce4] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 773.684419] env[62066]: ERROR nova.compute.manager [instance: d973b09f-7727-4055-b05e-123079df9ce4] updated_port = self._update_port( [ 773.684419] env[62066]: ERROR nova.compute.manager [instance: d973b09f-7727-4055-b05e-123079df9ce4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 773.684419] env[62066]: ERROR nova.compute.manager [instance: d973b09f-7727-4055-b05e-123079df9ce4] _ensure_no_port_binding_failure(port) [ 773.684419] env[62066]: ERROR nova.compute.manager [instance: d973b09f-7727-4055-b05e-123079df9ce4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 773.684419] env[62066]: ERROR nova.compute.manager [instance: d973b09f-7727-4055-b05e-123079df9ce4] raise exception.PortBindingFailed(port_id=port['id']) [ 773.684419] env[62066]: ERROR nova.compute.manager [instance: d973b09f-7727-4055-b05e-123079df9ce4] nova.exception.PortBindingFailed: Binding failed for port 72daca54-f990-4f16-9362-cccaded8dba7, please check neutron logs for more information. [ 773.684419] env[62066]: ERROR nova.compute.manager [instance: d973b09f-7727-4055-b05e-123079df9ce4] [ 773.684683] env[62066]: DEBUG nova.compute.utils [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] [instance: d973b09f-7727-4055-b05e-123079df9ce4] Binding failed for port 72daca54-f990-4f16-9362-cccaded8dba7, please check neutron logs for more information. {{(pid=62066) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 773.686965] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 26.916s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 773.690498] env[62066]: DEBUG nova.compute.manager [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] [instance: d973b09f-7727-4055-b05e-123079df9ce4] Build of instance d973b09f-7727-4055-b05e-123079df9ce4 was re-scheduled: Binding failed for port 72daca54-f990-4f16-9362-cccaded8dba7, please check neutron logs for more information. {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 773.691898] env[62066]: DEBUG nova.compute.manager [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] [instance: d973b09f-7727-4055-b05e-123079df9ce4] Unplugging VIFs for instance {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 773.692325] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] Acquiring lock "refresh_cache-d973b09f-7727-4055-b05e-123079df9ce4" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 773.692748] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] Acquired lock "refresh_cache-d973b09f-7727-4055-b05e-123079df9ce4" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 773.694256] env[62066]: DEBUG nova.network.neutron [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] [instance: d973b09f-7727-4055-b05e-123079df9ce4] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 773.816618] env[62066]: DEBUG nova.compute.manager [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 0ebba3b2-f82d-4c1d-b01a-f75c3559c117] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 773.849047] env[62066]: DEBUG nova.virt.hardware [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-23T13:40:41Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-23T13:40:23Z,direct_url=,disk_format='vmdk',id=50ff584c-3b50-4395-af07-3e66769bc9f7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='eb52f7069a374c61ae946f052007c6d9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-23T13:40:24Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 773.849210] env[62066]: DEBUG nova.virt.hardware [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 773.849405] env[62066]: DEBUG nova.virt.hardware [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 773.849598] env[62066]: DEBUG nova.virt.hardware [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 773.849743] env[62066]: DEBUG nova.virt.hardware [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 773.849890] env[62066]: DEBUG nova.virt.hardware [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 773.850144] env[62066]: DEBUG nova.virt.hardware [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 773.850319] env[62066]: DEBUG nova.virt.hardware [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 773.850488] env[62066]: DEBUG nova.virt.hardware [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 773.850652] env[62066]: DEBUG nova.virt.hardware [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 773.850821] env[62066]: DEBUG nova.virt.hardware [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 773.851793] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-634b73d4-b792-4014-907d-e9cbd909e9db {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.861551] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03089012-fbd9-4c62-bcb7-3c9a8d52bf2e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.179383] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c70bcb7f-78ff-424d-8121-684c45ecd12b tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 774.217102] env[62066]: DEBUG nova.network.neutron [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] [instance: d973b09f-7727-4055-b05e-123079df9ce4] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 774.300293] env[62066]: DEBUG nova.network.neutron [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] [instance: d973b09f-7727-4055-b05e-123079df9ce4] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 774.442502] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-617c2762-7f7c-49e0-b210-698a00f832d0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.450056] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3ab72c1-a5ea-4bd0-bdef-6556ee6f6681 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.480039] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b592125-980e-4908-ba4a-17c304ce35b1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.487331] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cffd89c-383e-40a8-a5d6-46735b3f84a1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.500397] env[62066]: DEBUG nova.compute.provider_tree [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 774.771338] env[62066]: DEBUG nova.compute.manager [req-510704e9-271a-430c-a7ae-1457094ed88b req-a9a1a281-86f8-4dcc-aa81-6fcf913a0c3d service nova] [instance: 0ebba3b2-f82d-4c1d-b01a-f75c3559c117] Received event network-vif-plugged-d9ded989-5fd8-4b73-86d4-dd8d8247d504 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 774.771971] env[62066]: DEBUG oslo_concurrency.lockutils [req-510704e9-271a-430c-a7ae-1457094ed88b req-a9a1a281-86f8-4dcc-aa81-6fcf913a0c3d service nova] Acquiring lock "0ebba3b2-f82d-4c1d-b01a-f75c3559c117-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 774.772956] env[62066]: DEBUG oslo_concurrency.lockutils [req-510704e9-271a-430c-a7ae-1457094ed88b req-a9a1a281-86f8-4dcc-aa81-6fcf913a0c3d service nova] Lock "0ebba3b2-f82d-4c1d-b01a-f75c3559c117-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 774.774157] env[62066]: DEBUG oslo_concurrency.lockutils [req-510704e9-271a-430c-a7ae-1457094ed88b req-a9a1a281-86f8-4dcc-aa81-6fcf913a0c3d service nova] Lock "0ebba3b2-f82d-4c1d-b01a-f75c3559c117-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 774.774157] env[62066]: DEBUG nova.compute.manager [req-510704e9-271a-430c-a7ae-1457094ed88b req-a9a1a281-86f8-4dcc-aa81-6fcf913a0c3d service nova] [instance: 0ebba3b2-f82d-4c1d-b01a-f75c3559c117] No waiting events found dispatching network-vif-plugged-d9ded989-5fd8-4b73-86d4-dd8d8247d504 {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 774.774157] env[62066]: WARNING nova.compute.manager [req-510704e9-271a-430c-a7ae-1457094ed88b req-a9a1a281-86f8-4dcc-aa81-6fcf913a0c3d service nova] [instance: 0ebba3b2-f82d-4c1d-b01a-f75c3559c117] Received unexpected event network-vif-plugged-d9ded989-5fd8-4b73-86d4-dd8d8247d504 for instance with vm_state building and task_state spawning. [ 774.804776] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] Releasing lock "refresh_cache-d973b09f-7727-4055-b05e-123079df9ce4" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 774.804776] env[62066]: DEBUG nova.compute.manager [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 774.804776] env[62066]: DEBUG nova.compute.manager [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] [instance: d973b09f-7727-4055-b05e-123079df9ce4] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 774.805014] env[62066]: DEBUG nova.network.neutron [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] [instance: d973b09f-7727-4055-b05e-123079df9ce4] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 774.833186] env[62066]: DEBUG nova.network.neutron [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] [instance: d973b09f-7727-4055-b05e-123079df9ce4] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 774.922377] env[62066]: DEBUG nova.network.neutron [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 0ebba3b2-f82d-4c1d-b01a-f75c3559c117] Successfully updated port: d9ded989-5fd8-4b73-86d4-dd8d8247d504 {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 775.003392] env[62066]: DEBUG nova.scheduler.client.report [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 775.336266] env[62066]: DEBUG nova.network.neutron [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] [instance: d973b09f-7727-4055-b05e-123079df9ce4] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 775.424482] env[62066]: DEBUG oslo_concurrency.lockutils [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Acquiring lock "refresh_cache-0ebba3b2-f82d-4c1d-b01a-f75c3559c117" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 775.424616] env[62066]: DEBUG oslo_concurrency.lockutils [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Acquired lock "refresh_cache-0ebba3b2-f82d-4c1d-b01a-f75c3559c117" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 775.424752] env[62066]: DEBUG nova.network.neutron [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 0ebba3b2-f82d-4c1d-b01a-f75c3559c117] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 775.508802] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.823s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 775.509482] env[62066]: ERROR nova.compute.manager [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port a7a012fe-04f6-4d88-b008-27cf1ef05505, please check neutron logs for more information. [ 775.509482] env[62066]: ERROR nova.compute.manager [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] Traceback (most recent call last): [ 775.509482] env[62066]: ERROR nova.compute.manager [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 775.509482] env[62066]: ERROR nova.compute.manager [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] self.driver.spawn(context, instance, image_meta, [ 775.509482] env[62066]: ERROR nova.compute.manager [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 775.509482] env[62066]: ERROR nova.compute.manager [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 775.509482] env[62066]: ERROR nova.compute.manager [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 775.509482] env[62066]: ERROR nova.compute.manager [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] vm_ref = self.build_virtual_machine(instance, [ 775.509482] env[62066]: ERROR nova.compute.manager [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 775.509482] env[62066]: ERROR nova.compute.manager [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] vif_infos = vmwarevif.get_vif_info(self._session, [ 775.509482] env[62066]: ERROR nova.compute.manager [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 775.509825] env[62066]: ERROR nova.compute.manager [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] for vif in network_info: [ 775.509825] env[62066]: ERROR nova.compute.manager [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 775.509825] env[62066]: ERROR nova.compute.manager [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] return self._sync_wrapper(fn, *args, **kwargs) [ 775.509825] env[62066]: ERROR nova.compute.manager [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 775.509825] env[62066]: ERROR nova.compute.manager [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] self.wait() [ 775.509825] env[62066]: ERROR nova.compute.manager [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 775.509825] env[62066]: ERROR nova.compute.manager [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] self[:] = self._gt.wait() [ 775.509825] env[62066]: ERROR nova.compute.manager [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 775.509825] env[62066]: ERROR nova.compute.manager [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] return self._exit_event.wait() [ 775.509825] env[62066]: ERROR nova.compute.manager [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 775.509825] env[62066]: ERROR nova.compute.manager [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] current.throw(*self._exc) [ 775.509825] env[62066]: ERROR nova.compute.manager [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 775.509825] env[62066]: ERROR nova.compute.manager [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] result = function(*args, **kwargs) [ 775.510208] env[62066]: ERROR nova.compute.manager [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 775.510208] env[62066]: ERROR nova.compute.manager [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] return func(*args, **kwargs) [ 775.510208] env[62066]: ERROR nova.compute.manager [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 775.510208] env[62066]: ERROR nova.compute.manager [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] raise e [ 775.510208] env[62066]: ERROR nova.compute.manager [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 775.510208] env[62066]: ERROR nova.compute.manager [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] nwinfo = self.network_api.allocate_for_instance( [ 775.510208] env[62066]: ERROR nova.compute.manager [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 775.510208] env[62066]: ERROR nova.compute.manager [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] created_port_ids = self._update_ports_for_instance( [ 775.510208] env[62066]: ERROR nova.compute.manager [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 775.510208] env[62066]: ERROR nova.compute.manager [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] with excutils.save_and_reraise_exception(): [ 775.510208] env[62066]: ERROR nova.compute.manager [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 775.510208] env[62066]: ERROR nova.compute.manager [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] self.force_reraise() [ 775.510208] env[62066]: ERROR nova.compute.manager [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 775.510564] env[62066]: ERROR nova.compute.manager [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] raise self.value [ 775.510564] env[62066]: ERROR nova.compute.manager [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 775.510564] env[62066]: ERROR nova.compute.manager [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] updated_port = self._update_port( [ 775.510564] env[62066]: ERROR nova.compute.manager [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 775.510564] env[62066]: ERROR nova.compute.manager [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] _ensure_no_port_binding_failure(port) [ 775.510564] env[62066]: ERROR nova.compute.manager [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 775.510564] env[62066]: ERROR nova.compute.manager [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] raise exception.PortBindingFailed(port_id=port['id']) [ 775.510564] env[62066]: ERROR nova.compute.manager [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] nova.exception.PortBindingFailed: Binding failed for port a7a012fe-04f6-4d88-b008-27cf1ef05505, please check neutron logs for more information. [ 775.510564] env[62066]: ERROR nova.compute.manager [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] [ 775.510564] env[62066]: DEBUG nova.compute.utils [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] Binding failed for port a7a012fe-04f6-4d88-b008-27cf1ef05505, please check neutron logs for more information. {{(pid=62066) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 775.511709] env[62066]: DEBUG nova.compute.manager [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] Build of instance 71c44a07-fdf0-4847-868a-47f16c033c3d was re-scheduled: Binding failed for port a7a012fe-04f6-4d88-b008-27cf1ef05505, please check neutron logs for more information. {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 775.512125] env[62066]: DEBUG nova.compute.manager [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] Unplugging VIFs for instance {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 775.512352] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] Acquiring lock "refresh_cache-71c44a07-fdf0-4847-868a-47f16c033c3d" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 775.512496] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] Acquired lock "refresh_cache-71c44a07-fdf0-4847-868a-47f16c033c3d" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 775.512650] env[62066]: DEBUG nova.network.neutron [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 775.513638] env[62066]: DEBUG oslo_concurrency.lockutils [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.354s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 775.515367] env[62066]: INFO nova.compute.claims [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 98fb270c-f3f8-4375-8b5c-c2279305d476] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 775.839015] env[62066]: INFO nova.compute.manager [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] [instance: d973b09f-7727-4055-b05e-123079df9ce4] Took 1.03 seconds to deallocate network for instance. [ 775.965692] env[62066]: DEBUG nova.network.neutron [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 0ebba3b2-f82d-4c1d-b01a-f75c3559c117] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 776.034419] env[62066]: DEBUG nova.network.neutron [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 776.121469] env[62066]: DEBUG nova.network.neutron [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 776.134640] env[62066]: DEBUG nova.network.neutron [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 0ebba3b2-f82d-4c1d-b01a-f75c3559c117] Updating instance_info_cache with network_info: [{"id": "d9ded989-5fd8-4b73-86d4-dd8d8247d504", "address": "fa:16:3e:07:a4:fb", "network": {"id": "6bef6488-4bd5-4d76-921d-d2ab43051386", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-859765632-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7664e11610664ce5a85a2ad8758986f8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d733acc2-07d0-479e-918c-ec8a21925389", "external-id": "nsx-vlan-transportzone-459", "segmentation_id": 459, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd9ded989-5f", "ovs_interfaceid": "d9ded989-5fd8-4b73-86d4-dd8d8247d504", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 776.624484] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] Releasing lock "refresh_cache-71c44a07-fdf0-4847-868a-47f16c033c3d" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 776.624714] env[62066]: DEBUG nova.compute.manager [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 776.624894] env[62066]: DEBUG nova.compute.manager [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 776.625072] env[62066]: DEBUG nova.network.neutron [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 776.637270] env[62066]: DEBUG oslo_concurrency.lockutils [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Releasing lock "refresh_cache-0ebba3b2-f82d-4c1d-b01a-f75c3559c117" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 776.637639] env[62066]: DEBUG nova.compute.manager [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 0ebba3b2-f82d-4c1d-b01a-f75c3559c117] Instance network_info: |[{"id": "d9ded989-5fd8-4b73-86d4-dd8d8247d504", "address": "fa:16:3e:07:a4:fb", "network": {"id": "6bef6488-4bd5-4d76-921d-d2ab43051386", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-859765632-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7664e11610664ce5a85a2ad8758986f8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d733acc2-07d0-479e-918c-ec8a21925389", "external-id": "nsx-vlan-transportzone-459", "segmentation_id": 459, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd9ded989-5f", "ovs_interfaceid": "d9ded989-5fd8-4b73-86d4-dd8d8247d504", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 776.638179] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 0ebba3b2-f82d-4c1d-b01a-f75c3559c117] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:07:a4:fb', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd733acc2-07d0-479e-918c-ec8a21925389', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd9ded989-5fd8-4b73-86d4-dd8d8247d504', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 776.646095] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Creating folder: Project (7664e11610664ce5a85a2ad8758986f8). Parent ref: group-v285980. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 776.646984] env[62066]: DEBUG nova.network.neutron [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 776.651618] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-22f816d5-2fae-44b2-bc02-9c80928356c6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.662328] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Created folder: Project (7664e11610664ce5a85a2ad8758986f8) in parent group-v285980. [ 776.662514] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Creating folder: Instances. Parent ref: group-v286004. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 776.664860] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1962465f-b950-45cd-ba04-78e2553f0832 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.674669] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Created folder: Instances in parent group-v286004. [ 776.674910] env[62066]: DEBUG oslo.service.loopingcall [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 776.675118] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0ebba3b2-f82d-4c1d-b01a-f75c3559c117] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 776.675324] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e348999e-bb70-4439-a715-d4e2078fc527 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.695718] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 776.695718] env[62066]: value = "task-1340698" [ 776.695718] env[62066]: _type = "Task" [ 776.695718] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 776.704835] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1340698, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 776.791591] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c041ec0a-bea4-41d5-bd4c-bf8e14a924ee {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.800546] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24e53b4a-0cf9-4131-bdca-d570cc8bba16 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.832161] env[62066]: DEBUG nova.compute.manager [req-3202b730-8db6-4b80-bcb2-f53c89a5cef3 req-8c5327f3-9597-455c-97d4-e037c18b227a service nova] [instance: 0ebba3b2-f82d-4c1d-b01a-f75c3559c117] Received event network-changed-d9ded989-5fd8-4b73-86d4-dd8d8247d504 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 776.832364] env[62066]: DEBUG nova.compute.manager [req-3202b730-8db6-4b80-bcb2-f53c89a5cef3 req-8c5327f3-9597-455c-97d4-e037c18b227a service nova] [instance: 0ebba3b2-f82d-4c1d-b01a-f75c3559c117] Refreshing instance network info cache due to event network-changed-d9ded989-5fd8-4b73-86d4-dd8d8247d504. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 776.832573] env[62066]: DEBUG oslo_concurrency.lockutils [req-3202b730-8db6-4b80-bcb2-f53c89a5cef3 req-8c5327f3-9597-455c-97d4-e037c18b227a service nova] Acquiring lock "refresh_cache-0ebba3b2-f82d-4c1d-b01a-f75c3559c117" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 776.832714] env[62066]: DEBUG oslo_concurrency.lockutils [req-3202b730-8db6-4b80-bcb2-f53c89a5cef3 req-8c5327f3-9597-455c-97d4-e037c18b227a service nova] Acquired lock "refresh_cache-0ebba3b2-f82d-4c1d-b01a-f75c3559c117" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 776.832871] env[62066]: DEBUG nova.network.neutron [req-3202b730-8db6-4b80-bcb2-f53c89a5cef3 req-8c5327f3-9597-455c-97d4-e037c18b227a service nova] [instance: 0ebba3b2-f82d-4c1d-b01a-f75c3559c117] Refreshing network info cache for port d9ded989-5fd8-4b73-86d4-dd8d8247d504 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 776.834961] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac3c0ae5-766a-46f5-84fd-fce1abba2a55 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.845351] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d8c32f7-46f6-439d-abdd-a7c99f1ea617 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.863482] env[62066]: DEBUG nova.compute.provider_tree [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 776.869363] env[62066]: INFO nova.scheduler.client.report [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] Deleted allocations for instance d973b09f-7727-4055-b05e-123079df9ce4 [ 777.150023] env[62066]: DEBUG nova.network.neutron [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 777.205755] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1340698, 'name': CreateVM_Task, 'duration_secs': 0.304023} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 777.205755] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0ebba3b2-f82d-4c1d-b01a-f75c3559c117] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 777.212707] env[62066]: DEBUG oslo_concurrency.lockutils [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 777.212891] env[62066]: DEBUG oslo_concurrency.lockutils [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Acquired lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 777.213250] env[62066]: DEBUG oslo_concurrency.lockutils [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 777.213519] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3cb2d903-369f-4de3-957c-aacbfb2f8119 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.218223] env[62066]: DEBUG oslo_vmware.api [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Waiting for the task: (returnval){ [ 777.218223] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]5278c9a7-de89-aebd-2947-ad3a92eed435" [ 777.218223] env[62066]: _type = "Task" [ 777.218223] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 777.226091] env[62066]: DEBUG oslo_vmware.api [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]5278c9a7-de89-aebd-2947-ad3a92eed435, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.365650] env[62066]: DEBUG nova.scheduler.client.report [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 777.376792] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] Lock "d973b09f-7727-4055-b05e-123079df9ce4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 154.079s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 777.521548] env[62066]: DEBUG nova.network.neutron [req-3202b730-8db6-4b80-bcb2-f53c89a5cef3 req-8c5327f3-9597-455c-97d4-e037c18b227a service nova] [instance: 0ebba3b2-f82d-4c1d-b01a-f75c3559c117] Updated VIF entry in instance network info cache for port d9ded989-5fd8-4b73-86d4-dd8d8247d504. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 777.521933] env[62066]: DEBUG nova.network.neutron [req-3202b730-8db6-4b80-bcb2-f53c89a5cef3 req-8c5327f3-9597-455c-97d4-e037c18b227a service nova] [instance: 0ebba3b2-f82d-4c1d-b01a-f75c3559c117] Updating instance_info_cache with network_info: [{"id": "d9ded989-5fd8-4b73-86d4-dd8d8247d504", "address": "fa:16:3e:07:a4:fb", "network": {"id": "6bef6488-4bd5-4d76-921d-d2ab43051386", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-859765632-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7664e11610664ce5a85a2ad8758986f8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d733acc2-07d0-479e-918c-ec8a21925389", "external-id": "nsx-vlan-transportzone-459", "segmentation_id": 459, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd9ded989-5f", "ovs_interfaceid": "d9ded989-5fd8-4b73-86d4-dd8d8247d504", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 777.653230] env[62066]: INFO nova.compute.manager [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] [instance: 71c44a07-fdf0-4847-868a-47f16c033c3d] Took 1.03 seconds to deallocate network for instance. [ 777.728413] env[62066]: DEBUG oslo_vmware.api [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]5278c9a7-de89-aebd-2947-ad3a92eed435, 'name': SearchDatastore_Task, 'duration_secs': 0.009635} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 777.728715] env[62066]: DEBUG oslo_concurrency.lockutils [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Releasing lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 777.728992] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 0ebba3b2-f82d-4c1d-b01a-f75c3559c117] Processing image 50ff584c-3b50-4395-af07-3e66769bc9f7 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 777.729223] env[62066]: DEBUG oslo_concurrency.lockutils [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 777.729380] env[62066]: DEBUG oslo_concurrency.lockutils [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Acquired lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 777.729663] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 777.729857] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6c751e33-6ec5-465b-a001-5d335d989136 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.737643] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 777.737828] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 777.738522] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1f08c38c-8727-43d0-add6-9781332d197f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.743367] env[62066]: DEBUG oslo_vmware.api [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Waiting for the task: (returnval){ [ 777.743367] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]5292dd72-d309-e3e3-4563-5bfd1396ff6b" [ 777.743367] env[62066]: _type = "Task" [ 777.743367] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 777.751494] env[62066]: DEBUG oslo_vmware.api [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]5292dd72-d309-e3e3-4563-5bfd1396ff6b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.873643] env[62066]: DEBUG oslo_concurrency.lockutils [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.360s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 777.874507] env[62066]: DEBUG nova.compute.manager [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 98fb270c-f3f8-4375-8b5c-c2279305d476] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 777.880030] env[62066]: DEBUG oslo_concurrency.lockutils [None req-064a3fac-0a45-4f27-8eed-cb07d9679634 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.730s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 777.880030] env[62066]: INFO nova.compute.claims [None req-064a3fac-0a45-4f27-8eed-cb07d9679634 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] [instance: 1a4b0637-1a56-41ef-b89b-6b56d24ed206] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 777.882320] env[62066]: DEBUG nova.compute.manager [None req-816774ed-d63b-4dd8-94be-e4c70c8f10e7 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: 459f432b-e817-405f-9c3e-dfa5d957788c] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 778.024214] env[62066]: DEBUG oslo_concurrency.lockutils [req-3202b730-8db6-4b80-bcb2-f53c89a5cef3 req-8c5327f3-9597-455c-97d4-e037c18b227a service nova] Releasing lock "refresh_cache-0ebba3b2-f82d-4c1d-b01a-f75c3559c117" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 778.254941] env[62066]: DEBUG oslo_vmware.api [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]5292dd72-d309-e3e3-4563-5bfd1396ff6b, 'name': SearchDatastore_Task, 'duration_secs': 0.00831} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 778.255707] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d997a322-bafc-4d57-9f15-4818c23219da {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.261030] env[62066]: DEBUG oslo_vmware.api [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Waiting for the task: (returnval){ [ 778.261030] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]5259b8f9-7bf9-56c3-d7c9-caf386f433d0" [ 778.261030] env[62066]: _type = "Task" [ 778.261030] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 778.268660] env[62066]: DEBUG oslo_vmware.api [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]5259b8f9-7bf9-56c3-d7c9-caf386f433d0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 778.378830] env[62066]: DEBUG nova.compute.utils [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 778.380310] env[62066]: DEBUG nova.compute.manager [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 98fb270c-f3f8-4375-8b5c-c2279305d476] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 778.380484] env[62066]: DEBUG nova.network.neutron [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 98fb270c-f3f8-4375-8b5c-c2279305d476] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 778.399850] env[62066]: DEBUG oslo_concurrency.lockutils [None req-816774ed-d63b-4dd8-94be-e4c70c8f10e7 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 778.419554] env[62066]: DEBUG nova.policy [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '236eede7716d4664a75f5df76aa403ce', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7664e11610664ce5a85a2ad8758986f8', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 778.666592] env[62066]: DEBUG nova.network.neutron [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 98fb270c-f3f8-4375-8b5c-c2279305d476] Successfully created port: 78f327de-1b83-4e33-a2b7-8d1269a96df1 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 778.682848] env[62066]: INFO nova.scheduler.client.report [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] Deleted allocations for instance 71c44a07-fdf0-4847-868a-47f16c033c3d [ 778.778567] env[62066]: DEBUG oslo_vmware.api [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]5259b8f9-7bf9-56c3-d7c9-caf386f433d0, 'name': SearchDatastore_Task, 'duration_secs': 0.0084} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 778.779784] env[62066]: DEBUG oslo_concurrency.lockutils [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Releasing lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 778.782867] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk to [datastore2] 0ebba3b2-f82d-4c1d-b01a-f75c3559c117/0ebba3b2-f82d-4c1d-b01a-f75c3559c117.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 778.782867] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c05dcfcc-958c-409e-b624-7c79515ea4ba {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.789039] env[62066]: DEBUG oslo_vmware.api [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Waiting for the task: (returnval){ [ 778.789039] env[62066]: value = "task-1340699" [ 778.789039] env[62066]: _type = "Task" [ 778.789039] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 778.798420] env[62066]: DEBUG oslo_vmware.api [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Task: {'id': task-1340699, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 778.883845] env[62066]: DEBUG nova.compute.manager [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 98fb270c-f3f8-4375-8b5c-c2279305d476] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 779.157576] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49d85f0a-ca7d-4f18-9ee6-cd2f92599e26 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.166441] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-781baa8f-c0c3-4aee-8ca6-4ded02291e24 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.198580] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] Lock "71c44a07-fdf0-4847-868a-47f16c033c3d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 155.863s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 779.202937] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95d27580-7965-4893-9708-7e0e396b51b4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.209701] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2fbc29b-a57a-4e3a-a347-30a6c3f145b0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.225994] env[62066]: DEBUG nova.compute.provider_tree [None req-064a3fac-0a45-4f27-8eed-cb07d9679634 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 779.300747] env[62066]: DEBUG oslo_vmware.api [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Task: {'id': task-1340699, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.459716} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 779.301034] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk to [datastore2] 0ebba3b2-f82d-4c1d-b01a-f75c3559c117/0ebba3b2-f82d-4c1d-b01a-f75c3559c117.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 779.301260] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 0ebba3b2-f82d-4c1d-b01a-f75c3559c117] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 779.301514] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1c9dc6b3-b181-4bc6-ac35-8d6b56098341 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.307988] env[62066]: DEBUG oslo_vmware.api [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Waiting for the task: (returnval){ [ 779.307988] env[62066]: value = "task-1340700" [ 779.307988] env[62066]: _type = "Task" [ 779.307988] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 779.315829] env[62066]: DEBUG oslo_vmware.api [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Task: {'id': task-1340700, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.705579] env[62066]: DEBUG nova.compute.manager [None req-febd07a3-de81-4924-888e-f9c7d8ba2329 tempest-ServerMetadataNegativeTestJSON-1549470887 tempest-ServerMetadataNegativeTestJSON-1549470887-project-member] [instance: 31d47299-83a3-4f27-aeb7-95c4cd36c5ac] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 779.729436] env[62066]: DEBUG nova.scheduler.client.report [None req-064a3fac-0a45-4f27-8eed-cb07d9679634 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 779.817675] env[62066]: DEBUG oslo_vmware.api [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Task: {'id': task-1340700, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064449} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 779.817939] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 0ebba3b2-f82d-4c1d-b01a-f75c3559c117] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 779.818730] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-177c22b0-273b-4918-8073-c64c747c3eaa {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.841584] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 0ebba3b2-f82d-4c1d-b01a-f75c3559c117] Reconfiguring VM instance instance-00000038 to attach disk [datastore2] 0ebba3b2-f82d-4c1d-b01a-f75c3559c117/0ebba3b2-f82d-4c1d-b01a-f75c3559c117.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 779.841847] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-95e791f2-3b29-49f9-acc0-5f989ee3f1b0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.861218] env[62066]: DEBUG oslo_vmware.api [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Waiting for the task: (returnval){ [ 779.861218] env[62066]: value = "task-1340701" [ 779.861218] env[62066]: _type = "Task" [ 779.861218] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 779.869154] env[62066]: DEBUG oslo_vmware.api [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Task: {'id': task-1340701, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.897015] env[62066]: DEBUG nova.compute.manager [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 98fb270c-f3f8-4375-8b5c-c2279305d476] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 779.922686] env[62066]: DEBUG nova.virt.hardware [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-23T13:40:41Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-23T13:40:23Z,direct_url=,disk_format='vmdk',id=50ff584c-3b50-4395-af07-3e66769bc9f7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='eb52f7069a374c61ae946f052007c6d9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-23T13:40:24Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 779.922965] env[62066]: DEBUG nova.virt.hardware [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 779.923158] env[62066]: DEBUG nova.virt.hardware [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 779.923355] env[62066]: DEBUG nova.virt.hardware [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 779.923516] env[62066]: DEBUG nova.virt.hardware [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 779.923678] env[62066]: DEBUG nova.virt.hardware [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 779.923890] env[62066]: DEBUG nova.virt.hardware [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 779.924077] env[62066]: DEBUG nova.virt.hardware [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 779.924265] env[62066]: DEBUG nova.virt.hardware [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 779.924431] env[62066]: DEBUG nova.virt.hardware [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 779.924677] env[62066]: DEBUG nova.virt.hardware [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 779.925511] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf57b66e-9687-42c4-89ab-6c7a92eabce9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.933675] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69ada526-c50e-41b9-a9e7-8a44bda33a9b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.216954] env[62066]: DEBUG nova.compute.manager [req-f5a1fd1a-0261-461d-8d9a-6dc51db08d07 req-cc7c7d7d-8934-462b-8f39-74e2cb02ddde service nova] [instance: 98fb270c-f3f8-4375-8b5c-c2279305d476] Received event network-vif-plugged-78f327de-1b83-4e33-a2b7-8d1269a96df1 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 780.217292] env[62066]: DEBUG oslo_concurrency.lockutils [req-f5a1fd1a-0261-461d-8d9a-6dc51db08d07 req-cc7c7d7d-8934-462b-8f39-74e2cb02ddde service nova] Acquiring lock "98fb270c-f3f8-4375-8b5c-c2279305d476-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 780.218068] env[62066]: DEBUG oslo_concurrency.lockutils [req-f5a1fd1a-0261-461d-8d9a-6dc51db08d07 req-cc7c7d7d-8934-462b-8f39-74e2cb02ddde service nova] Lock "98fb270c-f3f8-4375-8b5c-c2279305d476-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 780.218068] env[62066]: DEBUG oslo_concurrency.lockutils [req-f5a1fd1a-0261-461d-8d9a-6dc51db08d07 req-cc7c7d7d-8934-462b-8f39-74e2cb02ddde service nova] Lock "98fb270c-f3f8-4375-8b5c-c2279305d476-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 780.218068] env[62066]: DEBUG nova.compute.manager [req-f5a1fd1a-0261-461d-8d9a-6dc51db08d07 req-cc7c7d7d-8934-462b-8f39-74e2cb02ddde service nova] [instance: 98fb270c-f3f8-4375-8b5c-c2279305d476] No waiting events found dispatching network-vif-plugged-78f327de-1b83-4e33-a2b7-8d1269a96df1 {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 780.218068] env[62066]: WARNING nova.compute.manager [req-f5a1fd1a-0261-461d-8d9a-6dc51db08d07 req-cc7c7d7d-8934-462b-8f39-74e2cb02ddde service nova] [instance: 98fb270c-f3f8-4375-8b5c-c2279305d476] Received unexpected event network-vif-plugged-78f327de-1b83-4e33-a2b7-8d1269a96df1 for instance with vm_state building and task_state spawning. [ 780.227395] env[62066]: DEBUG oslo_concurrency.lockutils [None req-febd07a3-de81-4924-888e-f9c7d8ba2329 tempest-ServerMetadataNegativeTestJSON-1549470887 tempest-ServerMetadataNegativeTestJSON-1549470887-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 780.234255] env[62066]: DEBUG oslo_concurrency.lockutils [None req-064a3fac-0a45-4f27-8eed-cb07d9679634 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.357s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 780.234734] env[62066]: DEBUG nova.compute.manager [None req-064a3fac-0a45-4f27-8eed-cb07d9679634 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] [instance: 1a4b0637-1a56-41ef-b89b-6b56d24ed206] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 780.237446] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 26.268s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 780.307015] env[62066]: DEBUG nova.network.neutron [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 98fb270c-f3f8-4375-8b5c-c2279305d476] Successfully updated port: 78f327de-1b83-4e33-a2b7-8d1269a96df1 {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 780.371275] env[62066]: DEBUG oslo_vmware.api [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Task: {'id': task-1340701, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.742073] env[62066]: DEBUG nova.compute.utils [None req-064a3fac-0a45-4f27-8eed-cb07d9679634 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 780.745806] env[62066]: DEBUG nova.compute.manager [None req-064a3fac-0a45-4f27-8eed-cb07d9679634 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] [instance: 1a4b0637-1a56-41ef-b89b-6b56d24ed206] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 780.745969] env[62066]: DEBUG nova.network.neutron [None req-064a3fac-0a45-4f27-8eed-cb07d9679634 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] [instance: 1a4b0637-1a56-41ef-b89b-6b56d24ed206] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 780.804704] env[62066]: DEBUG nova.policy [None req-064a3fac-0a45-4f27-8eed-cb07d9679634 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '34ad59e83ec04c8b82bf1fbca2988d48', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '638c4c550925459ea6fc75090dbacde0', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 780.812939] env[62066]: DEBUG oslo_concurrency.lockutils [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Acquiring lock "refresh_cache-98fb270c-f3f8-4375-8b5c-c2279305d476" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 780.813102] env[62066]: DEBUG oslo_concurrency.lockutils [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Acquired lock "refresh_cache-98fb270c-f3f8-4375-8b5c-c2279305d476" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 780.813256] env[62066]: DEBUG nova.network.neutron [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 98fb270c-f3f8-4375-8b5c-c2279305d476] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 780.872685] env[62066]: DEBUG oslo_vmware.api [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Task: {'id': task-1340701, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.980021] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-515e1870-1346-4d4b-9cc3-65054e6db759 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.989247] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b853a1f8-b13b-42f2-8186-f0af8aea9a61 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.035762] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c661e9e-222f-4a26-a198-f5134e072d55 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.044563] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-223ec8d2-640f-413e-8f8b-aebf6419c506 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.060651] env[62066]: DEBUG nova.compute.provider_tree [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 781.129323] env[62066]: DEBUG nova.network.neutron [None req-064a3fac-0a45-4f27-8eed-cb07d9679634 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] [instance: 1a4b0637-1a56-41ef-b89b-6b56d24ed206] Successfully created port: 84936b80-0721-4235-a92a-7bc4adf450d8 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 781.248766] env[62066]: DEBUG nova.compute.manager [None req-064a3fac-0a45-4f27-8eed-cb07d9679634 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] [instance: 1a4b0637-1a56-41ef-b89b-6b56d24ed206] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 781.347577] env[62066]: DEBUG nova.network.neutron [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 98fb270c-f3f8-4375-8b5c-c2279305d476] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 781.371885] env[62066]: DEBUG oslo_vmware.api [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Task: {'id': task-1340701, 'name': ReconfigVM_Task, 'duration_secs': 1.317734} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.372293] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 0ebba3b2-f82d-4c1d-b01a-f75c3559c117] Reconfigured VM instance instance-00000038 to attach disk [datastore2] 0ebba3b2-f82d-4c1d-b01a-f75c3559c117/0ebba3b2-f82d-4c1d-b01a-f75c3559c117.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 781.372904] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c8aa57c1-0b59-4dd5-85fa-4ac8e97d80ac {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.378821] env[62066]: DEBUG oslo_vmware.api [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Waiting for the task: (returnval){ [ 781.378821] env[62066]: value = "task-1340702" [ 781.378821] env[62066]: _type = "Task" [ 781.378821] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.386164] env[62066]: DEBUG oslo_vmware.api [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Task: {'id': task-1340702, 'name': Rename_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.564801] env[62066]: DEBUG nova.scheduler.client.report [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 781.600229] env[62066]: DEBUG nova.network.neutron [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 98fb270c-f3f8-4375-8b5c-c2279305d476] Updating instance_info_cache with network_info: [{"id": "78f327de-1b83-4e33-a2b7-8d1269a96df1", "address": "fa:16:3e:9a:d5:44", "network": {"id": "6bef6488-4bd5-4d76-921d-d2ab43051386", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-859765632-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7664e11610664ce5a85a2ad8758986f8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d733acc2-07d0-479e-918c-ec8a21925389", "external-id": "nsx-vlan-transportzone-459", "segmentation_id": 459, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap78f327de-1b", "ovs_interfaceid": "78f327de-1b83-4e33-a2b7-8d1269a96df1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 781.888549] env[62066]: DEBUG oslo_vmware.api [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Task: {'id': task-1340702, 'name': Rename_Task, 'duration_secs': 0.135461} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.888769] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 0ebba3b2-f82d-4c1d-b01a-f75c3559c117] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 781.889017] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d98cfad4-88b5-4403-9b4a-8409edc00d75 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.894860] env[62066]: DEBUG oslo_vmware.api [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Waiting for the task: (returnval){ [ 781.894860] env[62066]: value = "task-1340703" [ 781.894860] env[62066]: _type = "Task" [ 781.894860] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.901807] env[62066]: DEBUG oslo_vmware.api [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Task: {'id': task-1340703, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.070121] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.833s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 782.070903] env[62066]: ERROR nova.compute.manager [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 85acf49f-2497-490a-8686-8131ebcad644, please check neutron logs for more information. [ 782.070903] env[62066]: ERROR nova.compute.manager [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] Traceback (most recent call last): [ 782.070903] env[62066]: ERROR nova.compute.manager [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 782.070903] env[62066]: ERROR nova.compute.manager [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] self.driver.spawn(context, instance, image_meta, [ 782.070903] env[62066]: ERROR nova.compute.manager [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 782.070903] env[62066]: ERROR nova.compute.manager [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] self._vmops.spawn(context, instance, image_meta, injected_files, [ 782.070903] env[62066]: ERROR nova.compute.manager [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 782.070903] env[62066]: ERROR nova.compute.manager [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] vm_ref = self.build_virtual_machine(instance, [ 782.070903] env[62066]: ERROR nova.compute.manager [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 782.070903] env[62066]: ERROR nova.compute.manager [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] vif_infos = vmwarevif.get_vif_info(self._session, [ 782.070903] env[62066]: ERROR nova.compute.manager [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 782.071467] env[62066]: ERROR nova.compute.manager [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] for vif in network_info: [ 782.071467] env[62066]: ERROR nova.compute.manager [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 782.071467] env[62066]: ERROR nova.compute.manager [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] return self._sync_wrapper(fn, *args, **kwargs) [ 782.071467] env[62066]: ERROR nova.compute.manager [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 782.071467] env[62066]: ERROR nova.compute.manager [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] self.wait() [ 782.071467] env[62066]: ERROR nova.compute.manager [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 782.071467] env[62066]: ERROR nova.compute.manager [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] self[:] = self._gt.wait() [ 782.071467] env[62066]: ERROR nova.compute.manager [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 782.071467] env[62066]: ERROR nova.compute.manager [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] return self._exit_event.wait() [ 782.071467] env[62066]: ERROR nova.compute.manager [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 782.071467] env[62066]: ERROR nova.compute.manager [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] current.throw(*self._exc) [ 782.071467] env[62066]: ERROR nova.compute.manager [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 782.071467] env[62066]: ERROR nova.compute.manager [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] result = function(*args, **kwargs) [ 782.072052] env[62066]: ERROR nova.compute.manager [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 782.072052] env[62066]: ERROR nova.compute.manager [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] return func(*args, **kwargs) [ 782.072052] env[62066]: ERROR nova.compute.manager [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 782.072052] env[62066]: ERROR nova.compute.manager [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] raise e [ 782.072052] env[62066]: ERROR nova.compute.manager [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 782.072052] env[62066]: ERROR nova.compute.manager [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] nwinfo = self.network_api.allocate_for_instance( [ 782.072052] env[62066]: ERROR nova.compute.manager [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 782.072052] env[62066]: ERROR nova.compute.manager [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] created_port_ids = self._update_ports_for_instance( [ 782.072052] env[62066]: ERROR nova.compute.manager [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 782.072052] env[62066]: ERROR nova.compute.manager [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] with excutils.save_and_reraise_exception(): [ 782.072052] env[62066]: ERROR nova.compute.manager [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 782.072052] env[62066]: ERROR nova.compute.manager [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] self.force_reraise() [ 782.072052] env[62066]: ERROR nova.compute.manager [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 782.072882] env[62066]: ERROR nova.compute.manager [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] raise self.value [ 782.072882] env[62066]: ERROR nova.compute.manager [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 782.072882] env[62066]: ERROR nova.compute.manager [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] updated_port = self._update_port( [ 782.072882] env[62066]: ERROR nova.compute.manager [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 782.072882] env[62066]: ERROR nova.compute.manager [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] _ensure_no_port_binding_failure(port) [ 782.072882] env[62066]: ERROR nova.compute.manager [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 782.072882] env[62066]: ERROR nova.compute.manager [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] raise exception.PortBindingFailed(port_id=port['id']) [ 782.072882] env[62066]: ERROR nova.compute.manager [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] nova.exception.PortBindingFailed: Binding failed for port 85acf49f-2497-490a-8686-8131ebcad644, please check neutron logs for more information. [ 782.072882] env[62066]: ERROR nova.compute.manager [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] [ 782.072882] env[62066]: DEBUG nova.compute.utils [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] Binding failed for port 85acf49f-2497-490a-8686-8131ebcad644, please check neutron logs for more information. {{(pid=62066) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 782.073585] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3baf3239-22c0-4858-82f6-d3b75bd59a7c tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.812s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 782.076024] env[62066]: INFO nova.compute.claims [None req-3baf3239-22c0-4858-82f6-d3b75bd59a7c tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] [instance: 1e1a2ede-4ad8-4600-851e-6f2046b3f919] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 782.079500] env[62066]: DEBUG nova.compute.manager [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] Build of instance f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5 was re-scheduled: Binding failed for port 85acf49f-2497-490a-8686-8131ebcad644, please check neutron logs for more information. {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 782.080182] env[62066]: DEBUG nova.compute.manager [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] Unplugging VIFs for instance {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 782.080527] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] Acquiring lock "refresh_cache-f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 782.080783] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] Acquired lock "refresh_cache-f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 782.081073] env[62066]: DEBUG nova.network.neutron [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 782.102313] env[62066]: DEBUG oslo_concurrency.lockutils [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Releasing lock "refresh_cache-98fb270c-f3f8-4375-8b5c-c2279305d476" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 782.102623] env[62066]: DEBUG nova.compute.manager [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 98fb270c-f3f8-4375-8b5c-c2279305d476] Instance network_info: |[{"id": "78f327de-1b83-4e33-a2b7-8d1269a96df1", "address": "fa:16:3e:9a:d5:44", "network": {"id": "6bef6488-4bd5-4d76-921d-d2ab43051386", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-859765632-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7664e11610664ce5a85a2ad8758986f8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d733acc2-07d0-479e-918c-ec8a21925389", "external-id": "nsx-vlan-transportzone-459", "segmentation_id": 459, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap78f327de-1b", "ovs_interfaceid": "78f327de-1b83-4e33-a2b7-8d1269a96df1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 782.103240] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 98fb270c-f3f8-4375-8b5c-c2279305d476] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:9a:d5:44', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd733acc2-07d0-479e-918c-ec8a21925389', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '78f327de-1b83-4e33-a2b7-8d1269a96df1', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 782.111564] env[62066]: DEBUG oslo.service.loopingcall [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 782.112684] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 98fb270c-f3f8-4375-8b5c-c2279305d476] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 782.112927] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-55bed9cd-01c0-44af-b88f-74873a501d87 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.133304] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 782.133304] env[62066]: value = "task-1340704" [ 782.133304] env[62066]: _type = "Task" [ 782.133304] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.141149] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1340704, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.260069] env[62066]: DEBUG nova.compute.manager [None req-064a3fac-0a45-4f27-8eed-cb07d9679634 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] [instance: 1a4b0637-1a56-41ef-b89b-6b56d24ed206] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 782.272295] env[62066]: DEBUG nova.compute.manager [req-aac7992b-b800-4597-a628-ef1d6bfc2dd5 req-8d0b34f5-c6b8-4cf2-b45d-1ce08d04a673 service nova] [instance: 98fb270c-f3f8-4375-8b5c-c2279305d476] Received event network-changed-78f327de-1b83-4e33-a2b7-8d1269a96df1 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 782.272492] env[62066]: DEBUG nova.compute.manager [req-aac7992b-b800-4597-a628-ef1d6bfc2dd5 req-8d0b34f5-c6b8-4cf2-b45d-1ce08d04a673 service nova] [instance: 98fb270c-f3f8-4375-8b5c-c2279305d476] Refreshing instance network info cache due to event network-changed-78f327de-1b83-4e33-a2b7-8d1269a96df1. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 782.273365] env[62066]: DEBUG oslo_concurrency.lockutils [req-aac7992b-b800-4597-a628-ef1d6bfc2dd5 req-8d0b34f5-c6b8-4cf2-b45d-1ce08d04a673 service nova] Acquiring lock "refresh_cache-98fb270c-f3f8-4375-8b5c-c2279305d476" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 782.273365] env[62066]: DEBUG oslo_concurrency.lockutils [req-aac7992b-b800-4597-a628-ef1d6bfc2dd5 req-8d0b34f5-c6b8-4cf2-b45d-1ce08d04a673 service nova] Acquired lock "refresh_cache-98fb270c-f3f8-4375-8b5c-c2279305d476" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 782.273365] env[62066]: DEBUG nova.network.neutron [req-aac7992b-b800-4597-a628-ef1d6bfc2dd5 req-8d0b34f5-c6b8-4cf2-b45d-1ce08d04a673 service nova] [instance: 98fb270c-f3f8-4375-8b5c-c2279305d476] Refreshing network info cache for port 78f327de-1b83-4e33-a2b7-8d1269a96df1 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 782.285565] env[62066]: DEBUG nova.virt.hardware [None req-064a3fac-0a45-4f27-8eed-cb07d9679634 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-23T13:40:41Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-23T13:40:23Z,direct_url=,disk_format='vmdk',id=50ff584c-3b50-4395-af07-3e66769bc9f7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='eb52f7069a374c61ae946f052007c6d9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-23T13:40:24Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 782.285813] env[62066]: DEBUG nova.virt.hardware [None req-064a3fac-0a45-4f27-8eed-cb07d9679634 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 782.287040] env[62066]: DEBUG nova.virt.hardware [None req-064a3fac-0a45-4f27-8eed-cb07d9679634 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 782.287040] env[62066]: DEBUG nova.virt.hardware [None req-064a3fac-0a45-4f27-8eed-cb07d9679634 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 782.287040] env[62066]: DEBUG nova.virt.hardware [None req-064a3fac-0a45-4f27-8eed-cb07d9679634 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 782.287040] env[62066]: DEBUG nova.virt.hardware [None req-064a3fac-0a45-4f27-8eed-cb07d9679634 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 782.287040] env[62066]: DEBUG nova.virt.hardware [None req-064a3fac-0a45-4f27-8eed-cb07d9679634 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 782.287285] env[62066]: DEBUG nova.virt.hardware [None req-064a3fac-0a45-4f27-8eed-cb07d9679634 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 782.287285] env[62066]: DEBUG nova.virt.hardware [None req-064a3fac-0a45-4f27-8eed-cb07d9679634 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 782.287285] env[62066]: DEBUG nova.virt.hardware [None req-064a3fac-0a45-4f27-8eed-cb07d9679634 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 782.287371] env[62066]: DEBUG nova.virt.hardware [None req-064a3fac-0a45-4f27-8eed-cb07d9679634 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 782.288586] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b863bf5-9ef8-4922-b394-3d44a00987ce {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.298358] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3f7603a-8040-47b1-a297-3151ad27cf7b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.407456] env[62066]: DEBUG oslo_vmware.api [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Task: {'id': task-1340703, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.602890] env[62066]: DEBUG nova.network.neutron [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 782.631081] env[62066]: DEBUG nova.network.neutron [None req-064a3fac-0a45-4f27-8eed-cb07d9679634 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] [instance: 1a4b0637-1a56-41ef-b89b-6b56d24ed206] Successfully updated port: 84936b80-0721-4235-a92a-7bc4adf450d8 {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 782.646940] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1340704, 'name': CreateVM_Task, 'duration_secs': 0.382206} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.648862] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 98fb270c-f3f8-4375-8b5c-c2279305d476] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 782.650644] env[62066]: DEBUG oslo_concurrency.lockutils [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 782.651666] env[62066]: DEBUG oslo_concurrency.lockutils [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Acquired lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 782.652051] env[62066]: DEBUG oslo_concurrency.lockutils [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 782.652585] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0cffa82a-0d0a-4cd6-b746-bf8f8d56b8c3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.661158] env[62066]: DEBUG oslo_vmware.api [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Waiting for the task: (returnval){ [ 782.661158] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]5247cd45-bade-e2b5-cce9-eb1a8f8ebbe2" [ 782.661158] env[62066]: _type = "Task" [ 782.661158] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.669264] env[62066]: DEBUG oslo_vmware.api [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]5247cd45-bade-e2b5-cce9-eb1a8f8ebbe2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.728897] env[62066]: DEBUG nova.network.neutron [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 782.904505] env[62066]: DEBUG oslo_vmware.api [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Task: {'id': task-1340703, 'name': PowerOnVM_Task, 'duration_secs': 0.639939} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.904834] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 0ebba3b2-f82d-4c1d-b01a-f75c3559c117] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 782.904957] env[62066]: INFO nova.compute.manager [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 0ebba3b2-f82d-4c1d-b01a-f75c3559c117] Took 9.09 seconds to spawn the instance on the hypervisor. [ 782.905151] env[62066]: DEBUG nova.compute.manager [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 0ebba3b2-f82d-4c1d-b01a-f75c3559c117] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 782.905898] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf6b1bb2-c984-4582-9c37-b5b89fd058d3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.955914] env[62066]: DEBUG nova.network.neutron [req-aac7992b-b800-4597-a628-ef1d6bfc2dd5 req-8d0b34f5-c6b8-4cf2-b45d-1ce08d04a673 service nova] [instance: 98fb270c-f3f8-4375-8b5c-c2279305d476] Updated VIF entry in instance network info cache for port 78f327de-1b83-4e33-a2b7-8d1269a96df1. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 782.956286] env[62066]: DEBUG nova.network.neutron [req-aac7992b-b800-4597-a628-ef1d6bfc2dd5 req-8d0b34f5-c6b8-4cf2-b45d-1ce08d04a673 service nova] [instance: 98fb270c-f3f8-4375-8b5c-c2279305d476] Updating instance_info_cache with network_info: [{"id": "78f327de-1b83-4e33-a2b7-8d1269a96df1", "address": "fa:16:3e:9a:d5:44", "network": {"id": "6bef6488-4bd5-4d76-921d-d2ab43051386", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-859765632-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7664e11610664ce5a85a2ad8758986f8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d733acc2-07d0-479e-918c-ec8a21925389", "external-id": "nsx-vlan-transportzone-459", "segmentation_id": 459, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap78f327de-1b", "ovs_interfaceid": "78f327de-1b83-4e33-a2b7-8d1269a96df1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 783.133901] env[62066]: DEBUG oslo_concurrency.lockutils [None req-064a3fac-0a45-4f27-8eed-cb07d9679634 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Acquiring lock "refresh_cache-1a4b0637-1a56-41ef-b89b-6b56d24ed206" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 783.134154] env[62066]: DEBUG oslo_concurrency.lockutils [None req-064a3fac-0a45-4f27-8eed-cb07d9679634 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Acquired lock "refresh_cache-1a4b0637-1a56-41ef-b89b-6b56d24ed206" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 783.134229] env[62066]: DEBUG nova.network.neutron [None req-064a3fac-0a45-4f27-8eed-cb07d9679634 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] [instance: 1a4b0637-1a56-41ef-b89b-6b56d24ed206] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 783.171931] env[62066]: DEBUG oslo_vmware.api [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]5247cd45-bade-e2b5-cce9-eb1a8f8ebbe2, 'name': SearchDatastore_Task, 'duration_secs': 0.011617} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 783.172249] env[62066]: DEBUG oslo_concurrency.lockutils [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Releasing lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 783.172481] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 98fb270c-f3f8-4375-8b5c-c2279305d476] Processing image 50ff584c-3b50-4395-af07-3e66769bc9f7 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 783.172716] env[62066]: DEBUG oslo_concurrency.lockutils [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 783.172859] env[62066]: DEBUG oslo_concurrency.lockutils [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Acquired lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 783.173046] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 783.173306] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6a90950c-92d0-4ab0-ae44-a837a8999d36 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.181197] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 783.181382] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 783.182099] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3a64a091-7c41-4500-9b0e-2e8e20417130 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.186984] env[62066]: DEBUG oslo_vmware.api [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Waiting for the task: (returnval){ [ 783.186984] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]5265348e-8428-365a-d85d-24ca484ffab8" [ 783.186984] env[62066]: _type = "Task" [ 783.186984] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 783.196131] env[62066]: DEBUG oslo_vmware.api [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]5265348e-8428-365a-d85d-24ca484ffab8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.231720] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] Releasing lock "refresh_cache-f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 783.233016] env[62066]: DEBUG nova.compute.manager [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 783.233234] env[62066]: DEBUG nova.compute.manager [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 783.234135] env[62066]: DEBUG nova.network.neutron [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 783.270631] env[62066]: DEBUG nova.network.neutron [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 783.314426] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ec4f0c6-ae71-4ea0-bee3-d986c8e15e3a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.322317] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcaf0630-17b4-49e7-af3a-d4a4c9f7e3c0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.366250] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b192be4-0000-4ac8-a64e-bb6991934454 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.374043] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ca65832-9369-42ca-ae69-dccbd4d393bd {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.386060] env[62066]: DEBUG nova.compute.provider_tree [None req-3baf3239-22c0-4858-82f6-d3b75bd59a7c tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 783.422608] env[62066]: INFO nova.compute.manager [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 0ebba3b2-f82d-4c1d-b01a-f75c3559c117] Took 42.19 seconds to build instance. [ 783.458576] env[62066]: DEBUG oslo_concurrency.lockutils [req-aac7992b-b800-4597-a628-ef1d6bfc2dd5 req-8d0b34f5-c6b8-4cf2-b45d-1ce08d04a673 service nova] Releasing lock "refresh_cache-98fb270c-f3f8-4375-8b5c-c2279305d476" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 783.662669] env[62066]: DEBUG nova.network.neutron [None req-064a3fac-0a45-4f27-8eed-cb07d9679634 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] [instance: 1a4b0637-1a56-41ef-b89b-6b56d24ed206] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 783.697087] env[62066]: DEBUG oslo_vmware.api [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]5265348e-8428-365a-d85d-24ca484ffab8, 'name': SearchDatastore_Task, 'duration_secs': 0.008077} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 783.697857] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7012151e-522c-4a1d-a21d-2d2b86698ef6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.704840] env[62066]: DEBUG oslo_vmware.api [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Waiting for the task: (returnval){ [ 783.704840] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]525b5a50-612f-c937-8f31-2d645c5cf372" [ 783.704840] env[62066]: _type = "Task" [ 783.704840] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 783.712099] env[62066]: DEBUG oslo_vmware.api [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]525b5a50-612f-c937-8f31-2d645c5cf372, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.775951] env[62066]: DEBUG nova.network.neutron [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 783.831382] env[62066]: DEBUG nova.network.neutron [None req-064a3fac-0a45-4f27-8eed-cb07d9679634 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] [instance: 1a4b0637-1a56-41ef-b89b-6b56d24ed206] Updating instance_info_cache with network_info: [{"id": "84936b80-0721-4235-a92a-7bc4adf450d8", "address": "fa:16:3e:f7:5a:62", "network": {"id": "518e5084-b25d-4931-9319-a62810398063", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.233", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "eb52f7069a374c61ae946f052007c6d9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c6365036-aa37-44d2-90d1-ca1c3516ded9", "external-id": "nsx-vlan-transportzone-66", "segmentation_id": 66, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap84936b80-07", "ovs_interfaceid": "84936b80-0721-4235-a92a-7bc4adf450d8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 783.891465] env[62066]: DEBUG nova.scheduler.client.report [None req-3baf3239-22c0-4858-82f6-d3b75bd59a7c tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 783.924340] env[62066]: DEBUG oslo_concurrency.lockutils [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Lock "0ebba3b2-f82d-4c1d-b01a-f75c3559c117" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 152.648s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 784.214096] env[62066]: DEBUG oslo_vmware.api [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]525b5a50-612f-c937-8f31-2d645c5cf372, 'name': SearchDatastore_Task, 'duration_secs': 0.009116} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 784.214423] env[62066]: DEBUG oslo_concurrency.lockutils [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Releasing lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 784.214597] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk to [datastore2] 98fb270c-f3f8-4375-8b5c-c2279305d476/98fb270c-f3f8-4375-8b5c-c2279305d476.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 784.214845] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a1fd7064-ca24-4f21-97e0-437b202abb34 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.220590] env[62066]: DEBUG oslo_vmware.api [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Waiting for the task: (returnval){ [ 784.220590] env[62066]: value = "task-1340705" [ 784.220590] env[62066]: _type = "Task" [ 784.220590] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 784.228233] env[62066]: DEBUG oslo_vmware.api [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Task: {'id': task-1340705, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.279246] env[62066]: INFO nova.compute.manager [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] [instance: f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5] Took 1.05 seconds to deallocate network for instance. [ 784.300122] env[62066]: DEBUG nova.compute.manager [req-22fe4902-9412-456b-a13f-99994b349398 req-bbbc2a06-656c-4a5c-8355-6771d3933a01 service nova] [instance: 1a4b0637-1a56-41ef-b89b-6b56d24ed206] Received event network-vif-plugged-84936b80-0721-4235-a92a-7bc4adf450d8 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 784.300350] env[62066]: DEBUG oslo_concurrency.lockutils [req-22fe4902-9412-456b-a13f-99994b349398 req-bbbc2a06-656c-4a5c-8355-6771d3933a01 service nova] Acquiring lock "1a4b0637-1a56-41ef-b89b-6b56d24ed206-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 784.300855] env[62066]: DEBUG oslo_concurrency.lockutils [req-22fe4902-9412-456b-a13f-99994b349398 req-bbbc2a06-656c-4a5c-8355-6771d3933a01 service nova] Lock "1a4b0637-1a56-41ef-b89b-6b56d24ed206-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 784.300855] env[62066]: DEBUG oslo_concurrency.lockutils [req-22fe4902-9412-456b-a13f-99994b349398 req-bbbc2a06-656c-4a5c-8355-6771d3933a01 service nova] Lock "1a4b0637-1a56-41ef-b89b-6b56d24ed206-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 784.300855] env[62066]: DEBUG nova.compute.manager [req-22fe4902-9412-456b-a13f-99994b349398 req-bbbc2a06-656c-4a5c-8355-6771d3933a01 service nova] [instance: 1a4b0637-1a56-41ef-b89b-6b56d24ed206] No waiting events found dispatching network-vif-plugged-84936b80-0721-4235-a92a-7bc4adf450d8 {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 784.300976] env[62066]: WARNING nova.compute.manager [req-22fe4902-9412-456b-a13f-99994b349398 req-bbbc2a06-656c-4a5c-8355-6771d3933a01 service nova] [instance: 1a4b0637-1a56-41ef-b89b-6b56d24ed206] Received unexpected event network-vif-plugged-84936b80-0721-4235-a92a-7bc4adf450d8 for instance with vm_state building and task_state spawning. [ 784.301134] env[62066]: DEBUG nova.compute.manager [req-22fe4902-9412-456b-a13f-99994b349398 req-bbbc2a06-656c-4a5c-8355-6771d3933a01 service nova] [instance: 1a4b0637-1a56-41ef-b89b-6b56d24ed206] Received event network-changed-84936b80-0721-4235-a92a-7bc4adf450d8 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 784.301507] env[62066]: DEBUG nova.compute.manager [req-22fe4902-9412-456b-a13f-99994b349398 req-bbbc2a06-656c-4a5c-8355-6771d3933a01 service nova] [instance: 1a4b0637-1a56-41ef-b89b-6b56d24ed206] Refreshing instance network info cache due to event network-changed-84936b80-0721-4235-a92a-7bc4adf450d8. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 784.301703] env[62066]: DEBUG oslo_concurrency.lockutils [req-22fe4902-9412-456b-a13f-99994b349398 req-bbbc2a06-656c-4a5c-8355-6771d3933a01 service nova] Acquiring lock "refresh_cache-1a4b0637-1a56-41ef-b89b-6b56d24ed206" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 784.333746] env[62066]: DEBUG oslo_concurrency.lockutils [None req-064a3fac-0a45-4f27-8eed-cb07d9679634 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Releasing lock "refresh_cache-1a4b0637-1a56-41ef-b89b-6b56d24ed206" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 784.334080] env[62066]: DEBUG nova.compute.manager [None req-064a3fac-0a45-4f27-8eed-cb07d9679634 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] [instance: 1a4b0637-1a56-41ef-b89b-6b56d24ed206] Instance network_info: |[{"id": "84936b80-0721-4235-a92a-7bc4adf450d8", "address": "fa:16:3e:f7:5a:62", "network": {"id": "518e5084-b25d-4931-9319-a62810398063", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.233", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "eb52f7069a374c61ae946f052007c6d9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c6365036-aa37-44d2-90d1-ca1c3516ded9", "external-id": "nsx-vlan-transportzone-66", "segmentation_id": 66, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap84936b80-07", "ovs_interfaceid": "84936b80-0721-4235-a92a-7bc4adf450d8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 784.334611] env[62066]: DEBUG oslo_concurrency.lockutils [req-22fe4902-9412-456b-a13f-99994b349398 req-bbbc2a06-656c-4a5c-8355-6771d3933a01 service nova] Acquired lock "refresh_cache-1a4b0637-1a56-41ef-b89b-6b56d24ed206" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 784.334801] env[62066]: DEBUG nova.network.neutron [req-22fe4902-9412-456b-a13f-99994b349398 req-bbbc2a06-656c-4a5c-8355-6771d3933a01 service nova] [instance: 1a4b0637-1a56-41ef-b89b-6b56d24ed206] Refreshing network info cache for port 84936b80-0721-4235-a92a-7bc4adf450d8 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 784.337074] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-064a3fac-0a45-4f27-8eed-cb07d9679634 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] [instance: 1a4b0637-1a56-41ef-b89b-6b56d24ed206] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f7:5a:62', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c6365036-aa37-44d2-90d1-ca1c3516ded9', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '84936b80-0721-4235-a92a-7bc4adf450d8', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 784.343865] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-064a3fac-0a45-4f27-8eed-cb07d9679634 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Creating folder: Project (638c4c550925459ea6fc75090dbacde0). Parent ref: group-v285980. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 784.344908] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-045d394a-621d-4039-b58a-87642ecf0075 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.356041] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-064a3fac-0a45-4f27-8eed-cb07d9679634 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Created folder: Project (638c4c550925459ea6fc75090dbacde0) in parent group-v285980. [ 784.356041] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-064a3fac-0a45-4f27-8eed-cb07d9679634 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Creating folder: Instances. Parent ref: group-v286008. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 784.356223] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f8f65a29-e889-4de5-aa0f-3b2484602a46 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.365258] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-064a3fac-0a45-4f27-8eed-cb07d9679634 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Created folder: Instances in parent group-v286008. [ 784.365426] env[62066]: DEBUG oslo.service.loopingcall [None req-064a3fac-0a45-4f27-8eed-cb07d9679634 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 784.365654] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1a4b0637-1a56-41ef-b89b-6b56d24ed206] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 784.365907] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6ad4410d-d885-4a28-a161-884e718957db {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.385746] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 784.385746] env[62066]: value = "task-1340708" [ 784.385746] env[62066]: _type = "Task" [ 784.385746] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 784.393835] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1340708, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.396551] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3baf3239-22c0-4858-82f6-d3b75bd59a7c tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.323s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 784.397084] env[62066]: DEBUG nova.compute.manager [None req-3baf3239-22c0-4858-82f6-d3b75bd59a7c tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] [instance: 1e1a2ede-4ad8-4600-851e-6f2046b3f919] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 784.399991] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7fc7916e-6e82-4e11-a2d0-03a259fc6439 tempest-ServerAddressesNegativeTestJSON-1804754632 tempest-ServerAddressesNegativeTestJSON-1804754632-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.484s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 784.427419] env[62066]: DEBUG nova.compute.manager [None req-df48af02-4c4b-4364-8d38-efc343bf8241 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 49f46244-34fa-48a1-95a2-8e95850f345d] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 784.730505] env[62066]: DEBUG oslo_vmware.api [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Task: {'id': task-1340705, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.471725} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 784.730762] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk to [datastore2] 98fb270c-f3f8-4375-8b5c-c2279305d476/98fb270c-f3f8-4375-8b5c-c2279305d476.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 784.731024] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 98fb270c-f3f8-4375-8b5c-c2279305d476] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 784.731281] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-371c5aa0-942c-4482-8a30-b22ec9a6db3c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.737302] env[62066]: DEBUG oslo_vmware.api [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Waiting for the task: (returnval){ [ 784.737302] env[62066]: value = "task-1340709" [ 784.737302] env[62066]: _type = "Task" [ 784.737302] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 784.745178] env[62066]: DEBUG oslo_vmware.api [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Task: {'id': task-1340709, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.897328] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1340708, 'name': CreateVM_Task} progress is 99%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.905743] env[62066]: DEBUG nova.compute.utils [None req-3baf3239-22c0-4858-82f6-d3b75bd59a7c tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 784.910927] env[62066]: DEBUG nova.compute.manager [None req-3baf3239-22c0-4858-82f6-d3b75bd59a7c tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] [instance: 1e1a2ede-4ad8-4600-851e-6f2046b3f919] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 784.911176] env[62066]: DEBUG nova.network.neutron [None req-3baf3239-22c0-4858-82f6-d3b75bd59a7c tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] [instance: 1e1a2ede-4ad8-4600-851e-6f2046b3f919] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 784.945575] env[62066]: DEBUG oslo_concurrency.lockutils [None req-df48af02-4c4b-4364-8d38-efc343bf8241 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 784.968286] env[62066]: DEBUG nova.policy [None req-3baf3239-22c0-4858-82f6-d3b75bd59a7c tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '34ad59e83ec04c8b82bf1fbca2988d48', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '638c4c550925459ea6fc75090dbacde0', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 785.152326] env[62066]: DEBUG nova.network.neutron [req-22fe4902-9412-456b-a13f-99994b349398 req-bbbc2a06-656c-4a5c-8355-6771d3933a01 service nova] [instance: 1a4b0637-1a56-41ef-b89b-6b56d24ed206] Updated VIF entry in instance network info cache for port 84936b80-0721-4235-a92a-7bc4adf450d8. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 785.152326] env[62066]: DEBUG nova.network.neutron [req-22fe4902-9412-456b-a13f-99994b349398 req-bbbc2a06-656c-4a5c-8355-6771d3933a01 service nova] [instance: 1a4b0637-1a56-41ef-b89b-6b56d24ed206] Updating instance_info_cache with network_info: [{"id": "84936b80-0721-4235-a92a-7bc4adf450d8", "address": "fa:16:3e:f7:5a:62", "network": {"id": "518e5084-b25d-4931-9319-a62810398063", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.233", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "eb52f7069a374c61ae946f052007c6d9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c6365036-aa37-44d2-90d1-ca1c3516ded9", "external-id": "nsx-vlan-transportzone-66", "segmentation_id": 66, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap84936b80-07", "ovs_interfaceid": "84936b80-0721-4235-a92a-7bc4adf450d8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 785.183718] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f631124f-a710-476c-94a3-8dfd662a46f0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.192410] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2393a51a-3108-4952-bfb1-249407563b8c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.223170] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4d5212d-5eac-4255-a83b-cec6a6dd3342 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.230782] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acd5cb2f-6ad9-4d27-9537-9fcbff39a52e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.244120] env[62066]: DEBUG nova.compute.provider_tree [None req-7fc7916e-6e82-4e11-a2d0-03a259fc6439 tempest-ServerAddressesNegativeTestJSON-1804754632 tempest-ServerAddressesNegativeTestJSON-1804754632-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 785.252639] env[62066]: DEBUG oslo_vmware.api [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Task: {'id': task-1340709, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.147731} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 785.252899] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 98fb270c-f3f8-4375-8b5c-c2279305d476] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 785.253811] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eeb246b3-bb26-410a-a84b-6968abd516c6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.277266] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 98fb270c-f3f8-4375-8b5c-c2279305d476] Reconfiguring VM instance instance-00000039 to attach disk [datastore2] 98fb270c-f3f8-4375-8b5c-c2279305d476/98fb270c-f3f8-4375-8b5c-c2279305d476.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 785.278268] env[62066]: DEBUG nova.network.neutron [None req-3baf3239-22c0-4858-82f6-d3b75bd59a7c tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] [instance: 1e1a2ede-4ad8-4600-851e-6f2046b3f919] Successfully created port: 9f537aea-f65c-4bcb-bd50-c419586c6aea {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 785.280499] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-39fd9393-bd2a-4ad6-bac1-2552fd49dff6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.304786] env[62066]: DEBUG oslo_vmware.api [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Waiting for the task: (returnval){ [ 785.304786] env[62066]: value = "task-1340710" [ 785.304786] env[62066]: _type = "Task" [ 785.304786] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 785.313387] env[62066]: DEBUG oslo_vmware.api [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Task: {'id': task-1340710, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.323266] env[62066]: INFO nova.scheduler.client.report [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] Deleted allocations for instance f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5 [ 785.396525] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1340708, 'name': CreateVM_Task, 'duration_secs': 0.610255} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 785.396704] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1a4b0637-1a56-41ef-b89b-6b56d24ed206] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 785.397383] env[62066]: DEBUG oslo_concurrency.lockutils [None req-064a3fac-0a45-4f27-8eed-cb07d9679634 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 785.397582] env[62066]: DEBUG oslo_concurrency.lockutils [None req-064a3fac-0a45-4f27-8eed-cb07d9679634 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Acquired lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 785.397866] env[62066]: DEBUG oslo_concurrency.lockutils [None req-064a3fac-0a45-4f27-8eed-cb07d9679634 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 785.398135] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f32d16ef-0797-453d-a9eb-61c4ce2ce355 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.403061] env[62066]: DEBUG oslo_vmware.api [None req-064a3fac-0a45-4f27-8eed-cb07d9679634 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Waiting for the task: (returnval){ [ 785.403061] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]52fc9418-b9cf-52e6-7270-401bb5edceb1" [ 785.403061] env[62066]: _type = "Task" [ 785.403061] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 785.411119] env[62066]: DEBUG oslo_vmware.api [None req-064a3fac-0a45-4f27-8eed-cb07d9679634 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52fc9418-b9cf-52e6-7270-401bb5edceb1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.412022] env[62066]: DEBUG nova.compute.manager [None req-3baf3239-22c0-4858-82f6-d3b75bd59a7c tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] [instance: 1e1a2ede-4ad8-4600-851e-6f2046b3f919] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 785.655647] env[62066]: DEBUG oslo_concurrency.lockutils [req-22fe4902-9412-456b-a13f-99994b349398 req-bbbc2a06-656c-4a5c-8355-6771d3933a01 service nova] Releasing lock "refresh_cache-1a4b0637-1a56-41ef-b89b-6b56d24ed206" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 785.749615] env[62066]: DEBUG nova.scheduler.client.report [None req-7fc7916e-6e82-4e11-a2d0-03a259fc6439 tempest-ServerAddressesNegativeTestJSON-1804754632 tempest-ServerAddressesNegativeTestJSON-1804754632-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 785.814850] env[62066]: DEBUG oslo_vmware.api [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Task: {'id': task-1340710, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.832891] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ef5fb6cd-37d2-4ad5-b2ee-d23095df949c tempest-ListServersNegativeTestJSON-582260731 tempest-ListServersNegativeTestJSON-582260731-project-member] Lock "f9512c6b-3e01-42f6-8d83-1dd7d3b9c9a5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 162.464s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 785.914481] env[62066]: DEBUG oslo_vmware.api [None req-064a3fac-0a45-4f27-8eed-cb07d9679634 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52fc9418-b9cf-52e6-7270-401bb5edceb1, 'name': SearchDatastore_Task, 'duration_secs': 0.009059} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 785.917848] env[62066]: DEBUG oslo_concurrency.lockutils [None req-064a3fac-0a45-4f27-8eed-cb07d9679634 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Releasing lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 785.918100] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-064a3fac-0a45-4f27-8eed-cb07d9679634 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] [instance: 1a4b0637-1a56-41ef-b89b-6b56d24ed206] Processing image 50ff584c-3b50-4395-af07-3e66769bc9f7 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 785.918336] env[62066]: DEBUG oslo_concurrency.lockutils [None req-064a3fac-0a45-4f27-8eed-cb07d9679634 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 785.918481] env[62066]: DEBUG oslo_concurrency.lockutils [None req-064a3fac-0a45-4f27-8eed-cb07d9679634 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Acquired lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 785.918656] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-064a3fac-0a45-4f27-8eed-cb07d9679634 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 785.919131] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5dd61466-f73e-4e92-8d57-5d9ac3044e5b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.927039] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-064a3fac-0a45-4f27-8eed-cb07d9679634 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 785.927198] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-064a3fac-0a45-4f27-8eed-cb07d9679634 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 785.927881] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8cb41cf4-69f9-444f-890e-ad2b26d3c3d0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.933247] env[62066]: DEBUG oslo_vmware.api [None req-064a3fac-0a45-4f27-8eed-cb07d9679634 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Waiting for the task: (returnval){ [ 785.933247] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]5223bc10-becb-64e0-95db-35b2350a76ff" [ 785.933247] env[62066]: _type = "Task" [ 785.933247] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 785.942402] env[62066]: DEBUG oslo_vmware.api [None req-064a3fac-0a45-4f27-8eed-cb07d9679634 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]5223bc10-becb-64e0-95db-35b2350a76ff, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.254784] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7fc7916e-6e82-4e11-a2d0-03a259fc6439 tempest-ServerAddressesNegativeTestJSON-1804754632 tempest-ServerAddressesNegativeTestJSON-1804754632-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.855s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 786.255460] env[62066]: ERROR nova.compute.manager [None req-7fc7916e-6e82-4e11-a2d0-03a259fc6439 tempest-ServerAddressesNegativeTestJSON-1804754632 tempest-ServerAddressesNegativeTestJSON-1804754632-project-member] [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 17996c34-360e-41b6-9b1e-f194859938a1, please check neutron logs for more information. [ 786.255460] env[62066]: ERROR nova.compute.manager [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] Traceback (most recent call last): [ 786.255460] env[62066]: ERROR nova.compute.manager [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 786.255460] env[62066]: ERROR nova.compute.manager [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] self.driver.spawn(context, instance, image_meta, [ 786.255460] env[62066]: ERROR nova.compute.manager [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 786.255460] env[62066]: ERROR nova.compute.manager [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] self._vmops.spawn(context, instance, image_meta, injected_files, [ 786.255460] env[62066]: ERROR nova.compute.manager [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 786.255460] env[62066]: ERROR nova.compute.manager [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] vm_ref = self.build_virtual_machine(instance, [ 786.255460] env[62066]: ERROR nova.compute.manager [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 786.255460] env[62066]: ERROR nova.compute.manager [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] vif_infos = vmwarevif.get_vif_info(self._session, [ 786.255460] env[62066]: ERROR nova.compute.manager [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 786.255756] env[62066]: ERROR nova.compute.manager [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] for vif in network_info: [ 786.255756] env[62066]: ERROR nova.compute.manager [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 786.255756] env[62066]: ERROR nova.compute.manager [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] return self._sync_wrapper(fn, *args, **kwargs) [ 786.255756] env[62066]: ERROR nova.compute.manager [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 786.255756] env[62066]: ERROR nova.compute.manager [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] self.wait() [ 786.255756] env[62066]: ERROR nova.compute.manager [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 786.255756] env[62066]: ERROR nova.compute.manager [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] self[:] = self._gt.wait() [ 786.255756] env[62066]: ERROR nova.compute.manager [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 786.255756] env[62066]: ERROR nova.compute.manager [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] return self._exit_event.wait() [ 786.255756] env[62066]: ERROR nova.compute.manager [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 786.255756] env[62066]: ERROR nova.compute.manager [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] result = hub.switch() [ 786.255756] env[62066]: ERROR nova.compute.manager [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 786.255756] env[62066]: ERROR nova.compute.manager [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] return self.greenlet.switch() [ 786.256119] env[62066]: ERROR nova.compute.manager [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 786.256119] env[62066]: ERROR nova.compute.manager [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] result = function(*args, **kwargs) [ 786.256119] env[62066]: ERROR nova.compute.manager [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 786.256119] env[62066]: ERROR nova.compute.manager [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] return func(*args, **kwargs) [ 786.256119] env[62066]: ERROR nova.compute.manager [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 786.256119] env[62066]: ERROR nova.compute.manager [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] raise e [ 786.256119] env[62066]: ERROR nova.compute.manager [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 786.256119] env[62066]: ERROR nova.compute.manager [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] nwinfo = self.network_api.allocate_for_instance( [ 786.256119] env[62066]: ERROR nova.compute.manager [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 786.256119] env[62066]: ERROR nova.compute.manager [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] created_port_ids = self._update_ports_for_instance( [ 786.256119] env[62066]: ERROR nova.compute.manager [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 786.256119] env[62066]: ERROR nova.compute.manager [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] with excutils.save_and_reraise_exception(): [ 786.256119] env[62066]: ERROR nova.compute.manager [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 786.256477] env[62066]: ERROR nova.compute.manager [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] self.force_reraise() [ 786.256477] env[62066]: ERROR nova.compute.manager [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 786.256477] env[62066]: ERROR nova.compute.manager [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] raise self.value [ 786.256477] env[62066]: ERROR nova.compute.manager [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 786.256477] env[62066]: ERROR nova.compute.manager [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] updated_port = self._update_port( [ 786.256477] env[62066]: ERROR nova.compute.manager [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 786.256477] env[62066]: ERROR nova.compute.manager [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] _ensure_no_port_binding_failure(port) [ 786.256477] env[62066]: ERROR nova.compute.manager [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 786.256477] env[62066]: ERROR nova.compute.manager [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] raise exception.PortBindingFailed(port_id=port['id']) [ 786.256477] env[62066]: ERROR nova.compute.manager [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] nova.exception.PortBindingFailed: Binding failed for port 17996c34-360e-41b6-9b1e-f194859938a1, please check neutron logs for more information. [ 786.256477] env[62066]: ERROR nova.compute.manager [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] [ 786.256768] env[62066]: DEBUG nova.compute.utils [None req-7fc7916e-6e82-4e11-a2d0-03a259fc6439 tempest-ServerAddressesNegativeTestJSON-1804754632 tempest-ServerAddressesNegativeTestJSON-1804754632-project-member] [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] Binding failed for port 17996c34-360e-41b6-9b1e-f194859938a1, please check neutron logs for more information. {{(pid=62066) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 786.257561] env[62066]: DEBUG oslo_concurrency.lockutils [None req-86820846-d757-4de7-918c-ad5501b560db tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.923s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 786.259262] env[62066]: INFO nova.compute.claims [None req-86820846-d757-4de7-918c-ad5501b560db tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 786.262287] env[62066]: DEBUG nova.compute.manager [None req-7fc7916e-6e82-4e11-a2d0-03a259fc6439 tempest-ServerAddressesNegativeTestJSON-1804754632 tempest-ServerAddressesNegativeTestJSON-1804754632-project-member] [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] Build of instance 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef was re-scheduled: Binding failed for port 17996c34-360e-41b6-9b1e-f194859938a1, please check neutron logs for more information. {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 786.262719] env[62066]: DEBUG nova.compute.manager [None req-7fc7916e-6e82-4e11-a2d0-03a259fc6439 tempest-ServerAddressesNegativeTestJSON-1804754632 tempest-ServerAddressesNegativeTestJSON-1804754632-project-member] [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] Unplugging VIFs for instance {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 786.262948] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7fc7916e-6e82-4e11-a2d0-03a259fc6439 tempest-ServerAddressesNegativeTestJSON-1804754632 tempest-ServerAddressesNegativeTestJSON-1804754632-project-member] Acquiring lock "refresh_cache-239f0b4b-32cf-4f9f-a47e-bac4235fb6ef" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 786.263111] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7fc7916e-6e82-4e11-a2d0-03a259fc6439 tempest-ServerAddressesNegativeTestJSON-1804754632 tempest-ServerAddressesNegativeTestJSON-1804754632-project-member] Acquired lock "refresh_cache-239f0b4b-32cf-4f9f-a47e-bac4235fb6ef" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 786.263276] env[62066]: DEBUG nova.network.neutron [None req-7fc7916e-6e82-4e11-a2d0-03a259fc6439 tempest-ServerAddressesNegativeTestJSON-1804754632 tempest-ServerAddressesNegativeTestJSON-1804754632-project-member] [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 786.315896] env[62066]: DEBUG oslo_vmware.api [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Task: {'id': task-1340710, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.335774] env[62066]: DEBUG nova.compute.manager [None req-8034bd7f-2294-4f4c-b924-f35bec7ebaac tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] [instance: 738ebb2f-7cc6-4d0f-871c-1428c9c21384] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 786.420762] env[62066]: DEBUG nova.compute.manager [None req-3baf3239-22c0-4858-82f6-d3b75bd59a7c tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] [instance: 1e1a2ede-4ad8-4600-851e-6f2046b3f919] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 786.446265] env[62066]: DEBUG oslo_vmware.api [None req-064a3fac-0a45-4f27-8eed-cb07d9679634 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]5223bc10-becb-64e0-95db-35b2350a76ff, 'name': SearchDatastore_Task, 'duration_secs': 0.01151} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 786.448556] env[62066]: DEBUG nova.virt.hardware [None req-3baf3239-22c0-4858-82f6-d3b75bd59a7c tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-23T13:40:41Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-23T13:40:23Z,direct_url=,disk_format='vmdk',id=50ff584c-3b50-4395-af07-3e66769bc9f7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='eb52f7069a374c61ae946f052007c6d9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-23T13:40:24Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 786.448783] env[62066]: DEBUG nova.virt.hardware [None req-3baf3239-22c0-4858-82f6-d3b75bd59a7c tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 786.448939] env[62066]: DEBUG nova.virt.hardware [None req-3baf3239-22c0-4858-82f6-d3b75bd59a7c tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 786.449171] env[62066]: DEBUG nova.virt.hardware [None req-3baf3239-22c0-4858-82f6-d3b75bd59a7c tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 786.449429] env[62066]: DEBUG nova.virt.hardware [None req-3baf3239-22c0-4858-82f6-d3b75bd59a7c tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 786.449653] env[62066]: DEBUG nova.virt.hardware [None req-3baf3239-22c0-4858-82f6-d3b75bd59a7c tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 786.449919] env[62066]: DEBUG nova.virt.hardware [None req-3baf3239-22c0-4858-82f6-d3b75bd59a7c tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 786.450143] env[62066]: DEBUG nova.virt.hardware [None req-3baf3239-22c0-4858-82f6-d3b75bd59a7c tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 786.450377] env[62066]: DEBUG nova.virt.hardware [None req-3baf3239-22c0-4858-82f6-d3b75bd59a7c tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 786.450565] env[62066]: DEBUG nova.virt.hardware [None req-3baf3239-22c0-4858-82f6-d3b75bd59a7c tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 786.450740] env[62066]: DEBUG nova.virt.hardware [None req-3baf3239-22c0-4858-82f6-d3b75bd59a7c tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 786.452018] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-705e1d40-7988-4887-86b5-bfc4244890a9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.454714] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-516024b7-0f9f-4e93-9a0d-7d693e1266cf {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.464110] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f69a147f-99d6-41a3-acc4-0117fcc3ef19 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.467843] env[62066]: DEBUG oslo_vmware.api [None req-064a3fac-0a45-4f27-8eed-cb07d9679634 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Waiting for the task: (returnval){ [ 786.467843] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]52b4e04d-6a9b-8eb9-2ee1-a69d325ca1eb" [ 786.467843] env[62066]: _type = "Task" [ 786.467843] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.485047] env[62066]: DEBUG oslo_vmware.api [None req-064a3fac-0a45-4f27-8eed-cb07d9679634 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52b4e04d-6a9b-8eb9-2ee1-a69d325ca1eb, 'name': SearchDatastore_Task, 'duration_secs': 0.008663} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 786.485341] env[62066]: DEBUG oslo_concurrency.lockutils [None req-064a3fac-0a45-4f27-8eed-cb07d9679634 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Releasing lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 786.485601] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-064a3fac-0a45-4f27-8eed-cb07d9679634 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk to [datastore2] 1a4b0637-1a56-41ef-b89b-6b56d24ed206/1a4b0637-1a56-41ef-b89b-6b56d24ed206.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 786.485879] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-49985d88-5d6d-4e84-b2e3-d22aa099f94a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.491581] env[62066]: DEBUG oslo_vmware.api [None req-064a3fac-0a45-4f27-8eed-cb07d9679634 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Waiting for the task: (returnval){ [ 786.491581] env[62066]: value = "task-1340711" [ 786.491581] env[62066]: _type = "Task" [ 786.491581] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.499726] env[62066]: DEBUG oslo_vmware.api [None req-064a3fac-0a45-4f27-8eed-cb07d9679634 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Task: {'id': task-1340711, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.806491] env[62066]: DEBUG nova.network.neutron [None req-7fc7916e-6e82-4e11-a2d0-03a259fc6439 tempest-ServerAddressesNegativeTestJSON-1804754632 tempest-ServerAddressesNegativeTestJSON-1804754632-project-member] [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 786.818174] env[62066]: DEBUG oslo_vmware.api [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Task: {'id': task-1340710, 'name': ReconfigVM_Task, 'duration_secs': 1.374966} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 786.818475] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 98fb270c-f3f8-4375-8b5c-c2279305d476] Reconfigured VM instance instance-00000039 to attach disk [datastore2] 98fb270c-f3f8-4375-8b5c-c2279305d476/98fb270c-f3f8-4375-8b5c-c2279305d476.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 786.819160] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-42fc6aec-33a5-4ede-80c4-cff696c790fc {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.825312] env[62066]: DEBUG nova.compute.manager [req-b5fa30ed-c049-482b-bad9-1df62db40447 req-6398c78d-62bd-44cb-a72e-d7a4d7ed10b3 service nova] [instance: 1e1a2ede-4ad8-4600-851e-6f2046b3f919] Received event network-vif-plugged-9f537aea-f65c-4bcb-bd50-c419586c6aea {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 786.825529] env[62066]: DEBUG oslo_concurrency.lockutils [req-b5fa30ed-c049-482b-bad9-1df62db40447 req-6398c78d-62bd-44cb-a72e-d7a4d7ed10b3 service nova] Acquiring lock "1e1a2ede-4ad8-4600-851e-6f2046b3f919-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 786.825826] env[62066]: DEBUG oslo_concurrency.lockutils [req-b5fa30ed-c049-482b-bad9-1df62db40447 req-6398c78d-62bd-44cb-a72e-d7a4d7ed10b3 service nova] Lock "1e1a2ede-4ad8-4600-851e-6f2046b3f919-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 786.825896] env[62066]: DEBUG oslo_concurrency.lockutils [req-b5fa30ed-c049-482b-bad9-1df62db40447 req-6398c78d-62bd-44cb-a72e-d7a4d7ed10b3 service nova] Lock "1e1a2ede-4ad8-4600-851e-6f2046b3f919-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 786.826207] env[62066]: DEBUG nova.compute.manager [req-b5fa30ed-c049-482b-bad9-1df62db40447 req-6398c78d-62bd-44cb-a72e-d7a4d7ed10b3 service nova] [instance: 1e1a2ede-4ad8-4600-851e-6f2046b3f919] No waiting events found dispatching network-vif-plugged-9f537aea-f65c-4bcb-bd50-c419586c6aea {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 786.826240] env[62066]: WARNING nova.compute.manager [req-b5fa30ed-c049-482b-bad9-1df62db40447 req-6398c78d-62bd-44cb-a72e-d7a4d7ed10b3 service nova] [instance: 1e1a2ede-4ad8-4600-851e-6f2046b3f919] Received unexpected event network-vif-plugged-9f537aea-f65c-4bcb-bd50-c419586c6aea for instance with vm_state building and task_state spawning. [ 786.828456] env[62066]: DEBUG oslo_vmware.api [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Waiting for the task: (returnval){ [ 786.828456] env[62066]: value = "task-1340712" [ 786.828456] env[62066]: _type = "Task" [ 786.828456] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.837721] env[62066]: DEBUG oslo_vmware.api [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Task: {'id': task-1340712, 'name': Rename_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.864639] env[62066]: DEBUG nova.network.neutron [None req-3baf3239-22c0-4858-82f6-d3b75bd59a7c tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] [instance: 1e1a2ede-4ad8-4600-851e-6f2046b3f919] Successfully updated port: 9f537aea-f65c-4bcb-bd50-c419586c6aea {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 786.880081] env[62066]: DEBUG oslo_concurrency.lockutils [None req-8034bd7f-2294-4f4c-b924-f35bec7ebaac tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 787.003221] env[62066]: DEBUG oslo_vmware.api [None req-064a3fac-0a45-4f27-8eed-cb07d9679634 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Task: {'id': task-1340711, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.482868} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 787.003503] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-064a3fac-0a45-4f27-8eed-cb07d9679634 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk to [datastore2] 1a4b0637-1a56-41ef-b89b-6b56d24ed206/1a4b0637-1a56-41ef-b89b-6b56d24ed206.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 787.003736] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-064a3fac-0a45-4f27-8eed-cb07d9679634 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] [instance: 1a4b0637-1a56-41ef-b89b-6b56d24ed206] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 787.004039] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ba43a489-4f4f-4671-b87d-b54aa3523111 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.011221] env[62066]: DEBUG oslo_vmware.api [None req-064a3fac-0a45-4f27-8eed-cb07d9679634 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Waiting for the task: (returnval){ [ 787.011221] env[62066]: value = "task-1340713" [ 787.011221] env[62066]: _type = "Task" [ 787.011221] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 787.020178] env[62066]: DEBUG oslo_vmware.api [None req-064a3fac-0a45-4f27-8eed-cb07d9679634 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Task: {'id': task-1340713, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.155688] env[62066]: DEBUG nova.network.neutron [None req-7fc7916e-6e82-4e11-a2d0-03a259fc6439 tempest-ServerAddressesNegativeTestJSON-1804754632 tempest-ServerAddressesNegativeTestJSON-1804754632-project-member] [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 787.340076] env[62066]: DEBUG oslo_vmware.api [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Task: {'id': task-1340712, 'name': Rename_Task, 'duration_secs': 0.209767} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 787.340377] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 98fb270c-f3f8-4375-8b5c-c2279305d476] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 787.340590] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5b94536b-d1da-4fbc-b380-bee0a155a4cb {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.349085] env[62066]: DEBUG oslo_vmware.api [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Waiting for the task: (returnval){ [ 787.349085] env[62066]: value = "task-1340714" [ 787.349085] env[62066]: _type = "Task" [ 787.349085] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 787.359022] env[62066]: DEBUG oslo_vmware.api [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Task: {'id': task-1340714, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.372611] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3baf3239-22c0-4858-82f6-d3b75bd59a7c tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Acquiring lock "refresh_cache-1e1a2ede-4ad8-4600-851e-6f2046b3f919" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 787.372755] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3baf3239-22c0-4858-82f6-d3b75bd59a7c tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Acquired lock "refresh_cache-1e1a2ede-4ad8-4600-851e-6f2046b3f919" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 787.372892] env[62066]: DEBUG nova.network.neutron [None req-3baf3239-22c0-4858-82f6-d3b75bd59a7c tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] [instance: 1e1a2ede-4ad8-4600-851e-6f2046b3f919] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 787.524107] env[62066]: DEBUG oslo_vmware.api [None req-064a3fac-0a45-4f27-8eed-cb07d9679634 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Task: {'id': task-1340713, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064133} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 787.524107] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-064a3fac-0a45-4f27-8eed-cb07d9679634 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] [instance: 1a4b0637-1a56-41ef-b89b-6b56d24ed206] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 787.524925] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe2e4af7-2f00-4d08-9208-1962cb822dee {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.531106] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d50d458-7f5a-43b7-854d-f287612d3460 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.553388] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-064a3fac-0a45-4f27-8eed-cb07d9679634 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] [instance: 1a4b0637-1a56-41ef-b89b-6b56d24ed206] Reconfiguring VM instance instance-0000003a to attach disk [datastore2] 1a4b0637-1a56-41ef-b89b-6b56d24ed206/1a4b0637-1a56-41ef-b89b-6b56d24ed206.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 787.555434] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2e848cb7-8496-42be-80b6-c5ca9ab70c83 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.570509] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9626f827-6952-4a5c-8d9e-fc29a0da8e0c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.602905] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90a85628-31d4-45f5-9d23-65114a9908d8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.605658] env[62066]: DEBUG oslo_vmware.api [None req-064a3fac-0a45-4f27-8eed-cb07d9679634 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Waiting for the task: (returnval){ [ 787.605658] env[62066]: value = "task-1340715" [ 787.605658] env[62066]: _type = "Task" [ 787.605658] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 787.612400] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00b82dce-e5f8-49a3-b61d-2ad2c07ca8c7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.619329] env[62066]: DEBUG oslo_vmware.api [None req-064a3fac-0a45-4f27-8eed-cb07d9679634 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Task: {'id': task-1340715, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.631138] env[62066]: DEBUG nova.compute.provider_tree [None req-86820846-d757-4de7-918c-ad5501b560db tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 787.658909] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7fc7916e-6e82-4e11-a2d0-03a259fc6439 tempest-ServerAddressesNegativeTestJSON-1804754632 tempest-ServerAddressesNegativeTestJSON-1804754632-project-member] Releasing lock "refresh_cache-239f0b4b-32cf-4f9f-a47e-bac4235fb6ef" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 787.659402] env[62066]: DEBUG nova.compute.manager [None req-7fc7916e-6e82-4e11-a2d0-03a259fc6439 tempest-ServerAddressesNegativeTestJSON-1804754632 tempest-ServerAddressesNegativeTestJSON-1804754632-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62066) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 787.659448] env[62066]: DEBUG nova.compute.manager [None req-7fc7916e-6e82-4e11-a2d0-03a259fc6439 tempest-ServerAddressesNegativeTestJSON-1804754632 tempest-ServerAddressesNegativeTestJSON-1804754632-project-member] [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 787.659615] env[62066]: DEBUG nova.network.neutron [None req-7fc7916e-6e82-4e11-a2d0-03a259fc6439 tempest-ServerAddressesNegativeTestJSON-1804754632 tempest-ServerAddressesNegativeTestJSON-1804754632-project-member] [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 787.675399] env[62066]: DEBUG nova.network.neutron [None req-7fc7916e-6e82-4e11-a2d0-03a259fc6439 tempest-ServerAddressesNegativeTestJSON-1804754632 tempest-ServerAddressesNegativeTestJSON-1804754632-project-member] [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 787.861739] env[62066]: DEBUG oslo_vmware.api [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Task: {'id': task-1340714, 'name': PowerOnVM_Task} progress is 64%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.942367] env[62066]: DEBUG nova.network.neutron [None req-3baf3239-22c0-4858-82f6-d3b75bd59a7c tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] [instance: 1e1a2ede-4ad8-4600-851e-6f2046b3f919] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 788.116276] env[62066]: DEBUG oslo_vmware.api [None req-064a3fac-0a45-4f27-8eed-cb07d9679634 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Task: {'id': task-1340715, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.134837] env[62066]: DEBUG nova.scheduler.client.report [None req-86820846-d757-4de7-918c-ad5501b560db tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 788.177740] env[62066]: DEBUG nova.network.neutron [None req-7fc7916e-6e82-4e11-a2d0-03a259fc6439 tempest-ServerAddressesNegativeTestJSON-1804754632 tempest-ServerAddressesNegativeTestJSON-1804754632-project-member] [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 788.203159] env[62066]: DEBUG nova.network.neutron [None req-3baf3239-22c0-4858-82f6-d3b75bd59a7c tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] [instance: 1e1a2ede-4ad8-4600-851e-6f2046b3f919] Updating instance_info_cache with network_info: [{"id": "9f537aea-f65c-4bcb-bd50-c419586c6aea", "address": "fa:16:3e:9c:15:b7", "network": {"id": "518e5084-b25d-4931-9319-a62810398063", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "eb52f7069a374c61ae946f052007c6d9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c6365036-aa37-44d2-90d1-ca1c3516ded9", "external-id": "nsx-vlan-transportzone-66", "segmentation_id": 66, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9f537aea-f6", "ovs_interfaceid": "9f537aea-f65c-4bcb-bd50-c419586c6aea", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 788.360957] env[62066]: DEBUG oslo_vmware.api [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Task: {'id': task-1340714, 'name': PowerOnVM_Task} progress is 82%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.617525] env[62066]: DEBUG oslo_vmware.api [None req-064a3fac-0a45-4f27-8eed-cb07d9679634 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Task: {'id': task-1340715, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.639555] env[62066]: DEBUG oslo_concurrency.lockutils [None req-86820846-d757-4de7-918c-ad5501b560db tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.382s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 788.639998] env[62066]: DEBUG nova.compute.manager [None req-86820846-d757-4de7-918c-ad5501b560db tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 788.644783] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c70bcb7f-78ff-424d-8121-684c45ecd12b tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.464s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 788.644783] env[62066]: INFO nova.compute.claims [None req-c70bcb7f-78ff-424d-8121-684c45ecd12b tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 02fb3e92-5dd6-4b1e-a6e4-d60d3fc85b07] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 788.680529] env[62066]: INFO nova.compute.manager [None req-7fc7916e-6e82-4e11-a2d0-03a259fc6439 tempest-ServerAddressesNegativeTestJSON-1804754632 tempest-ServerAddressesNegativeTestJSON-1804754632-project-member] [instance: 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef] Took 1.02 seconds to deallocate network for instance. [ 788.705577] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3baf3239-22c0-4858-82f6-d3b75bd59a7c tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Releasing lock "refresh_cache-1e1a2ede-4ad8-4600-851e-6f2046b3f919" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 788.705881] env[62066]: DEBUG nova.compute.manager [None req-3baf3239-22c0-4858-82f6-d3b75bd59a7c tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] [instance: 1e1a2ede-4ad8-4600-851e-6f2046b3f919] Instance network_info: |[{"id": "9f537aea-f65c-4bcb-bd50-c419586c6aea", "address": "fa:16:3e:9c:15:b7", "network": {"id": "518e5084-b25d-4931-9319-a62810398063", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "eb52f7069a374c61ae946f052007c6d9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c6365036-aa37-44d2-90d1-ca1c3516ded9", "external-id": "nsx-vlan-transportzone-66", "segmentation_id": 66, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9f537aea-f6", "ovs_interfaceid": "9f537aea-f65c-4bcb-bd50-c419586c6aea", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 788.706384] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-3baf3239-22c0-4858-82f6-d3b75bd59a7c tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] [instance: 1e1a2ede-4ad8-4600-851e-6f2046b3f919] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:9c:15:b7', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c6365036-aa37-44d2-90d1-ca1c3516ded9', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9f537aea-f65c-4bcb-bd50-c419586c6aea', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 788.714505] env[62066]: DEBUG oslo.service.loopingcall [None req-3baf3239-22c0-4858-82f6-d3b75bd59a7c tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 788.715324] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1e1a2ede-4ad8-4600-851e-6f2046b3f919] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 788.715553] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0b203f1c-afd6-479c-88ae-81abc677c6ea {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.739408] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 788.739408] env[62066]: value = "task-1340716" [ 788.739408] env[62066]: _type = "Task" [ 788.739408] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 788.746958] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1340716, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.861880] env[62066]: DEBUG oslo_vmware.api [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Task: {'id': task-1340714, 'name': PowerOnVM_Task, 'duration_secs': 1.095034} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 788.862140] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 98fb270c-f3f8-4375-8b5c-c2279305d476] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 788.862328] env[62066]: INFO nova.compute.manager [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 98fb270c-f3f8-4375-8b5c-c2279305d476] Took 8.97 seconds to spawn the instance on the hypervisor. [ 788.862520] env[62066]: DEBUG nova.compute.manager [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 98fb270c-f3f8-4375-8b5c-c2279305d476] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 788.863364] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e36fc976-c38c-425c-8826-95e6caf39fe5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.901050] env[62066]: DEBUG nova.compute.manager [req-1e39eab3-ce6f-4c63-ba82-1c067aea304e req-7983674e-ee81-4f0a-8430-51ea23278195 service nova] [instance: 1e1a2ede-4ad8-4600-851e-6f2046b3f919] Received event network-changed-9f537aea-f65c-4bcb-bd50-c419586c6aea {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 788.901050] env[62066]: DEBUG nova.compute.manager [req-1e39eab3-ce6f-4c63-ba82-1c067aea304e req-7983674e-ee81-4f0a-8430-51ea23278195 service nova] [instance: 1e1a2ede-4ad8-4600-851e-6f2046b3f919] Refreshing instance network info cache due to event network-changed-9f537aea-f65c-4bcb-bd50-c419586c6aea. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 788.901050] env[62066]: DEBUG oslo_concurrency.lockutils [req-1e39eab3-ce6f-4c63-ba82-1c067aea304e req-7983674e-ee81-4f0a-8430-51ea23278195 service nova] Acquiring lock "refresh_cache-1e1a2ede-4ad8-4600-851e-6f2046b3f919" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 788.901050] env[62066]: DEBUG oslo_concurrency.lockutils [req-1e39eab3-ce6f-4c63-ba82-1c067aea304e req-7983674e-ee81-4f0a-8430-51ea23278195 service nova] Acquired lock "refresh_cache-1e1a2ede-4ad8-4600-851e-6f2046b3f919" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 788.901050] env[62066]: DEBUG nova.network.neutron [req-1e39eab3-ce6f-4c63-ba82-1c067aea304e req-7983674e-ee81-4f0a-8430-51ea23278195 service nova] [instance: 1e1a2ede-4ad8-4600-851e-6f2046b3f919] Refreshing network info cache for port 9f537aea-f65c-4bcb-bd50-c419586c6aea {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 789.117694] env[62066]: DEBUG oslo_vmware.api [None req-064a3fac-0a45-4f27-8eed-cb07d9679634 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Task: {'id': task-1340715, 'name': ReconfigVM_Task, 'duration_secs': 1.059771} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.117962] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-064a3fac-0a45-4f27-8eed-cb07d9679634 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] [instance: 1a4b0637-1a56-41ef-b89b-6b56d24ed206] Reconfigured VM instance instance-0000003a to attach disk [datastore2] 1a4b0637-1a56-41ef-b89b-6b56d24ed206/1a4b0637-1a56-41ef-b89b-6b56d24ed206.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 789.118577] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-06a900f0-8ad8-4dbf-a298-f25cd3d94ccb {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.124624] env[62066]: DEBUG oslo_vmware.api [None req-064a3fac-0a45-4f27-8eed-cb07d9679634 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Waiting for the task: (returnval){ [ 789.124624] env[62066]: value = "task-1340717" [ 789.124624] env[62066]: _type = "Task" [ 789.124624] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.133323] env[62066]: DEBUG oslo_vmware.api [None req-064a3fac-0a45-4f27-8eed-cb07d9679634 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Task: {'id': task-1340717, 'name': Rename_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.152269] env[62066]: DEBUG nova.compute.utils [None req-86820846-d757-4de7-918c-ad5501b560db tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 789.154446] env[62066]: DEBUG nova.compute.manager [None req-86820846-d757-4de7-918c-ad5501b560db tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 789.154446] env[62066]: DEBUG nova.network.neutron [None req-86820846-d757-4de7-918c-ad5501b560db tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 789.220320] env[62066]: DEBUG nova.policy [None req-86820846-d757-4de7-918c-ad5501b560db tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a8455bed94694b669d1e6365ffa272e1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0b8c246df0e8404ca3743c7207cfc808', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 789.251416] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1340716, 'name': CreateVM_Task} progress is 99%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.379978] env[62066]: INFO nova.compute.manager [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 98fb270c-f3f8-4375-8b5c-c2279305d476] Took 41.24 seconds to build instance. [ 789.602667] env[62066]: DEBUG nova.network.neutron [None req-86820846-d757-4de7-918c-ad5501b560db tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] Successfully created port: 88380b24-b187-4b82-b92e-b93adeeab3eb {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 789.636562] env[62066]: DEBUG oslo_vmware.api [None req-064a3fac-0a45-4f27-8eed-cb07d9679634 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Task: {'id': task-1340717, 'name': Rename_Task, 'duration_secs': 0.185665} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.637053] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-064a3fac-0a45-4f27-8eed-cb07d9679634 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] [instance: 1a4b0637-1a56-41ef-b89b-6b56d24ed206] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 789.637426] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9ee999c7-caeb-4a7c-881d-924286b3f2bd {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.644377] env[62066]: DEBUG oslo_vmware.api [None req-064a3fac-0a45-4f27-8eed-cb07d9679634 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Waiting for the task: (returnval){ [ 789.644377] env[62066]: value = "task-1340718" [ 789.644377] env[62066]: _type = "Task" [ 789.644377] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.652741] env[62066]: DEBUG oslo_vmware.api [None req-064a3fac-0a45-4f27-8eed-cb07d9679634 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Task: {'id': task-1340718, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.657657] env[62066]: DEBUG nova.compute.manager [None req-86820846-d757-4de7-918c-ad5501b560db tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 789.721703] env[62066]: INFO nova.scheduler.client.report [None req-7fc7916e-6e82-4e11-a2d0-03a259fc6439 tempest-ServerAddressesNegativeTestJSON-1804754632 tempest-ServerAddressesNegativeTestJSON-1804754632-project-member] Deleted allocations for instance 239f0b4b-32cf-4f9f-a47e-bac4235fb6ef [ 789.754516] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1340716, 'name': CreateVM_Task, 'duration_secs': 0.545532} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.756656] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1e1a2ede-4ad8-4600-851e-6f2046b3f919] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 789.760045] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3baf3239-22c0-4858-82f6-d3b75bd59a7c tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 789.760351] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3baf3239-22c0-4858-82f6-d3b75bd59a7c tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Acquired lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 789.762166] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3baf3239-22c0-4858-82f6-d3b75bd59a7c tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 789.764591] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dd559cc4-4776-46f4-a09e-dffd4a5150f1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.769234] env[62066]: DEBUG nova.network.neutron [req-1e39eab3-ce6f-4c63-ba82-1c067aea304e req-7983674e-ee81-4f0a-8430-51ea23278195 service nova] [instance: 1e1a2ede-4ad8-4600-851e-6f2046b3f919] Updated VIF entry in instance network info cache for port 9f537aea-f65c-4bcb-bd50-c419586c6aea. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 789.770528] env[62066]: DEBUG nova.network.neutron [req-1e39eab3-ce6f-4c63-ba82-1c067aea304e req-7983674e-ee81-4f0a-8430-51ea23278195 service nova] [instance: 1e1a2ede-4ad8-4600-851e-6f2046b3f919] Updating instance_info_cache with network_info: [{"id": "9f537aea-f65c-4bcb-bd50-c419586c6aea", "address": "fa:16:3e:9c:15:b7", "network": {"id": "518e5084-b25d-4931-9319-a62810398063", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "eb52f7069a374c61ae946f052007c6d9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c6365036-aa37-44d2-90d1-ca1c3516ded9", "external-id": "nsx-vlan-transportzone-66", "segmentation_id": 66, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9f537aea-f6", "ovs_interfaceid": "9f537aea-f65c-4bcb-bd50-c419586c6aea", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 789.777261] env[62066]: DEBUG oslo_vmware.api [None req-3baf3239-22c0-4858-82f6-d3b75bd59a7c tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Waiting for the task: (returnval){ [ 789.777261] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]52638b37-dab2-103e-cea8-e0a90360b08b" [ 789.777261] env[62066]: _type = "Task" [ 789.777261] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.796499] env[62066]: DEBUG oslo_vmware.api [None req-3baf3239-22c0-4858-82f6-d3b75bd59a7c tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52638b37-dab2-103e-cea8-e0a90360b08b, 'name': SearchDatastore_Task, 'duration_secs': 0.009077} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.796499] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3baf3239-22c0-4858-82f6-d3b75bd59a7c tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Releasing lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 789.796724] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-3baf3239-22c0-4858-82f6-d3b75bd59a7c tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] [instance: 1e1a2ede-4ad8-4600-851e-6f2046b3f919] Processing image 50ff584c-3b50-4395-af07-3e66769bc9f7 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 789.797097] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3baf3239-22c0-4858-82f6-d3b75bd59a7c tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 789.797097] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3baf3239-22c0-4858-82f6-d3b75bd59a7c tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Acquired lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 789.797339] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-3baf3239-22c0-4858-82f6-d3b75bd59a7c tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 789.797860] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a4d17151-b4ac-4e8a-a880-4a562fab91b8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.805871] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-3baf3239-22c0-4858-82f6-d3b75bd59a7c tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 789.806062] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-3baf3239-22c0-4858-82f6-d3b75bd59a7c tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 789.809297] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f313926b-b64a-4527-8206-54ce3d07fda0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.815977] env[62066]: DEBUG oslo_vmware.api [None req-3baf3239-22c0-4858-82f6-d3b75bd59a7c tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Waiting for the task: (returnval){ [ 789.815977] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]5273aef6-4600-c66c-6a2f-8b19b1b96e61" [ 789.815977] env[62066]: _type = "Task" [ 789.815977] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.823489] env[62066]: DEBUG oslo_vmware.api [None req-3baf3239-22c0-4858-82f6-d3b75bd59a7c tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]5273aef6-4600-c66c-6a2f-8b19b1b96e61, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.882761] env[62066]: DEBUG oslo_concurrency.lockutils [None req-6c73d8be-87d4-4de0-9d01-9c612974dc91 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Lock "98fb270c-f3f8-4375-8b5c-c2279305d476" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 158.582s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 789.981498] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1e22fe3-1a6f-4691-9dbb-72bfc90f86bc {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.990020] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c9edbaf-d177-4a59-80a9-8f91d9672a9e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.025287] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92f738f8-4cf2-4234-abd7-2ee8fd10ed0a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.033583] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94b7f08a-e54e-47c0-9476-975a503fe2f0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.045933] env[62066]: DEBUG nova.compute.provider_tree [None req-c70bcb7f-78ff-424d-8121-684c45ecd12b tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 790.155307] env[62066]: DEBUG oslo_vmware.api [None req-064a3fac-0a45-4f27-8eed-cb07d9679634 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Task: {'id': task-1340718, 'name': PowerOnVM_Task} progress is 64%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.163273] env[62066]: INFO nova.virt.block_device [None req-86820846-d757-4de7-918c-ad5501b560db tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] Booting with volume 26ceca45-1cae-48c9-a567-e7cc4f9970c1 at /dev/sda [ 790.210939] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f11ec12a-0838-4754-bdab-71badbde1146 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.220271] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e296caa6-2b62-44c4-9f23-acc868d11907 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.231546] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7fc7916e-6e82-4e11-a2d0-03a259fc6439 tempest-ServerAddressesNegativeTestJSON-1804754632 tempest-ServerAddressesNegativeTestJSON-1804754632-project-member] Lock "239f0b4b-32cf-4f9f-a47e-bac4235fb6ef" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 166.363s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 790.246850] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-174b0290-1803-4a7d-9c57-a260f3247ad4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.255188] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d6ae343-6d03-4e2a-8ef0-e8a455cc47aa {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.265788] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ba3a3138-4a4e-4a10-bbee-93dd3a372fa2 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Acquiring lock "0ebba3b2-f82d-4c1d-b01a-f75c3559c117" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 790.266045] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ba3a3138-4a4e-4a10-bbee-93dd3a372fa2 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Lock "0ebba3b2-f82d-4c1d-b01a-f75c3559c117" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 790.269949] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ba3a3138-4a4e-4a10-bbee-93dd3a372fa2 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Acquiring lock "0ebba3b2-f82d-4c1d-b01a-f75c3559c117-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 790.269949] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ba3a3138-4a4e-4a10-bbee-93dd3a372fa2 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Lock "0ebba3b2-f82d-4c1d-b01a-f75c3559c117-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 790.269949] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ba3a3138-4a4e-4a10-bbee-93dd3a372fa2 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Lock "0ebba3b2-f82d-4c1d-b01a-f75c3559c117-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 790.269949] env[62066]: INFO nova.compute.manager [None req-ba3a3138-4a4e-4a10-bbee-93dd3a372fa2 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 0ebba3b2-f82d-4c1d-b01a-f75c3559c117] Terminating instance [ 790.274152] env[62066]: DEBUG nova.compute.manager [None req-ba3a3138-4a4e-4a10-bbee-93dd3a372fa2 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 0ebba3b2-f82d-4c1d-b01a-f75c3559c117] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 790.274152] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-ba3a3138-4a4e-4a10-bbee-93dd3a372fa2 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 0ebba3b2-f82d-4c1d-b01a-f75c3559c117] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 790.274152] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe0f886b-c63a-43b4-b93d-8202b88c3dea {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.276687] env[62066]: DEBUG oslo_concurrency.lockutils [req-1e39eab3-ce6f-4c63-ba82-1c067aea304e req-7983674e-ee81-4f0a-8430-51ea23278195 service nova] Releasing lock "refresh_cache-1e1a2ede-4ad8-4600-851e-6f2046b3f919" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 790.283859] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c1d15fe-536d-4ede-9b83-90d1b8b54de9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.288474] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-ba3a3138-4a4e-4a10-bbee-93dd3a372fa2 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 0ebba3b2-f82d-4c1d-b01a-f75c3559c117] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 790.288986] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-92edf145-577c-4611-a1eb-10d2cf2ff21d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.293077] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69fc8021-9b67-41ae-a06a-3bc6c044a90d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.297466] env[62066]: DEBUG oslo_vmware.api [None req-ba3a3138-4a4e-4a10-bbee-93dd3a372fa2 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Waiting for the task: (returnval){ [ 790.297466] env[62066]: value = "task-1340719" [ 790.297466] env[62066]: _type = "Task" [ 790.297466] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.305585] env[62066]: DEBUG oslo_vmware.api [None req-ba3a3138-4a4e-4a10-bbee-93dd3a372fa2 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Task: {'id': task-1340719, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.309383] env[62066]: DEBUG nova.virt.block_device [None req-86820846-d757-4de7-918c-ad5501b560db tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] Updating existing volume attachment record: 0334076e-6b3a-4c6b-9474-ed73d3ceeb0c {{(pid=62066) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 790.326176] env[62066]: DEBUG oslo_vmware.api [None req-3baf3239-22c0-4858-82f6-d3b75bd59a7c tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]5273aef6-4600-c66c-6a2f-8b19b1b96e61, 'name': SearchDatastore_Task, 'duration_secs': 0.008377} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.326176] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2255c36c-514f-4b89-bd5d-c7d82a795089 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.331920] env[62066]: DEBUG oslo_vmware.api [None req-3baf3239-22c0-4858-82f6-d3b75bd59a7c tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Waiting for the task: (returnval){ [ 790.331920] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]52be62e9-c0a1-38f7-b4a5-789c5313cbb7" [ 790.331920] env[62066]: _type = "Task" [ 790.331920] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.341176] env[62066]: DEBUG oslo_vmware.api [None req-3baf3239-22c0-4858-82f6-d3b75bd59a7c tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52be62e9-c0a1-38f7-b4a5-789c5313cbb7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.357952] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e5dc4f2e-1d02-4ba9-acb0-6628ba7125f6 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Acquiring lock "98fb270c-f3f8-4375-8b5c-c2279305d476" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 790.358240] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e5dc4f2e-1d02-4ba9-acb0-6628ba7125f6 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Lock "98fb270c-f3f8-4375-8b5c-c2279305d476" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 790.358447] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e5dc4f2e-1d02-4ba9-acb0-6628ba7125f6 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Acquiring lock "98fb270c-f3f8-4375-8b5c-c2279305d476-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 790.358628] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e5dc4f2e-1d02-4ba9-acb0-6628ba7125f6 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Lock "98fb270c-f3f8-4375-8b5c-c2279305d476-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 790.358799] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e5dc4f2e-1d02-4ba9-acb0-6628ba7125f6 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Lock "98fb270c-f3f8-4375-8b5c-c2279305d476-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 790.366766] env[62066]: INFO nova.compute.manager [None req-e5dc4f2e-1d02-4ba9-acb0-6628ba7125f6 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 98fb270c-f3f8-4375-8b5c-c2279305d476] Terminating instance [ 790.367500] env[62066]: DEBUG nova.compute.manager [None req-e5dc4f2e-1d02-4ba9-acb0-6628ba7125f6 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 98fb270c-f3f8-4375-8b5c-c2279305d476] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 790.367613] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-e5dc4f2e-1d02-4ba9-acb0-6628ba7125f6 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 98fb270c-f3f8-4375-8b5c-c2279305d476] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 790.368549] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f941567b-9195-4b9e-bccf-de1a5339735c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.377218] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-e5dc4f2e-1d02-4ba9-acb0-6628ba7125f6 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 98fb270c-f3f8-4375-8b5c-c2279305d476] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 790.377531] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-885285f2-6b14-4d8c-8d4d-bac158096fff {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.385454] env[62066]: DEBUG nova.compute.manager [None req-f6cc809d-278e-4fb4-906b-f468fa41739d tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] [instance: 5f645eaa-6edc-4362-82ad-38c4c57b2be2] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 790.387740] env[62066]: DEBUG oslo_vmware.api [None req-e5dc4f2e-1d02-4ba9-acb0-6628ba7125f6 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Waiting for the task: (returnval){ [ 790.387740] env[62066]: value = "task-1340720" [ 790.387740] env[62066]: _type = "Task" [ 790.387740] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.396526] env[62066]: DEBUG oslo_vmware.api [None req-e5dc4f2e-1d02-4ba9-acb0-6628ba7125f6 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Task: {'id': task-1340720, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.549627] env[62066]: DEBUG nova.scheduler.client.report [None req-c70bcb7f-78ff-424d-8121-684c45ecd12b tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 790.655293] env[62066]: DEBUG oslo_vmware.api [None req-064a3fac-0a45-4f27-8eed-cb07d9679634 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Task: {'id': task-1340718, 'name': PowerOnVM_Task, 'duration_secs': 0.892301} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.655643] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-064a3fac-0a45-4f27-8eed-cb07d9679634 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] [instance: 1a4b0637-1a56-41ef-b89b-6b56d24ed206] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 790.655854] env[62066]: INFO nova.compute.manager [None req-064a3fac-0a45-4f27-8eed-cb07d9679634 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] [instance: 1a4b0637-1a56-41ef-b89b-6b56d24ed206] Took 8.40 seconds to spawn the instance on the hypervisor. [ 790.656117] env[62066]: DEBUG nova.compute.manager [None req-064a3fac-0a45-4f27-8eed-cb07d9679634 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] [instance: 1a4b0637-1a56-41ef-b89b-6b56d24ed206] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 790.656894] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67c43214-62dd-4cd2-8b43-c85cb11fee05 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.734018] env[62066]: DEBUG nova.compute.manager [None req-90f5cbab-7c0d-4b69-92aa-09bf24d9f7da tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] [instance: 050040cb-6ee3-4ad2-960a-fcebb53ac394] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 790.809729] env[62066]: DEBUG oslo_vmware.api [None req-ba3a3138-4a4e-4a10-bbee-93dd3a372fa2 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Task: {'id': task-1340719, 'name': PowerOffVM_Task, 'duration_secs': 0.179215} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.810107] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-ba3a3138-4a4e-4a10-bbee-93dd3a372fa2 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 0ebba3b2-f82d-4c1d-b01a-f75c3559c117] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 790.810372] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-ba3a3138-4a4e-4a10-bbee-93dd3a372fa2 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 0ebba3b2-f82d-4c1d-b01a-f75c3559c117] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 790.811579] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6eca3bb0-93a1-4c60-86f7-9f1a942a464d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.846727] env[62066]: DEBUG oslo_vmware.api [None req-3baf3239-22c0-4858-82f6-d3b75bd59a7c tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52be62e9-c0a1-38f7-b4a5-789c5313cbb7, 'name': SearchDatastore_Task, 'duration_secs': 0.010597} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.847030] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3baf3239-22c0-4858-82f6-d3b75bd59a7c tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Releasing lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 790.847273] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-3baf3239-22c0-4858-82f6-d3b75bd59a7c tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk to [datastore2] 1e1a2ede-4ad8-4600-851e-6f2046b3f919/1e1a2ede-4ad8-4600-851e-6f2046b3f919.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 790.847550] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ea97935e-76f9-474e-9c73-eaa1d62e94b5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.853982] env[62066]: DEBUG oslo_vmware.api [None req-3baf3239-22c0-4858-82f6-d3b75bd59a7c tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Waiting for the task: (returnval){ [ 790.853982] env[62066]: value = "task-1340722" [ 790.853982] env[62066]: _type = "Task" [ 790.853982] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.862372] env[62066]: DEBUG oslo_vmware.api [None req-3baf3239-22c0-4858-82f6-d3b75bd59a7c tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Task: {'id': task-1340722, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.876257] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-ba3a3138-4a4e-4a10-bbee-93dd3a372fa2 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 0ebba3b2-f82d-4c1d-b01a-f75c3559c117] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 790.876630] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-ba3a3138-4a4e-4a10-bbee-93dd3a372fa2 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 0ebba3b2-f82d-4c1d-b01a-f75c3559c117] Deleting contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 790.876883] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-ba3a3138-4a4e-4a10-bbee-93dd3a372fa2 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Deleting the datastore file [datastore2] 0ebba3b2-f82d-4c1d-b01a-f75c3559c117 {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 790.877209] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-08b32f2b-774d-47b7-8508-3bdc02526e7f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.886458] env[62066]: DEBUG oslo_vmware.api [None req-ba3a3138-4a4e-4a10-bbee-93dd3a372fa2 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Waiting for the task: (returnval){ [ 790.886458] env[62066]: value = "task-1340723" [ 790.886458] env[62066]: _type = "Task" [ 790.886458] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.902655] env[62066]: DEBUG oslo_vmware.api [None req-ba3a3138-4a4e-4a10-bbee-93dd3a372fa2 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Task: {'id': task-1340723, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.909307] env[62066]: DEBUG oslo_vmware.api [None req-e5dc4f2e-1d02-4ba9-acb0-6628ba7125f6 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Task: {'id': task-1340720, 'name': PowerOffVM_Task, 'duration_secs': 0.192118} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.909307] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-e5dc4f2e-1d02-4ba9-acb0-6628ba7125f6 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 98fb270c-f3f8-4375-8b5c-c2279305d476] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 790.909307] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-e5dc4f2e-1d02-4ba9-acb0-6628ba7125f6 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 98fb270c-f3f8-4375-8b5c-c2279305d476] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 790.909307] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-df73786f-9f4b-4fdb-a98c-4e1117475200 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.918687] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f6cc809d-278e-4fb4-906b-f468fa41739d tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 790.968905] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-e5dc4f2e-1d02-4ba9-acb0-6628ba7125f6 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 98fb270c-f3f8-4375-8b5c-c2279305d476] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 790.972144] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-e5dc4f2e-1d02-4ba9-acb0-6628ba7125f6 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 98fb270c-f3f8-4375-8b5c-c2279305d476] Deleting contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 790.972144] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-e5dc4f2e-1d02-4ba9-acb0-6628ba7125f6 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Deleting the datastore file [datastore2] 98fb270c-f3f8-4375-8b5c-c2279305d476 {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 790.972144] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-63b61e8c-3f4f-4e35-9c4e-6721d5d47072 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.979132] env[62066]: DEBUG oslo_vmware.api [None req-e5dc4f2e-1d02-4ba9-acb0-6628ba7125f6 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Waiting for the task: (returnval){ [ 790.979132] env[62066]: value = "task-1340725" [ 790.979132] env[62066]: _type = "Task" [ 790.979132] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.985595] env[62066]: DEBUG oslo_vmware.api [None req-e5dc4f2e-1d02-4ba9-acb0-6628ba7125f6 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Task: {'id': task-1340725, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.059734] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c70bcb7f-78ff-424d-8121-684c45ecd12b tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.413s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 791.059734] env[62066]: DEBUG nova.compute.manager [None req-c70bcb7f-78ff-424d-8121-684c45ecd12b tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 02fb3e92-5dd6-4b1e-a6e4-d60d3fc85b07] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 791.062711] env[62066]: DEBUG oslo_concurrency.lockutils [None req-816774ed-d63b-4dd8-94be-e4c70c8f10e7 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.663s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 791.064938] env[62066]: INFO nova.compute.claims [None req-816774ed-d63b-4dd8-94be-e4c70c8f10e7 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: 459f432b-e817-405f-9c3e-dfa5d957788c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 791.180320] env[62066]: INFO nova.compute.manager [None req-064a3fac-0a45-4f27-8eed-cb07d9679634 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] [instance: 1a4b0637-1a56-41ef-b89b-6b56d24ed206] Took 41.05 seconds to build instance. [ 791.256830] env[62066]: DEBUG oslo_concurrency.lockutils [None req-90f5cbab-7c0d-4b69-92aa-09bf24d9f7da tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 791.363563] env[62066]: DEBUG oslo_vmware.api [None req-3baf3239-22c0-4858-82f6-d3b75bd59a7c tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Task: {'id': task-1340722, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.395739] env[62066]: DEBUG oslo_vmware.api [None req-ba3a3138-4a4e-4a10-bbee-93dd3a372fa2 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Task: {'id': task-1340723, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.276528} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 791.396079] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-ba3a3138-4a4e-4a10-bbee-93dd3a372fa2 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 791.396415] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-ba3a3138-4a4e-4a10-bbee-93dd3a372fa2 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 0ebba3b2-f82d-4c1d-b01a-f75c3559c117] Deleted contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 791.396654] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-ba3a3138-4a4e-4a10-bbee-93dd3a372fa2 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 0ebba3b2-f82d-4c1d-b01a-f75c3559c117] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 791.400017] env[62066]: INFO nova.compute.manager [None req-ba3a3138-4a4e-4a10-bbee-93dd3a372fa2 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 0ebba3b2-f82d-4c1d-b01a-f75c3559c117] Took 1.13 seconds to destroy the instance on the hypervisor. [ 791.400017] env[62066]: DEBUG oslo.service.loopingcall [None req-ba3a3138-4a4e-4a10-bbee-93dd3a372fa2 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 791.400017] env[62066]: DEBUG nova.compute.manager [-] [instance: 0ebba3b2-f82d-4c1d-b01a-f75c3559c117] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 791.400017] env[62066]: DEBUG nova.network.neutron [-] [instance: 0ebba3b2-f82d-4c1d-b01a-f75c3559c117] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 791.492891] env[62066]: DEBUG oslo_vmware.api [None req-e5dc4f2e-1d02-4ba9-acb0-6628ba7125f6 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Task: {'id': task-1340725, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.453948} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 791.493575] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-e5dc4f2e-1d02-4ba9-acb0-6628ba7125f6 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 791.493885] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-e5dc4f2e-1d02-4ba9-acb0-6628ba7125f6 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 98fb270c-f3f8-4375-8b5c-c2279305d476] Deleted contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 791.494414] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-e5dc4f2e-1d02-4ba9-acb0-6628ba7125f6 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 98fb270c-f3f8-4375-8b5c-c2279305d476] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 791.494724] env[62066]: INFO nova.compute.manager [None req-e5dc4f2e-1d02-4ba9-acb0-6628ba7125f6 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] [instance: 98fb270c-f3f8-4375-8b5c-c2279305d476] Took 1.13 seconds to destroy the instance on the hypervisor. [ 791.495366] env[62066]: DEBUG oslo.service.loopingcall [None req-e5dc4f2e-1d02-4ba9-acb0-6628ba7125f6 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 791.495978] env[62066]: DEBUG nova.compute.manager [-] [instance: 98fb270c-f3f8-4375-8b5c-c2279305d476] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 791.496202] env[62066]: DEBUG nova.network.neutron [-] [instance: 98fb270c-f3f8-4375-8b5c-c2279305d476] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 791.566424] env[62066]: DEBUG nova.compute.utils [None req-c70bcb7f-78ff-424d-8121-684c45ecd12b tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 791.568223] env[62066]: DEBUG nova.compute.manager [None req-c70bcb7f-78ff-424d-8121-684c45ecd12b tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 02fb3e92-5dd6-4b1e-a6e4-d60d3fc85b07] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 791.568609] env[62066]: DEBUG nova.network.neutron [None req-c70bcb7f-78ff-424d-8121-684c45ecd12b tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 02fb3e92-5dd6-4b1e-a6e4-d60d3fc85b07] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 791.682086] env[62066]: DEBUG oslo_concurrency.lockutils [None req-064a3fac-0a45-4f27-8eed-cb07d9679634 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Lock "1a4b0637-1a56-41ef-b89b-6b56d24ed206" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 147.136s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 791.712754] env[62066]: DEBUG nova.policy [None req-c70bcb7f-78ff-424d-8121-684c45ecd12b tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8910a229218b4ec5ad72c893badfc598', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6497ab02f327476d8ff81c2ecc0371e5', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 791.869539] env[62066]: DEBUG oslo_vmware.api [None req-3baf3239-22c0-4858-82f6-d3b75bd59a7c tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Task: {'id': task-1340722, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.519396} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 791.871145] env[62066]: DEBUG nova.network.neutron [None req-86820846-d757-4de7-918c-ad5501b560db tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] Successfully updated port: 88380b24-b187-4b82-b92e-b93adeeab3eb {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 791.872798] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-3baf3239-22c0-4858-82f6-d3b75bd59a7c tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk to [datastore2] 1e1a2ede-4ad8-4600-851e-6f2046b3f919/1e1a2ede-4ad8-4600-851e-6f2046b3f919.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 791.873315] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-3baf3239-22c0-4858-82f6-d3b75bd59a7c tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] [instance: 1e1a2ede-4ad8-4600-851e-6f2046b3f919] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 791.873545] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0ed14b58-d213-4172-8837-a8383557ced9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.881039] env[62066]: DEBUG oslo_vmware.api [None req-3baf3239-22c0-4858-82f6-d3b75bd59a7c tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Waiting for the task: (returnval){ [ 791.881039] env[62066]: value = "task-1340726" [ 791.881039] env[62066]: _type = "Task" [ 791.881039] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 791.895460] env[62066]: DEBUG oslo_vmware.api [None req-3baf3239-22c0-4858-82f6-d3b75bd59a7c tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Task: {'id': task-1340726, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.918154] env[62066]: DEBUG nova.compute.manager [req-6bb374b3-3c52-4020-a715-ea0aeb09d666 req-67ffe645-69ce-4a29-90f4-fcc609bb4044 service nova] [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] Received event network-vif-plugged-88380b24-b187-4b82-b92e-b93adeeab3eb {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 791.918154] env[62066]: DEBUG oslo_concurrency.lockutils [req-6bb374b3-3c52-4020-a715-ea0aeb09d666 req-67ffe645-69ce-4a29-90f4-fcc609bb4044 service nova] Acquiring lock "daffaf51-4c45-44aa-8fc2-4db066a09971-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 791.918154] env[62066]: DEBUG oslo_concurrency.lockutils [req-6bb374b3-3c52-4020-a715-ea0aeb09d666 req-67ffe645-69ce-4a29-90f4-fcc609bb4044 service nova] Lock "daffaf51-4c45-44aa-8fc2-4db066a09971-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 791.918154] env[62066]: DEBUG oslo_concurrency.lockutils [req-6bb374b3-3c52-4020-a715-ea0aeb09d666 req-67ffe645-69ce-4a29-90f4-fcc609bb4044 service nova] Lock "daffaf51-4c45-44aa-8fc2-4db066a09971-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 791.919117] env[62066]: DEBUG nova.compute.manager [req-6bb374b3-3c52-4020-a715-ea0aeb09d666 req-67ffe645-69ce-4a29-90f4-fcc609bb4044 service nova] [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] No waiting events found dispatching network-vif-plugged-88380b24-b187-4b82-b92e-b93adeeab3eb {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 791.919648] env[62066]: WARNING nova.compute.manager [req-6bb374b3-3c52-4020-a715-ea0aeb09d666 req-67ffe645-69ce-4a29-90f4-fcc609bb4044 service nova] [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] Received unexpected event network-vif-plugged-88380b24-b187-4b82-b92e-b93adeeab3eb for instance with vm_state building and task_state block_device_mapping. [ 792.072465] env[62066]: DEBUG nova.compute.manager [None req-c70bcb7f-78ff-424d-8121-684c45ecd12b tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 02fb3e92-5dd6-4b1e-a6e4-d60d3fc85b07] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 792.187622] env[62066]: DEBUG nova.compute.manager [None req-0188668d-25e1-420a-be60-fb5771349282 tempest-ServerAddressesTestJSON-1053293248 tempest-ServerAddressesTestJSON-1053293248-project-member] [instance: 49bc1350-0095-406b-bc68-005eb6b681a8] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 792.345494] env[62066]: DEBUG nova.compute.manager [req-e3276036-2f8d-420f-8a21-805f927bcb83 req-48bccb9d-9308-4303-9ea1-4e227953ba45 service nova] [instance: 0ebba3b2-f82d-4c1d-b01a-f75c3559c117] Received event network-vif-deleted-d9ded989-5fd8-4b73-86d4-dd8d8247d504 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 792.346342] env[62066]: INFO nova.compute.manager [req-e3276036-2f8d-420f-8a21-805f927bcb83 req-48bccb9d-9308-4303-9ea1-4e227953ba45 service nova] [instance: 0ebba3b2-f82d-4c1d-b01a-f75c3559c117] Neutron deleted interface d9ded989-5fd8-4b73-86d4-dd8d8247d504; detaching it from the instance and deleting it from the info cache [ 792.346342] env[62066]: DEBUG nova.network.neutron [req-e3276036-2f8d-420f-8a21-805f927bcb83 req-48bccb9d-9308-4303-9ea1-4e227953ba45 service nova] [instance: 0ebba3b2-f82d-4c1d-b01a-f75c3559c117] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 792.374484] env[62066]: DEBUG oslo_concurrency.lockutils [None req-86820846-d757-4de7-918c-ad5501b560db tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] Acquiring lock "refresh_cache-daffaf51-4c45-44aa-8fc2-4db066a09971" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 792.375693] env[62066]: DEBUG oslo_concurrency.lockutils [None req-86820846-d757-4de7-918c-ad5501b560db tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] Acquired lock "refresh_cache-daffaf51-4c45-44aa-8fc2-4db066a09971" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 792.375693] env[62066]: DEBUG nova.network.neutron [None req-86820846-d757-4de7-918c-ad5501b560db tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 792.391620] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32461b9b-9d48-45bd-a130-f7165e6b4264 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.396990] env[62066]: DEBUG oslo_vmware.api [None req-3baf3239-22c0-4858-82f6-d3b75bd59a7c tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Task: {'id': task-1340726, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.078171} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 792.398383] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-3baf3239-22c0-4858-82f6-d3b75bd59a7c tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] [instance: 1e1a2ede-4ad8-4600-851e-6f2046b3f919] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 792.398554] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4821d188-a743-4976-a539-77f6e46f32bd {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.404720] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c93ff1b3-e390-45d1-8bee-25a47671e9e8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.411052] env[62066]: DEBUG nova.network.neutron [None req-c70bcb7f-78ff-424d-8121-684c45ecd12b tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 02fb3e92-5dd6-4b1e-a6e4-d60d3fc85b07] Successfully created port: 16bb647a-353c-4bf8-832d-8f6e49a6ce4b {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 792.432769] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-3baf3239-22c0-4858-82f6-d3b75bd59a7c tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] [instance: 1e1a2ede-4ad8-4600-851e-6f2046b3f919] Reconfiguring VM instance instance-0000003b to attach disk [datastore2] 1e1a2ede-4ad8-4600-851e-6f2046b3f919/1e1a2ede-4ad8-4600-851e-6f2046b3f919.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 792.434759] env[62066]: DEBUG nova.compute.manager [None req-86820846-d757-4de7-918c-ad5501b560db tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 792.435292] env[62066]: DEBUG nova.virt.hardware [None req-86820846-d757-4de7-918c-ad5501b560db tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-23T13:40:41Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 792.435583] env[62066]: DEBUG nova.virt.hardware [None req-86820846-d757-4de7-918c-ad5501b560db tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 792.435656] env[62066]: DEBUG nova.virt.hardware [None req-86820846-d757-4de7-918c-ad5501b560db tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 792.436573] env[62066]: DEBUG nova.virt.hardware [None req-86820846-d757-4de7-918c-ad5501b560db tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 792.436573] env[62066]: DEBUG nova.virt.hardware [None req-86820846-d757-4de7-918c-ad5501b560db tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 792.436573] env[62066]: DEBUG nova.virt.hardware [None req-86820846-d757-4de7-918c-ad5501b560db tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 792.436573] env[62066]: DEBUG nova.virt.hardware [None req-86820846-d757-4de7-918c-ad5501b560db tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 792.436573] env[62066]: DEBUG nova.virt.hardware [None req-86820846-d757-4de7-918c-ad5501b560db tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 792.436732] env[62066]: DEBUG nova.virt.hardware [None req-86820846-d757-4de7-918c-ad5501b560db tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 792.436765] env[62066]: DEBUG nova.virt.hardware [None req-86820846-d757-4de7-918c-ad5501b560db tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 792.437134] env[62066]: DEBUG nova.virt.hardware [None req-86820846-d757-4de7-918c-ad5501b560db tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 792.437214] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-05415d34-bc0b-4404-97ef-17abf6c4e917 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.476297] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd080eef-c420-4e4e-98b6-0cee6bef77d7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.480662] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e46d41f3-06b5-4e63-822a-9d5f46b0d36f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.486712] env[62066]: DEBUG oslo_vmware.api [None req-3baf3239-22c0-4858-82f6-d3b75bd59a7c tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Waiting for the task: (returnval){ [ 792.486712] env[62066]: value = "task-1340727" [ 792.486712] env[62066]: _type = "Task" [ 792.486712] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.497074] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ce7979b-5ce0-4ca2-b01e-877e63a5b790 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.517439] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1e5d462-29ba-4ca4-a48f-5412521a3409 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.523234] env[62066]: DEBUG oslo_vmware.api [None req-3baf3239-22c0-4858-82f6-d3b75bd59a7c tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Task: {'id': task-1340727, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.531828] env[62066]: DEBUG nova.compute.provider_tree [None req-816774ed-d63b-4dd8-94be-e4c70c8f10e7 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 792.711133] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0188668d-25e1-420a-be60-fb5771349282 tempest-ServerAddressesTestJSON-1053293248 tempest-ServerAddressesTestJSON-1053293248-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 792.730705] env[62066]: DEBUG nova.network.neutron [-] [instance: 0ebba3b2-f82d-4c1d-b01a-f75c3559c117] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 792.763462] env[62066]: DEBUG nova.network.neutron [-] [instance: 98fb270c-f3f8-4375-8b5c-c2279305d476] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 792.856126] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5aafafc2-3c38-4ab1-b0be-c96941a7f2c3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.869876] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ee54ac1-aa6b-4f88-b641-8a61e3720256 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.898233] env[62066]: DEBUG nova.compute.manager [req-e3276036-2f8d-420f-8a21-805f927bcb83 req-48bccb9d-9308-4303-9ea1-4e227953ba45 service nova] [instance: 0ebba3b2-f82d-4c1d-b01a-f75c3559c117] Detach interface failed, port_id=d9ded989-5fd8-4b73-86d4-dd8d8247d504, reason: Instance 0ebba3b2-f82d-4c1d-b01a-f75c3559c117 could not be found. {{(pid=62066) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 792.934098] env[62066]: DEBUG nova.network.neutron [None req-86820846-d757-4de7-918c-ad5501b560db tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 793.003364] env[62066]: DEBUG oslo_vmware.api [None req-3baf3239-22c0-4858-82f6-d3b75bd59a7c tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Task: {'id': task-1340727, 'name': ReconfigVM_Task, 'duration_secs': 0.364719} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.003487] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-3baf3239-22c0-4858-82f6-d3b75bd59a7c tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] [instance: 1e1a2ede-4ad8-4600-851e-6f2046b3f919] Reconfigured VM instance instance-0000003b to attach disk [datastore2] 1e1a2ede-4ad8-4600-851e-6f2046b3f919/1e1a2ede-4ad8-4600-851e-6f2046b3f919.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 793.004454] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-840e0f35-23bd-498f-bcac-d50e1a1a7450 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.013169] env[62066]: DEBUG oslo_vmware.api [None req-3baf3239-22c0-4858-82f6-d3b75bd59a7c tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Waiting for the task: (returnval){ [ 793.013169] env[62066]: value = "task-1340728" [ 793.013169] env[62066]: _type = "Task" [ 793.013169] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.025159] env[62066]: DEBUG oslo_vmware.api [None req-3baf3239-22c0-4858-82f6-d3b75bd59a7c tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Task: {'id': task-1340728, 'name': Rename_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.037866] env[62066]: DEBUG nova.scheduler.client.report [None req-816774ed-d63b-4dd8-94be-e4c70c8f10e7 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 793.090344] env[62066]: DEBUG nova.compute.manager [None req-c70bcb7f-78ff-424d-8121-684c45ecd12b tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 02fb3e92-5dd6-4b1e-a6e4-d60d3fc85b07] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 793.107886] env[62066]: DEBUG oslo_concurrency.lockutils [None req-acef35f4-b906-4efb-ae97-e8cdc80e8e5a tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Acquiring lock "5e96def2-0cbd-4bd9-93f4-6a365a0142b0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 793.107886] env[62066]: DEBUG oslo_concurrency.lockutils [None req-acef35f4-b906-4efb-ae97-e8cdc80e8e5a tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Lock "5e96def2-0cbd-4bd9-93f4-6a365a0142b0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 793.121557] env[62066]: DEBUG nova.virt.hardware [None req-c70bcb7f-78ff-424d-8121-684c45ecd12b tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-23T13:40:41Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-23T13:40:23Z,direct_url=,disk_format='vmdk',id=50ff584c-3b50-4395-af07-3e66769bc9f7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='eb52f7069a374c61ae946f052007c6d9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-23T13:40:24Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 793.121962] env[62066]: DEBUG nova.virt.hardware [None req-c70bcb7f-78ff-424d-8121-684c45ecd12b tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 793.122026] env[62066]: DEBUG nova.virt.hardware [None req-c70bcb7f-78ff-424d-8121-684c45ecd12b tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 793.122270] env[62066]: DEBUG nova.virt.hardware [None req-c70bcb7f-78ff-424d-8121-684c45ecd12b tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 793.122324] env[62066]: DEBUG nova.virt.hardware [None req-c70bcb7f-78ff-424d-8121-684c45ecd12b tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 793.122469] env[62066]: DEBUG nova.virt.hardware [None req-c70bcb7f-78ff-424d-8121-684c45ecd12b tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 793.122708] env[62066]: DEBUG nova.virt.hardware [None req-c70bcb7f-78ff-424d-8121-684c45ecd12b tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 793.122888] env[62066]: DEBUG nova.virt.hardware [None req-c70bcb7f-78ff-424d-8121-684c45ecd12b tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 793.123588] env[62066]: DEBUG nova.virt.hardware [None req-c70bcb7f-78ff-424d-8121-684c45ecd12b tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 793.123809] env[62066]: DEBUG nova.virt.hardware [None req-c70bcb7f-78ff-424d-8121-684c45ecd12b tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 793.123991] env[62066]: DEBUG nova.virt.hardware [None req-c70bcb7f-78ff-424d-8121-684c45ecd12b tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 793.125197] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af37346e-5d7d-4b36-a6df-55a3e3034483 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.139104] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b91010a-1ffa-4792-aad4-2e32fe1ffb59 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.143873] env[62066]: DEBUG nova.network.neutron [None req-86820846-d757-4de7-918c-ad5501b560db tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] Updating instance_info_cache with network_info: [{"id": "88380b24-b187-4b82-b92e-b93adeeab3eb", "address": "fa:16:3e:5c:2a:4b", "network": {"id": "b08881d1-c195-4165-9400-8c2012516c9e", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-1637068010-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0b8c246df0e8404ca3743c7207cfc808", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e1c9bb98-73a9-48eb-856e-a541afe9b07b", "external-id": "nsx-vlan-transportzone-755", "segmentation_id": 755, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap88380b24-b1", "ovs_interfaceid": "88380b24-b187-4b82-b92e-b93adeeab3eb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 793.236664] env[62066]: INFO nova.compute.manager [-] [instance: 0ebba3b2-f82d-4c1d-b01a-f75c3559c117] Took 1.84 seconds to deallocate network for instance. [ 793.272019] env[62066]: INFO nova.compute.manager [-] [instance: 98fb270c-f3f8-4375-8b5c-c2279305d476] Took 1.78 seconds to deallocate network for instance. [ 793.523917] env[62066]: DEBUG oslo_vmware.api [None req-3baf3239-22c0-4858-82f6-d3b75bd59a7c tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Task: {'id': task-1340728, 'name': Rename_Task, 'duration_secs': 0.164533} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.524222] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-3baf3239-22c0-4858-82f6-d3b75bd59a7c tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] [instance: 1e1a2ede-4ad8-4600-851e-6f2046b3f919] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 793.524457] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7c1f68a8-5a36-4a05-9bf6-868d74ee8cbb {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.530848] env[62066]: DEBUG oslo_vmware.api [None req-3baf3239-22c0-4858-82f6-d3b75bd59a7c tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Waiting for the task: (returnval){ [ 793.530848] env[62066]: value = "task-1340729" [ 793.530848] env[62066]: _type = "Task" [ 793.530848] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.540043] env[62066]: DEBUG oslo_vmware.api [None req-3baf3239-22c0-4858-82f6-d3b75bd59a7c tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Task: {'id': task-1340729, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.544107] env[62066]: DEBUG oslo_concurrency.lockutils [None req-816774ed-d63b-4dd8-94be-e4c70c8f10e7 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.481s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 793.544825] env[62066]: DEBUG nova.compute.manager [None req-816774ed-d63b-4dd8-94be-e4c70c8f10e7 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: 459f432b-e817-405f-9c3e-dfa5d957788c] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 793.548869] env[62066]: DEBUG oslo_concurrency.lockutils [None req-febd07a3-de81-4924-888e-f9c7d8ba2329 tempest-ServerMetadataNegativeTestJSON-1549470887 tempest-ServerMetadataNegativeTestJSON-1549470887-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.321s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 793.550968] env[62066]: INFO nova.compute.claims [None req-febd07a3-de81-4924-888e-f9c7d8ba2329 tempest-ServerMetadataNegativeTestJSON-1549470887 tempest-ServerMetadataNegativeTestJSON-1549470887-project-member] [instance: 31d47299-83a3-4f27-aeb7-95c4cd36c5ac] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 793.648509] env[62066]: DEBUG oslo_concurrency.lockutils [None req-86820846-d757-4de7-918c-ad5501b560db tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] Releasing lock "refresh_cache-daffaf51-4c45-44aa-8fc2-4db066a09971" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 793.648759] env[62066]: DEBUG nova.compute.manager [None req-86820846-d757-4de7-918c-ad5501b560db tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] Instance network_info: |[{"id": "88380b24-b187-4b82-b92e-b93adeeab3eb", "address": "fa:16:3e:5c:2a:4b", "network": {"id": "b08881d1-c195-4165-9400-8c2012516c9e", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-1637068010-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0b8c246df0e8404ca3743c7207cfc808", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e1c9bb98-73a9-48eb-856e-a541afe9b07b", "external-id": "nsx-vlan-transportzone-755", "segmentation_id": 755, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap88380b24-b1", "ovs_interfaceid": "88380b24-b187-4b82-b92e-b93adeeab3eb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 793.649481] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-86820846-d757-4de7-918c-ad5501b560db tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:5c:2a:4b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e1c9bb98-73a9-48eb-856e-a541afe9b07b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '88380b24-b187-4b82-b92e-b93adeeab3eb', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 793.657151] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-86820846-d757-4de7-918c-ad5501b560db tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] Creating folder: Project (0b8c246df0e8404ca3743c7207cfc808). Parent ref: group-v285980. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 793.657786] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-82e2675a-d6c1-44e7-8577-0a5274447854 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.671029] env[62066]: WARNING suds.client [-] Web service reported a SOAP processing fault using an unexpected HTTP status code 200. Reporting as an internal server error. [ 793.671198] env[62066]: DEBUG oslo_vmware.api [-] Fault list: [DuplicateName] {{(pid=62066) _invoke_api /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:337}} [ 793.671510] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-86820846-d757-4de7-918c-ad5501b560db tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] Folder already exists: Project (0b8c246df0e8404ca3743c7207cfc808). Parent ref: group-v285980. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1609}} [ 793.671700] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-86820846-d757-4de7-918c-ad5501b560db tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] Creating folder: Instances. Parent ref: group-v285993. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 793.671926] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-854924b6-244d-4a69-bfff-6d6f73ca66fd {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.681052] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-86820846-d757-4de7-918c-ad5501b560db tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] Created folder: Instances in parent group-v285993. [ 793.681052] env[62066]: DEBUG oslo.service.loopingcall [None req-86820846-d757-4de7-918c-ad5501b560db tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 793.681052] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 793.681052] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4c28bd55-60a0-4b78-9a2c-4b29ac6752bc {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.699150] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 793.699150] env[62066]: value = "task-1340732" [ 793.699150] env[62066]: _type = "Task" [ 793.699150] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.706519] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1340732, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.743322] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ba3a3138-4a4e-4a10-bbee-93dd3a372fa2 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 793.778677] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e5dc4f2e-1d02-4ba9-acb0-6628ba7125f6 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 793.950430] env[62066]: DEBUG nova.compute.manager [req-f28622bf-f3a4-4f8b-aa2d-4d55b71de00a req-0f39c0da-97e4-453d-8eb2-1fd361f31664 service nova] [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] Received event network-changed-88380b24-b187-4b82-b92e-b93adeeab3eb {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 793.950655] env[62066]: DEBUG nova.compute.manager [req-f28622bf-f3a4-4f8b-aa2d-4d55b71de00a req-0f39c0da-97e4-453d-8eb2-1fd361f31664 service nova] [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] Refreshing instance network info cache due to event network-changed-88380b24-b187-4b82-b92e-b93adeeab3eb. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 793.950891] env[62066]: DEBUG oslo_concurrency.lockutils [req-f28622bf-f3a4-4f8b-aa2d-4d55b71de00a req-0f39c0da-97e4-453d-8eb2-1fd361f31664 service nova] Acquiring lock "refresh_cache-daffaf51-4c45-44aa-8fc2-4db066a09971" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 793.951048] env[62066]: DEBUG oslo_concurrency.lockutils [req-f28622bf-f3a4-4f8b-aa2d-4d55b71de00a req-0f39c0da-97e4-453d-8eb2-1fd361f31664 service nova] Acquired lock "refresh_cache-daffaf51-4c45-44aa-8fc2-4db066a09971" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 793.951218] env[62066]: DEBUG nova.network.neutron [req-f28622bf-f3a4-4f8b-aa2d-4d55b71de00a req-0f39c0da-97e4-453d-8eb2-1fd361f31664 service nova] [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] Refreshing network info cache for port 88380b24-b187-4b82-b92e-b93adeeab3eb {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 794.015330] env[62066]: DEBUG nova.network.neutron [None req-c70bcb7f-78ff-424d-8121-684c45ecd12b tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 02fb3e92-5dd6-4b1e-a6e4-d60d3fc85b07] Successfully updated port: 16bb647a-353c-4bf8-832d-8f6e49a6ce4b {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 794.040919] env[62066]: DEBUG oslo_vmware.api [None req-3baf3239-22c0-4858-82f6-d3b75bd59a7c tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Task: {'id': task-1340729, 'name': PowerOnVM_Task, 'duration_secs': 0.47103} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.041219] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-3baf3239-22c0-4858-82f6-d3b75bd59a7c tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] [instance: 1e1a2ede-4ad8-4600-851e-6f2046b3f919] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 794.041440] env[62066]: INFO nova.compute.manager [None req-3baf3239-22c0-4858-82f6-d3b75bd59a7c tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] [instance: 1e1a2ede-4ad8-4600-851e-6f2046b3f919] Took 7.62 seconds to spawn the instance on the hypervisor. [ 794.041640] env[62066]: DEBUG nova.compute.manager [None req-3baf3239-22c0-4858-82f6-d3b75bd59a7c tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] [instance: 1e1a2ede-4ad8-4600-851e-6f2046b3f919] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 794.042420] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0164007-e303-4168-aecc-566903676c91 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.055745] env[62066]: DEBUG nova.compute.utils [None req-816774ed-d63b-4dd8-94be-e4c70c8f10e7 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 794.058660] env[62066]: DEBUG nova.compute.manager [None req-816774ed-d63b-4dd8-94be-e4c70c8f10e7 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: 459f432b-e817-405f-9c3e-dfa5d957788c] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 794.058915] env[62066]: DEBUG nova.network.neutron [None req-816774ed-d63b-4dd8-94be-e4c70c8f10e7 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: 459f432b-e817-405f-9c3e-dfa5d957788c] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 794.098534] env[62066]: DEBUG nova.policy [None req-816774ed-d63b-4dd8-94be-e4c70c8f10e7 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '747026d272d4467bbf14a0c30589c3b7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0d4e33a99c7741fb8cdd97f4ec5dbbd8', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 794.211638] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1340732, 'name': CreateVM_Task, 'duration_secs': 0.303464} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.211834] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 794.212993] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-86820846-d757-4de7-918c-ad5501b560db tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] Block device information present: {'root_device_name': '/dev/sda', 'image': [], 'ephemerals': [], 'block_device_mapping': [{'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-285999', 'volume_id': '26ceca45-1cae-48c9-a567-e7cc4f9970c1', 'name': 'volume-26ceca45-1cae-48c9-a567-e7cc4f9970c1', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'daffaf51-4c45-44aa-8fc2-4db066a09971', 'attached_at': '', 'detached_at': '', 'volume_id': '26ceca45-1cae-48c9-a567-e7cc4f9970c1', 'serial': '26ceca45-1cae-48c9-a567-e7cc4f9970c1'}, 'delete_on_termination': True, 'attachment_id': '0334076e-6b3a-4c6b-9474-ed73d3ceeb0c', 'boot_index': 0, 'device_type': None, 'guest_format': None, 'disk_bus': None, 'mount_device': '/dev/sda', 'volume_type': None}], 'swap': None} {{(pid=62066) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 794.212993] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-86820846-d757-4de7-918c-ad5501b560db tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] Root volume attach. Driver type: vmdk {{(pid=62066) attach_root_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:661}} [ 794.214945] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e90ad898-20af-4cd9-b097-83f2bb5bc635 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.221030] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d9764a6-3aa2-40ce-a914-ad27547a1cca {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.227176] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fd33789-6cdc-4f25-aa56-deb2eec0e338 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.233868] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.RelocateVM_Task with opID=oslo.vmware-d916ccc8-3b50-457a-868b-5ba08bd1e51b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.240180] env[62066]: DEBUG oslo_vmware.api [None req-86820846-d757-4de7-918c-ad5501b560db tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] Waiting for the task: (returnval){ [ 794.240180] env[62066]: value = "task-1340733" [ 794.240180] env[62066]: _type = "Task" [ 794.240180] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.252104] env[62066]: DEBUG oslo_vmware.api [None req-86820846-d757-4de7-918c-ad5501b560db tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] Task: {'id': task-1340733, 'name': RelocateVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.374377] env[62066]: DEBUG nova.compute.manager [req-719cf617-9533-45e7-8f9f-5e63b1082f6f req-9d67d816-eb33-49ea-9cbb-b807ce98f310 service nova] [instance: 98fb270c-f3f8-4375-8b5c-c2279305d476] Received event network-vif-deleted-78f327de-1b83-4e33-a2b7-8d1269a96df1 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 794.374590] env[62066]: DEBUG nova.compute.manager [req-719cf617-9533-45e7-8f9f-5e63b1082f6f req-9d67d816-eb33-49ea-9cbb-b807ce98f310 service nova] [instance: 02fb3e92-5dd6-4b1e-a6e4-d60d3fc85b07] Received event network-vif-plugged-16bb647a-353c-4bf8-832d-8f6e49a6ce4b {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 794.374856] env[62066]: DEBUG oslo_concurrency.lockutils [req-719cf617-9533-45e7-8f9f-5e63b1082f6f req-9d67d816-eb33-49ea-9cbb-b807ce98f310 service nova] Acquiring lock "02fb3e92-5dd6-4b1e-a6e4-d60d3fc85b07-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 794.375063] env[62066]: DEBUG oslo_concurrency.lockutils [req-719cf617-9533-45e7-8f9f-5e63b1082f6f req-9d67d816-eb33-49ea-9cbb-b807ce98f310 service nova] Lock "02fb3e92-5dd6-4b1e-a6e4-d60d3fc85b07-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 794.375267] env[62066]: DEBUG oslo_concurrency.lockutils [req-719cf617-9533-45e7-8f9f-5e63b1082f6f req-9d67d816-eb33-49ea-9cbb-b807ce98f310 service nova] Lock "02fb3e92-5dd6-4b1e-a6e4-d60d3fc85b07-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 794.375459] env[62066]: DEBUG nova.compute.manager [req-719cf617-9533-45e7-8f9f-5e63b1082f6f req-9d67d816-eb33-49ea-9cbb-b807ce98f310 service nova] [instance: 02fb3e92-5dd6-4b1e-a6e4-d60d3fc85b07] No waiting events found dispatching network-vif-plugged-16bb647a-353c-4bf8-832d-8f6e49a6ce4b {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 794.375690] env[62066]: WARNING nova.compute.manager [req-719cf617-9533-45e7-8f9f-5e63b1082f6f req-9d67d816-eb33-49ea-9cbb-b807ce98f310 service nova] [instance: 02fb3e92-5dd6-4b1e-a6e4-d60d3fc85b07] Received unexpected event network-vif-plugged-16bb647a-353c-4bf8-832d-8f6e49a6ce4b for instance with vm_state building and task_state spawning. [ 794.375898] env[62066]: DEBUG nova.compute.manager [req-719cf617-9533-45e7-8f9f-5e63b1082f6f req-9d67d816-eb33-49ea-9cbb-b807ce98f310 service nova] [instance: 02fb3e92-5dd6-4b1e-a6e4-d60d3fc85b07] Received event network-changed-16bb647a-353c-4bf8-832d-8f6e49a6ce4b {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 794.376102] env[62066]: DEBUG nova.compute.manager [req-719cf617-9533-45e7-8f9f-5e63b1082f6f req-9d67d816-eb33-49ea-9cbb-b807ce98f310 service nova] [instance: 02fb3e92-5dd6-4b1e-a6e4-d60d3fc85b07] Refreshing instance network info cache due to event network-changed-16bb647a-353c-4bf8-832d-8f6e49a6ce4b. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 794.376345] env[62066]: DEBUG oslo_concurrency.lockutils [req-719cf617-9533-45e7-8f9f-5e63b1082f6f req-9d67d816-eb33-49ea-9cbb-b807ce98f310 service nova] Acquiring lock "refresh_cache-02fb3e92-5dd6-4b1e-a6e4-d60d3fc85b07" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 794.376513] env[62066]: DEBUG oslo_concurrency.lockutils [req-719cf617-9533-45e7-8f9f-5e63b1082f6f req-9d67d816-eb33-49ea-9cbb-b807ce98f310 service nova] Acquired lock "refresh_cache-02fb3e92-5dd6-4b1e-a6e4-d60d3fc85b07" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 794.376777] env[62066]: DEBUG nova.network.neutron [req-719cf617-9533-45e7-8f9f-5e63b1082f6f req-9d67d816-eb33-49ea-9cbb-b807ce98f310 service nova] [instance: 02fb3e92-5dd6-4b1e-a6e4-d60d3fc85b07] Refreshing network info cache for port 16bb647a-353c-4bf8-832d-8f6e49a6ce4b {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 794.391186] env[62066]: DEBUG nova.network.neutron [None req-816774ed-d63b-4dd8-94be-e4c70c8f10e7 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: 459f432b-e817-405f-9c3e-dfa5d957788c] Successfully created port: cb74e58f-b481-4a0f-861b-96381e2d52dc {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 794.518228] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c70bcb7f-78ff-424d-8121-684c45ecd12b tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Acquiring lock "refresh_cache-02fb3e92-5dd6-4b1e-a6e4-d60d3fc85b07" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 794.562007] env[62066]: INFO nova.compute.manager [None req-3baf3239-22c0-4858-82f6-d3b75bd59a7c tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] [instance: 1e1a2ede-4ad8-4600-851e-6f2046b3f919] Took 27.32 seconds to build instance. [ 794.563044] env[62066]: DEBUG nova.compute.manager [None req-816774ed-d63b-4dd8-94be-e4c70c8f10e7 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: 459f432b-e817-405f-9c3e-dfa5d957788c] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 794.751961] env[62066]: DEBUG oslo_vmware.api [None req-86820846-d757-4de7-918c-ad5501b560db tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] Task: {'id': task-1340733, 'name': RelocateVM_Task} progress is 19%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.762822] env[62066]: DEBUG nova.network.neutron [req-f28622bf-f3a4-4f8b-aa2d-4d55b71de00a req-0f39c0da-97e4-453d-8eb2-1fd361f31664 service nova] [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] Updated VIF entry in instance network info cache for port 88380b24-b187-4b82-b92e-b93adeeab3eb. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 794.763194] env[62066]: DEBUG nova.network.neutron [req-f28622bf-f3a4-4f8b-aa2d-4d55b71de00a req-0f39c0da-97e4-453d-8eb2-1fd361f31664 service nova] [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] Updating instance_info_cache with network_info: [{"id": "88380b24-b187-4b82-b92e-b93adeeab3eb", "address": "fa:16:3e:5c:2a:4b", "network": {"id": "b08881d1-c195-4165-9400-8c2012516c9e", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-1637068010-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0b8c246df0e8404ca3743c7207cfc808", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e1c9bb98-73a9-48eb-856e-a541afe9b07b", "external-id": "nsx-vlan-transportzone-755", "segmentation_id": 755, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap88380b24-b1", "ovs_interfaceid": "88380b24-b187-4b82-b92e-b93adeeab3eb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 794.828533] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-569584e0-a7ad-40e2-af60-36b1c0ed68ec {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.837215] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c81f4f3e-c324-44f3-bbef-fee7f77300c5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.871469] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1007a9bb-23dd-4c0f-8b39-c39b8d06b7eb {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.881302] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2360fd6-62d0-4b18-aec9-809f4a281151 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.897453] env[62066]: DEBUG nova.compute.provider_tree [None req-febd07a3-de81-4924-888e-f9c7d8ba2329 tempest-ServerMetadataNegativeTestJSON-1549470887 tempest-ServerMetadataNegativeTestJSON-1549470887-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 794.913168] env[62066]: DEBUG nova.network.neutron [req-719cf617-9533-45e7-8f9f-5e63b1082f6f req-9d67d816-eb33-49ea-9cbb-b807ce98f310 service nova] [instance: 02fb3e92-5dd6-4b1e-a6e4-d60d3fc85b07] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 794.990916] env[62066]: DEBUG nova.network.neutron [req-719cf617-9533-45e7-8f9f-5e63b1082f6f req-9d67d816-eb33-49ea-9cbb-b807ce98f310 service nova] [instance: 02fb3e92-5dd6-4b1e-a6e4-d60d3fc85b07] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 795.063501] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3baf3239-22c0-4858-82f6-d3b75bd59a7c tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Lock "1e1a2ede-4ad8-4600-851e-6f2046b3f919" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 150.300s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 795.252592] env[62066]: DEBUG oslo_vmware.api [None req-86820846-d757-4de7-918c-ad5501b560db tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] Task: {'id': task-1340733, 'name': RelocateVM_Task, 'duration_secs': 0.928302} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 795.252886] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-86820846-d757-4de7-918c-ad5501b560db tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] Volume attach. Driver type: vmdk {{(pid=62066) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 795.253107] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-86820846-d757-4de7-918c-ad5501b560db tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-285999', 'volume_id': '26ceca45-1cae-48c9-a567-e7cc4f9970c1', 'name': 'volume-26ceca45-1cae-48c9-a567-e7cc4f9970c1', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'daffaf51-4c45-44aa-8fc2-4db066a09971', 'attached_at': '', 'detached_at': '', 'volume_id': '26ceca45-1cae-48c9-a567-e7cc4f9970c1', 'serial': '26ceca45-1cae-48c9-a567-e7cc4f9970c1'} {{(pid=62066) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 795.253993] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-770cb6ca-9b68-4135-a49e-19e95c8c02ba {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.269183] env[62066]: DEBUG oslo_concurrency.lockutils [req-f28622bf-f3a4-4f8b-aa2d-4d55b71de00a req-0f39c0da-97e4-453d-8eb2-1fd361f31664 service nova] Releasing lock "refresh_cache-daffaf51-4c45-44aa-8fc2-4db066a09971" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 795.270318] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12624b71-934c-42fd-8ba5-3262098c636e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.293137] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-86820846-d757-4de7-918c-ad5501b560db tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] Reconfiguring VM instance instance-0000003c to attach disk [datastore2] volume-26ceca45-1cae-48c9-a567-e7cc4f9970c1/volume-26ceca45-1cae-48c9-a567-e7cc4f9970c1.vmdk or device None with type thin {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 795.293340] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f24900b1-7d0b-4ad5-928d-b8210547916e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.312993] env[62066]: DEBUG oslo_vmware.api [None req-86820846-d757-4de7-918c-ad5501b560db tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] Waiting for the task: (returnval){ [ 795.312993] env[62066]: value = "task-1340734" [ 795.312993] env[62066]: _type = "Task" [ 795.312993] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 795.320866] env[62066]: DEBUG oslo_vmware.api [None req-86820846-d757-4de7-918c-ad5501b560db tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] Task: {'id': task-1340734, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.401619] env[62066]: DEBUG nova.scheduler.client.report [None req-febd07a3-de81-4924-888e-f9c7d8ba2329 tempest-ServerMetadataNegativeTestJSON-1549470887 tempest-ServerMetadataNegativeTestJSON-1549470887-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 795.493450] env[62066]: DEBUG oslo_concurrency.lockutils [req-719cf617-9533-45e7-8f9f-5e63b1082f6f req-9d67d816-eb33-49ea-9cbb-b807ce98f310 service nova] Releasing lock "refresh_cache-02fb3e92-5dd6-4b1e-a6e4-d60d3fc85b07" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 795.493710] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c70bcb7f-78ff-424d-8121-684c45ecd12b tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Acquired lock "refresh_cache-02fb3e92-5dd6-4b1e-a6e4-d60d3fc85b07" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 795.493888] env[62066]: DEBUG nova.network.neutron [None req-c70bcb7f-78ff-424d-8121-684c45ecd12b tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 02fb3e92-5dd6-4b1e-a6e4-d60d3fc85b07] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 795.565910] env[62066]: DEBUG nova.compute.manager [None req-9e052131-3359-4888-9d78-89440b639dcc tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] [instance: 8ff6fbb9-c90f-498d-9a85-d220a8c2f794] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 795.571546] env[62066]: DEBUG nova.compute.manager [None req-816774ed-d63b-4dd8-94be-e4c70c8f10e7 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: 459f432b-e817-405f-9c3e-dfa5d957788c] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 795.598012] env[62066]: DEBUG nova.virt.hardware [None req-816774ed-d63b-4dd8-94be-e4c70c8f10e7 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-23T13:40:41Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-23T13:40:23Z,direct_url=,disk_format='vmdk',id=50ff584c-3b50-4395-af07-3e66769bc9f7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='eb52f7069a374c61ae946f052007c6d9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-23T13:40:24Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 795.598012] env[62066]: DEBUG nova.virt.hardware [None req-816774ed-d63b-4dd8-94be-e4c70c8f10e7 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 795.598012] env[62066]: DEBUG nova.virt.hardware [None req-816774ed-d63b-4dd8-94be-e4c70c8f10e7 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 795.598186] env[62066]: DEBUG nova.virt.hardware [None req-816774ed-d63b-4dd8-94be-e4c70c8f10e7 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 795.598186] env[62066]: DEBUG nova.virt.hardware [None req-816774ed-d63b-4dd8-94be-e4c70c8f10e7 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 795.598378] env[62066]: DEBUG nova.virt.hardware [None req-816774ed-d63b-4dd8-94be-e4c70c8f10e7 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 795.598587] env[62066]: DEBUG nova.virt.hardware [None req-816774ed-d63b-4dd8-94be-e4c70c8f10e7 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 795.598750] env[62066]: DEBUG nova.virt.hardware [None req-816774ed-d63b-4dd8-94be-e4c70c8f10e7 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 795.598918] env[62066]: DEBUG nova.virt.hardware [None req-816774ed-d63b-4dd8-94be-e4c70c8f10e7 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 795.599162] env[62066]: DEBUG nova.virt.hardware [None req-816774ed-d63b-4dd8-94be-e4c70c8f10e7 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 795.599368] env[62066]: DEBUG nova.virt.hardware [None req-816774ed-d63b-4dd8-94be-e4c70c8f10e7 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 795.600246] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edb9706a-4c86-41ab-91ed-9c3c947c2af2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.610017] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ba8f8de-51ca-4094-adad-d1c68bf8ee03 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.823749] env[62066]: DEBUG oslo_vmware.api [None req-86820846-d757-4de7-918c-ad5501b560db tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] Task: {'id': task-1340734, 'name': ReconfigVM_Task, 'duration_secs': 0.257187} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 795.824473] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-86820846-d757-4de7-918c-ad5501b560db tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] Reconfigured VM instance instance-0000003c to attach disk [datastore2] volume-26ceca45-1cae-48c9-a567-e7cc4f9970c1/volume-26ceca45-1cae-48c9-a567-e7cc4f9970c1.vmdk or device None with type thin {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 795.829816] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0c7f1741-2c4c-4812-85c0-aede38f2a888 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.844507] env[62066]: DEBUG oslo_vmware.api [None req-86820846-d757-4de7-918c-ad5501b560db tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] Waiting for the task: (returnval){ [ 795.844507] env[62066]: value = "task-1340735" [ 795.844507] env[62066]: _type = "Task" [ 795.844507] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 795.852617] env[62066]: DEBUG oslo_vmware.api [None req-86820846-d757-4de7-918c-ad5501b560db tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] Task: {'id': task-1340735, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.907015] env[62066]: DEBUG oslo_concurrency.lockutils [None req-febd07a3-de81-4924-888e-f9c7d8ba2329 tempest-ServerMetadataNegativeTestJSON-1549470887 tempest-ServerMetadataNegativeTestJSON-1549470887-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.358s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 795.907574] env[62066]: DEBUG nova.compute.manager [None req-febd07a3-de81-4924-888e-f9c7d8ba2329 tempest-ServerMetadataNegativeTestJSON-1549470887 tempest-ServerMetadataNegativeTestJSON-1549470887-project-member] [instance: 31d47299-83a3-4f27-aeb7-95c4cd36c5ac] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 795.910663] env[62066]: DEBUG oslo_concurrency.lockutils [None req-df48af02-4c4b-4364-8d38-efc343bf8241 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.965s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 795.912288] env[62066]: INFO nova.compute.claims [None req-df48af02-4c4b-4364-8d38-efc343bf8241 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 49f46244-34fa-48a1-95a2-8e95850f345d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 796.039475] env[62066]: DEBUG nova.network.neutron [None req-c70bcb7f-78ff-424d-8121-684c45ecd12b tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 02fb3e92-5dd6-4b1e-a6e4-d60d3fc85b07] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 796.067450] env[62066]: DEBUG nova.network.neutron [None req-816774ed-d63b-4dd8-94be-e4c70c8f10e7 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: 459f432b-e817-405f-9c3e-dfa5d957788c] Successfully updated port: cb74e58f-b481-4a0f-861b-96381e2d52dc {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 796.090315] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9e052131-3359-4888-9d78-89440b639dcc tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 796.192789] env[62066]: DEBUG nova.network.neutron [None req-c70bcb7f-78ff-424d-8121-684c45ecd12b tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 02fb3e92-5dd6-4b1e-a6e4-d60d3fc85b07] Updating instance_info_cache with network_info: [{"id": "16bb647a-353c-4bf8-832d-8f6e49a6ce4b", "address": "fa:16:3e:47:44:80", "network": {"id": "9f3396ce-ee48-4eaf-8cc7-3e1711f37537", "bridge": "br-int", "label": "tempest-ServersTestJSON-2129775093-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6497ab02f327476d8ff81c2ecc0371e5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7ab8d568-adb0-4f3b-b6cc-68413e6546ae", "external-id": "nsx-vlan-transportzone-86", "segmentation_id": 86, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap16bb647a-35", "ovs_interfaceid": "16bb647a-353c-4bf8-832d-8f6e49a6ce4b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 796.355268] env[62066]: DEBUG oslo_vmware.api [None req-86820846-d757-4de7-918c-ad5501b560db tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] Task: {'id': task-1340735, 'name': ReconfigVM_Task, 'duration_secs': 0.111552} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 796.355514] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-86820846-d757-4de7-918c-ad5501b560db tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-285999', 'volume_id': '26ceca45-1cae-48c9-a567-e7cc4f9970c1', 'name': 'volume-26ceca45-1cae-48c9-a567-e7cc4f9970c1', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'daffaf51-4c45-44aa-8fc2-4db066a09971', 'attached_at': '', 'detached_at': '', 'volume_id': '26ceca45-1cae-48c9-a567-e7cc4f9970c1', 'serial': '26ceca45-1cae-48c9-a567-e7cc4f9970c1'} {{(pid=62066) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 796.356066] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ee8bf96a-1b21-4873-a78a-0af02895ad78 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.362284] env[62066]: DEBUG oslo_vmware.api [None req-86820846-d757-4de7-918c-ad5501b560db tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] Waiting for the task: (returnval){ [ 796.362284] env[62066]: value = "task-1340736" [ 796.362284] env[62066]: _type = "Task" [ 796.362284] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.372298] env[62066]: DEBUG oslo_vmware.api [None req-86820846-d757-4de7-918c-ad5501b560db tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] Task: {'id': task-1340736, 'name': Rename_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.416914] env[62066]: DEBUG nova.compute.utils [None req-febd07a3-de81-4924-888e-f9c7d8ba2329 tempest-ServerMetadataNegativeTestJSON-1549470887 tempest-ServerMetadataNegativeTestJSON-1549470887-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 796.421398] env[62066]: DEBUG nova.compute.manager [None req-febd07a3-de81-4924-888e-f9c7d8ba2329 tempest-ServerMetadataNegativeTestJSON-1549470887 tempest-ServerMetadataNegativeTestJSON-1549470887-project-member] [instance: 31d47299-83a3-4f27-aeb7-95c4cd36c5ac] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 796.421604] env[62066]: DEBUG nova.network.neutron [None req-febd07a3-de81-4924-888e-f9c7d8ba2329 tempest-ServerMetadataNegativeTestJSON-1549470887 tempest-ServerMetadataNegativeTestJSON-1549470887-project-member] [instance: 31d47299-83a3-4f27-aeb7-95c4cd36c5ac] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 796.452566] env[62066]: DEBUG nova.compute.manager [req-66631c18-3119-4ecb-9cfc-f9e49cc7ad19 req-224001d5-2040-43fd-b6bb-646476cfaffd service nova] [instance: 459f432b-e817-405f-9c3e-dfa5d957788c] Received event network-vif-plugged-cb74e58f-b481-4a0f-861b-96381e2d52dc {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 796.453024] env[62066]: DEBUG oslo_concurrency.lockutils [req-66631c18-3119-4ecb-9cfc-f9e49cc7ad19 req-224001d5-2040-43fd-b6bb-646476cfaffd service nova] Acquiring lock "459f432b-e817-405f-9c3e-dfa5d957788c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 796.453024] env[62066]: DEBUG oslo_concurrency.lockutils [req-66631c18-3119-4ecb-9cfc-f9e49cc7ad19 req-224001d5-2040-43fd-b6bb-646476cfaffd service nova] Lock "459f432b-e817-405f-9c3e-dfa5d957788c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 796.453270] env[62066]: DEBUG oslo_concurrency.lockutils [req-66631c18-3119-4ecb-9cfc-f9e49cc7ad19 req-224001d5-2040-43fd-b6bb-646476cfaffd service nova] Lock "459f432b-e817-405f-9c3e-dfa5d957788c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 796.453324] env[62066]: DEBUG nova.compute.manager [req-66631c18-3119-4ecb-9cfc-f9e49cc7ad19 req-224001d5-2040-43fd-b6bb-646476cfaffd service nova] [instance: 459f432b-e817-405f-9c3e-dfa5d957788c] No waiting events found dispatching network-vif-plugged-cb74e58f-b481-4a0f-861b-96381e2d52dc {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 796.453496] env[62066]: WARNING nova.compute.manager [req-66631c18-3119-4ecb-9cfc-f9e49cc7ad19 req-224001d5-2040-43fd-b6bb-646476cfaffd service nova] [instance: 459f432b-e817-405f-9c3e-dfa5d957788c] Received unexpected event network-vif-plugged-cb74e58f-b481-4a0f-861b-96381e2d52dc for instance with vm_state building and task_state spawning. [ 796.453675] env[62066]: DEBUG nova.compute.manager [req-66631c18-3119-4ecb-9cfc-f9e49cc7ad19 req-224001d5-2040-43fd-b6bb-646476cfaffd service nova] [instance: 459f432b-e817-405f-9c3e-dfa5d957788c] Received event network-changed-cb74e58f-b481-4a0f-861b-96381e2d52dc {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 796.453910] env[62066]: DEBUG nova.compute.manager [req-66631c18-3119-4ecb-9cfc-f9e49cc7ad19 req-224001d5-2040-43fd-b6bb-646476cfaffd service nova] [instance: 459f432b-e817-405f-9c3e-dfa5d957788c] Refreshing instance network info cache due to event network-changed-cb74e58f-b481-4a0f-861b-96381e2d52dc. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 796.454426] env[62066]: DEBUG oslo_concurrency.lockutils [req-66631c18-3119-4ecb-9cfc-f9e49cc7ad19 req-224001d5-2040-43fd-b6bb-646476cfaffd service nova] Acquiring lock "refresh_cache-459f432b-e817-405f-9c3e-dfa5d957788c" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 796.454548] env[62066]: DEBUG oslo_concurrency.lockutils [req-66631c18-3119-4ecb-9cfc-f9e49cc7ad19 req-224001d5-2040-43fd-b6bb-646476cfaffd service nova] Acquired lock "refresh_cache-459f432b-e817-405f-9c3e-dfa5d957788c" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 796.454624] env[62066]: DEBUG nova.network.neutron [req-66631c18-3119-4ecb-9cfc-f9e49cc7ad19 req-224001d5-2040-43fd-b6bb-646476cfaffd service nova] [instance: 459f432b-e817-405f-9c3e-dfa5d957788c] Refreshing network info cache for port cb74e58f-b481-4a0f-861b-96381e2d52dc {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 796.467596] env[62066]: DEBUG nova.policy [None req-febd07a3-de81-4924-888e-f9c7d8ba2329 tempest-ServerMetadataNegativeTestJSON-1549470887 tempest-ServerMetadataNegativeTestJSON-1549470887-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '899ce0d363a64ee199eff40ae255a3d1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '47db3c06473c46a5a28865c0756ea745', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 796.572893] env[62066]: DEBUG oslo_concurrency.lockutils [None req-816774ed-d63b-4dd8-94be-e4c70c8f10e7 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Acquiring lock "refresh_cache-459f432b-e817-405f-9c3e-dfa5d957788c" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 796.698230] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c70bcb7f-78ff-424d-8121-684c45ecd12b tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Releasing lock "refresh_cache-02fb3e92-5dd6-4b1e-a6e4-d60d3fc85b07" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 796.698615] env[62066]: DEBUG nova.compute.manager [None req-c70bcb7f-78ff-424d-8121-684c45ecd12b tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 02fb3e92-5dd6-4b1e-a6e4-d60d3fc85b07] Instance network_info: |[{"id": "16bb647a-353c-4bf8-832d-8f6e49a6ce4b", "address": "fa:16:3e:47:44:80", "network": {"id": "9f3396ce-ee48-4eaf-8cc7-3e1711f37537", "bridge": "br-int", "label": "tempest-ServersTestJSON-2129775093-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6497ab02f327476d8ff81c2ecc0371e5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7ab8d568-adb0-4f3b-b6cc-68413e6546ae", "external-id": "nsx-vlan-transportzone-86", "segmentation_id": 86, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap16bb647a-35", "ovs_interfaceid": "16bb647a-353c-4bf8-832d-8f6e49a6ce4b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 796.701888] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-c70bcb7f-78ff-424d-8121-684c45ecd12b tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 02fb3e92-5dd6-4b1e-a6e4-d60d3fc85b07] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:47:44:80', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7ab8d568-adb0-4f3b-b6cc-68413e6546ae', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '16bb647a-353c-4bf8-832d-8f6e49a6ce4b', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 796.708700] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-c70bcb7f-78ff-424d-8121-684c45ecd12b tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Creating folder: Project (6497ab02f327476d8ff81c2ecc0371e5). Parent ref: group-v285980. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 796.709423] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-db65251a-85b0-4a4d-8e60-49dbd3207211 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.720248] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-c70bcb7f-78ff-424d-8121-684c45ecd12b tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Created folder: Project (6497ab02f327476d8ff81c2ecc0371e5) in parent group-v285980. [ 796.720468] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-c70bcb7f-78ff-424d-8121-684c45ecd12b tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Creating folder: Instances. Parent ref: group-v286014. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 796.720720] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1bb93173-09d5-48d0-924d-3d78f11e2b36 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.729480] env[62066]: DEBUG nova.network.neutron [None req-febd07a3-de81-4924-888e-f9c7d8ba2329 tempest-ServerMetadataNegativeTestJSON-1549470887 tempest-ServerMetadataNegativeTestJSON-1549470887-project-member] [instance: 31d47299-83a3-4f27-aeb7-95c4cd36c5ac] Successfully created port: 4d836af3-8ac2-4b67-842b-c868c679bd5e {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 796.732927] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-c70bcb7f-78ff-424d-8121-684c45ecd12b tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Created folder: Instances in parent group-v286014. [ 796.733195] env[62066]: DEBUG oslo.service.loopingcall [None req-c70bcb7f-78ff-424d-8121-684c45ecd12b tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 796.733571] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 02fb3e92-5dd6-4b1e-a6e4-d60d3fc85b07] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 796.733625] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d982e53a-c8b9-4acf-b1c6-280900560f3b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.756458] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 796.756458] env[62066]: value = "task-1340739" [ 796.756458] env[62066]: _type = "Task" [ 796.756458] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.764858] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1340739, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.872646] env[62066]: DEBUG oslo_vmware.api [None req-86820846-d757-4de7-918c-ad5501b560db tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] Task: {'id': task-1340736, 'name': Rename_Task, 'duration_secs': 0.128238} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 796.872946] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-86820846-d757-4de7-918c-ad5501b560db tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 796.873223] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-818f7fc9-1e0e-441e-93b5-ba85d15f70ce {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.879492] env[62066]: DEBUG oslo_vmware.api [None req-86820846-d757-4de7-918c-ad5501b560db tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] Waiting for the task: (returnval){ [ 796.879492] env[62066]: value = "task-1340740" [ 796.879492] env[62066]: _type = "Task" [ 796.879492] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.888901] env[62066]: DEBUG oslo_vmware.api [None req-86820846-d757-4de7-918c-ad5501b560db tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] Task: {'id': task-1340740, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.924672] env[62066]: DEBUG nova.compute.manager [None req-febd07a3-de81-4924-888e-f9c7d8ba2329 tempest-ServerMetadataNegativeTestJSON-1549470887 tempest-ServerMetadataNegativeTestJSON-1549470887-project-member] [instance: 31d47299-83a3-4f27-aeb7-95c4cd36c5ac] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 796.928700] env[62066]: DEBUG nova.compute.manager [None req-dcd99ffb-af75-402a-906d-756c3b1954c8 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] [instance: 1a4b0637-1a56-41ef-b89b-6b56d24ed206] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 796.930119] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75cfaad0-b4f1-436b-8ed4-876d6b9020d6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.999294] env[62066]: DEBUG nova.network.neutron [req-66631c18-3119-4ecb-9cfc-f9e49cc7ad19 req-224001d5-2040-43fd-b6bb-646476cfaffd service nova] [instance: 459f432b-e817-405f-9c3e-dfa5d957788c] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 797.132446] env[62066]: DEBUG nova.network.neutron [req-66631c18-3119-4ecb-9cfc-f9e49cc7ad19 req-224001d5-2040-43fd-b6bb-646476cfaffd service nova] [instance: 459f432b-e817-405f-9c3e-dfa5d957788c] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 797.173367] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b3dfe06-5d8d-454a-9e98-094587511541 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.183091] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-320fff53-0128-45ee-bc8f-73203e4daaab {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.220619] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32344049-0813-465c-9824-a577f89152f4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.228439] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aeac0f9d-6f48-422b-add9-532492637b80 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.242100] env[62066]: DEBUG nova.compute.provider_tree [None req-df48af02-4c4b-4364-8d38-efc343bf8241 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 797.266384] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1340739, 'name': CreateVM_Task, 'duration_secs': 0.482547} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.266585] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 02fb3e92-5dd6-4b1e-a6e4-d60d3fc85b07] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 797.268027] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c70bcb7f-78ff-424d-8121-684c45ecd12b tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 797.268027] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c70bcb7f-78ff-424d-8121-684c45ecd12b tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Acquired lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 797.268174] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c70bcb7f-78ff-424d-8121-684c45ecd12b tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 797.268435] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-941baefd-58ad-40b3-b878-ba7de032c864 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.275377] env[62066]: DEBUG oslo_vmware.api [None req-c70bcb7f-78ff-424d-8121-684c45ecd12b tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Waiting for the task: (returnval){ [ 797.275377] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]5276c0a6-8183-1478-f305-bf40348a7621" [ 797.275377] env[62066]: _type = "Task" [ 797.275377] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.284480] env[62066]: DEBUG oslo_vmware.api [None req-c70bcb7f-78ff-424d-8121-684c45ecd12b tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]5276c0a6-8183-1478-f305-bf40348a7621, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.390990] env[62066]: DEBUG oslo_vmware.api [None req-86820846-d757-4de7-918c-ad5501b560db tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] Task: {'id': task-1340740, 'name': PowerOnVM_Task} progress is 94%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.443054] env[62066]: INFO nova.compute.manager [None req-dcd99ffb-af75-402a-906d-756c3b1954c8 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] [instance: 1a4b0637-1a56-41ef-b89b-6b56d24ed206] instance snapshotting [ 797.445708] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad29cc62-b3aa-4871-90d3-396ecc596934 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.464829] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a87e9851-2b6e-4e18-95dc-82318de19ed2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.637621] env[62066]: DEBUG oslo_concurrency.lockutils [req-66631c18-3119-4ecb-9cfc-f9e49cc7ad19 req-224001d5-2040-43fd-b6bb-646476cfaffd service nova] Releasing lock "refresh_cache-459f432b-e817-405f-9c3e-dfa5d957788c" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 797.637983] env[62066]: DEBUG oslo_concurrency.lockutils [None req-816774ed-d63b-4dd8-94be-e4c70c8f10e7 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Acquired lock "refresh_cache-459f432b-e817-405f-9c3e-dfa5d957788c" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 797.638174] env[62066]: DEBUG nova.network.neutron [None req-816774ed-d63b-4dd8-94be-e4c70c8f10e7 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: 459f432b-e817-405f-9c3e-dfa5d957788c] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 797.745663] env[62066]: DEBUG nova.scheduler.client.report [None req-df48af02-4c4b-4364-8d38-efc343bf8241 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 797.786679] env[62066]: DEBUG oslo_vmware.api [None req-c70bcb7f-78ff-424d-8121-684c45ecd12b tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]5276c0a6-8183-1478-f305-bf40348a7621, 'name': SearchDatastore_Task, 'duration_secs': 0.043448} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.787057] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c70bcb7f-78ff-424d-8121-684c45ecd12b tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Releasing lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 797.787392] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-c70bcb7f-78ff-424d-8121-684c45ecd12b tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 02fb3e92-5dd6-4b1e-a6e4-d60d3fc85b07] Processing image 50ff584c-3b50-4395-af07-3e66769bc9f7 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 797.787724] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c70bcb7f-78ff-424d-8121-684c45ecd12b tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 797.787909] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c70bcb7f-78ff-424d-8121-684c45ecd12b tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Acquired lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 797.788128] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-c70bcb7f-78ff-424d-8121-684c45ecd12b tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 797.788415] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e227a213-adfd-4729-94fd-02a410f3b28b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.797200] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-c70bcb7f-78ff-424d-8121-684c45ecd12b tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 797.797402] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-c70bcb7f-78ff-424d-8121-684c45ecd12b tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 797.798190] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fcbc98f0-20f6-4ef3-8dc8-ed87a11337fc {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.803602] env[62066]: DEBUG oslo_vmware.api [None req-c70bcb7f-78ff-424d-8121-684c45ecd12b tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Waiting for the task: (returnval){ [ 797.803602] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]52990574-f132-fdb0-0b23-425894b9cf16" [ 797.803602] env[62066]: _type = "Task" [ 797.803602] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.810945] env[62066]: DEBUG oslo_vmware.api [None req-c70bcb7f-78ff-424d-8121-684c45ecd12b tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52990574-f132-fdb0-0b23-425894b9cf16, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.889430] env[62066]: DEBUG oslo_vmware.api [None req-86820846-d757-4de7-918c-ad5501b560db tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] Task: {'id': task-1340740, 'name': PowerOnVM_Task, 'duration_secs': 0.783039} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.889693] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-86820846-d757-4de7-918c-ad5501b560db tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 797.889900] env[62066]: INFO nova.compute.manager [None req-86820846-d757-4de7-918c-ad5501b560db tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] Took 5.45 seconds to spawn the instance on the hypervisor. [ 797.890126] env[62066]: DEBUG nova.compute.manager [None req-86820846-d757-4de7-918c-ad5501b560db tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 797.890843] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b2d4d19-9e66-4794-b4f7-dbb1fdd99c51 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.936689] env[62066]: DEBUG nova.compute.manager [None req-febd07a3-de81-4924-888e-f9c7d8ba2329 tempest-ServerMetadataNegativeTestJSON-1549470887 tempest-ServerMetadataNegativeTestJSON-1549470887-project-member] [instance: 31d47299-83a3-4f27-aeb7-95c4cd36c5ac] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 797.958644] env[62066]: DEBUG nova.virt.hardware [None req-febd07a3-de81-4924-888e-f9c7d8ba2329 tempest-ServerMetadataNegativeTestJSON-1549470887 tempest-ServerMetadataNegativeTestJSON-1549470887-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-23T13:40:41Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-23T13:40:23Z,direct_url=,disk_format='vmdk',id=50ff584c-3b50-4395-af07-3e66769bc9f7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='eb52f7069a374c61ae946f052007c6d9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-23T13:40:24Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 797.958901] env[62066]: DEBUG nova.virt.hardware [None req-febd07a3-de81-4924-888e-f9c7d8ba2329 tempest-ServerMetadataNegativeTestJSON-1549470887 tempest-ServerMetadataNegativeTestJSON-1549470887-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 797.959070] env[62066]: DEBUG nova.virt.hardware [None req-febd07a3-de81-4924-888e-f9c7d8ba2329 tempest-ServerMetadataNegativeTestJSON-1549470887 tempest-ServerMetadataNegativeTestJSON-1549470887-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 797.959338] env[62066]: DEBUG nova.virt.hardware [None req-febd07a3-de81-4924-888e-f9c7d8ba2329 tempest-ServerMetadataNegativeTestJSON-1549470887 tempest-ServerMetadataNegativeTestJSON-1549470887-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 797.959567] env[62066]: DEBUG nova.virt.hardware [None req-febd07a3-de81-4924-888e-f9c7d8ba2329 tempest-ServerMetadataNegativeTestJSON-1549470887 tempest-ServerMetadataNegativeTestJSON-1549470887-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 797.959642] env[62066]: DEBUG nova.virt.hardware [None req-febd07a3-de81-4924-888e-f9c7d8ba2329 tempest-ServerMetadataNegativeTestJSON-1549470887 tempest-ServerMetadataNegativeTestJSON-1549470887-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 797.959833] env[62066]: DEBUG nova.virt.hardware [None req-febd07a3-de81-4924-888e-f9c7d8ba2329 tempest-ServerMetadataNegativeTestJSON-1549470887 tempest-ServerMetadataNegativeTestJSON-1549470887-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 797.959990] env[62066]: DEBUG nova.virt.hardware [None req-febd07a3-de81-4924-888e-f9c7d8ba2329 tempest-ServerMetadataNegativeTestJSON-1549470887 tempest-ServerMetadataNegativeTestJSON-1549470887-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 797.960215] env[62066]: DEBUG nova.virt.hardware [None req-febd07a3-de81-4924-888e-f9c7d8ba2329 tempest-ServerMetadataNegativeTestJSON-1549470887 tempest-ServerMetadataNegativeTestJSON-1549470887-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 797.960428] env[62066]: DEBUG nova.virt.hardware [None req-febd07a3-de81-4924-888e-f9c7d8ba2329 tempest-ServerMetadataNegativeTestJSON-1549470887 tempest-ServerMetadataNegativeTestJSON-1549470887-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 797.960608] env[62066]: DEBUG nova.virt.hardware [None req-febd07a3-de81-4924-888e-f9c7d8ba2329 tempest-ServerMetadataNegativeTestJSON-1549470887 tempest-ServerMetadataNegativeTestJSON-1549470887-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 797.961482] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a870046a-6162-490c-9ef6-e1d7fd32148b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.969079] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ee42232-7f7d-497c-86d2-88416fcf33e3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.982945] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-dcd99ffb-af75-402a-906d-756c3b1954c8 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] [instance: 1a4b0637-1a56-41ef-b89b-6b56d24ed206] Creating Snapshot of the VM instance {{(pid=62066) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 797.983367] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-3aa05dde-ec79-4d26-9792-0fb058132ac0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.988842] env[62066]: DEBUG oslo_vmware.api [None req-dcd99ffb-af75-402a-906d-756c3b1954c8 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Waiting for the task: (returnval){ [ 797.988842] env[62066]: value = "task-1340741" [ 797.988842] env[62066]: _type = "Task" [ 797.988842] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.996197] env[62066]: DEBUG oslo_vmware.api [None req-dcd99ffb-af75-402a-906d-756c3b1954c8 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Task: {'id': task-1340741, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.193598] env[62066]: DEBUG nova.network.neutron [None req-816774ed-d63b-4dd8-94be-e4c70c8f10e7 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: 459f432b-e817-405f-9c3e-dfa5d957788c] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 798.250737] env[62066]: DEBUG oslo_concurrency.lockutils [None req-df48af02-4c4b-4364-8d38-efc343bf8241 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.340s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 798.251309] env[62066]: DEBUG nova.compute.manager [None req-df48af02-4c4b-4364-8d38-efc343bf8241 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 49f46244-34fa-48a1-95a2-8e95850f345d] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 798.257503] env[62066]: DEBUG oslo_concurrency.lockutils [None req-8034bd7f-2294-4f4c-b924-f35bec7ebaac tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.377s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 798.258979] env[62066]: INFO nova.compute.claims [None req-8034bd7f-2294-4f4c-b924-f35bec7ebaac tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] [instance: 738ebb2f-7cc6-4d0f-871c-1428c9c21384] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 798.321023] env[62066]: DEBUG oslo_vmware.api [None req-c70bcb7f-78ff-424d-8121-684c45ecd12b tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52990574-f132-fdb0-0b23-425894b9cf16, 'name': SearchDatastore_Task, 'duration_secs': 0.032707} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.321023] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2e4251da-e7e5-42ef-bb27-17e9e40452b7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.324909] env[62066]: DEBUG oslo_vmware.api [None req-c70bcb7f-78ff-424d-8121-684c45ecd12b tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Waiting for the task: (returnval){ [ 798.324909] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]52750cc1-1566-a01c-da3d-fc1f4d8071fa" [ 798.324909] env[62066]: _type = "Task" [ 798.324909] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.346079] env[62066]: DEBUG oslo_vmware.api [None req-c70bcb7f-78ff-424d-8121-684c45ecd12b tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52750cc1-1566-a01c-da3d-fc1f4d8071fa, 'name': SearchDatastore_Task, 'duration_secs': 0.009539} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.347668] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c70bcb7f-78ff-424d-8121-684c45ecd12b tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Releasing lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 798.348490] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-c70bcb7f-78ff-424d-8121-684c45ecd12b tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk to [datastore2] 02fb3e92-5dd6-4b1e-a6e4-d60d3fc85b07/02fb3e92-5dd6-4b1e-a6e4-d60d3fc85b07.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 798.350020] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9c4ef185-08ab-483b-bc9a-58b7f2ff4c80 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.356183] env[62066]: DEBUG oslo_vmware.api [None req-c70bcb7f-78ff-424d-8121-684c45ecd12b tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Waiting for the task: (returnval){ [ 798.356183] env[62066]: value = "task-1340742" [ 798.356183] env[62066]: _type = "Task" [ 798.356183] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.365342] env[62066]: DEBUG oslo_vmware.api [None req-c70bcb7f-78ff-424d-8121-684c45ecd12b tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': task-1340742, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.410946] env[62066]: INFO nova.compute.manager [None req-86820846-d757-4de7-918c-ad5501b560db tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] Took 26.09 seconds to build instance. [ 798.469518] env[62066]: DEBUG nova.network.neutron [None req-816774ed-d63b-4dd8-94be-e4c70c8f10e7 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: 459f432b-e817-405f-9c3e-dfa5d957788c] Updating instance_info_cache with network_info: [{"id": "cb74e58f-b481-4a0f-861b-96381e2d52dc", "address": "fa:16:3e:87:b9:2f", "network": {"id": "449dfe4e-bb49-43b6-9ba7-b57af74ebfb3", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-282029676-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0d4e33a99c7741fb8cdd97f4ec5dbbd8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db68bd64-5b56-49af-a075-13dcf85cb2e0", "external-id": "nsx-vlan-transportzone-590", "segmentation_id": 590, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcb74e58f-b4", "ovs_interfaceid": "cb74e58f-b481-4a0f-861b-96381e2d52dc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 798.499093] env[62066]: DEBUG oslo_vmware.api [None req-dcd99ffb-af75-402a-906d-756c3b1954c8 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Task: {'id': task-1340741, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.534594] env[62066]: DEBUG nova.compute.manager [req-57338d80-cfb0-41d2-9c6b-edb717400443 req-fa191111-097b-41c4-9e29-53ad2fc04257 service nova] [instance: 31d47299-83a3-4f27-aeb7-95c4cd36c5ac] Received event network-vif-plugged-4d836af3-8ac2-4b67-842b-c868c679bd5e {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 798.534815] env[62066]: DEBUG oslo_concurrency.lockutils [req-57338d80-cfb0-41d2-9c6b-edb717400443 req-fa191111-097b-41c4-9e29-53ad2fc04257 service nova] Acquiring lock "31d47299-83a3-4f27-aeb7-95c4cd36c5ac-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 798.534981] env[62066]: DEBUG oslo_concurrency.lockutils [req-57338d80-cfb0-41d2-9c6b-edb717400443 req-fa191111-097b-41c4-9e29-53ad2fc04257 service nova] Lock "31d47299-83a3-4f27-aeb7-95c4cd36c5ac-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 798.535176] env[62066]: DEBUG oslo_concurrency.lockutils [req-57338d80-cfb0-41d2-9c6b-edb717400443 req-fa191111-097b-41c4-9e29-53ad2fc04257 service nova] Lock "31d47299-83a3-4f27-aeb7-95c4cd36c5ac-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 798.535348] env[62066]: DEBUG nova.compute.manager [req-57338d80-cfb0-41d2-9c6b-edb717400443 req-fa191111-097b-41c4-9e29-53ad2fc04257 service nova] [instance: 31d47299-83a3-4f27-aeb7-95c4cd36c5ac] No waiting events found dispatching network-vif-plugged-4d836af3-8ac2-4b67-842b-c868c679bd5e {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 798.535511] env[62066]: WARNING nova.compute.manager [req-57338d80-cfb0-41d2-9c6b-edb717400443 req-fa191111-097b-41c4-9e29-53ad2fc04257 service nova] [instance: 31d47299-83a3-4f27-aeb7-95c4cd36c5ac] Received unexpected event network-vif-plugged-4d836af3-8ac2-4b67-842b-c868c679bd5e for instance with vm_state building and task_state spawning. [ 798.595321] env[62066]: DEBUG nova.network.neutron [None req-febd07a3-de81-4924-888e-f9c7d8ba2329 tempest-ServerMetadataNegativeTestJSON-1549470887 tempest-ServerMetadataNegativeTestJSON-1549470887-project-member] [instance: 31d47299-83a3-4f27-aeb7-95c4cd36c5ac] Successfully updated port: 4d836af3-8ac2-4b67-842b-c868c679bd5e {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 798.764586] env[62066]: DEBUG nova.compute.utils [None req-df48af02-4c4b-4364-8d38-efc343bf8241 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 798.771033] env[62066]: DEBUG nova.compute.manager [None req-df48af02-4c4b-4364-8d38-efc343bf8241 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 49f46244-34fa-48a1-95a2-8e95850f345d] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 798.771227] env[62066]: DEBUG nova.network.neutron [None req-df48af02-4c4b-4364-8d38-efc343bf8241 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 49f46244-34fa-48a1-95a2-8e95850f345d] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 798.820166] env[62066]: DEBUG nova.policy [None req-df48af02-4c4b-4364-8d38-efc343bf8241 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '60e2bc61839b4299912c53137bc23fbc', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '75f0352c852947369474c9e05766c584', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 798.872937] env[62066]: DEBUG oslo_vmware.api [None req-c70bcb7f-78ff-424d-8121-684c45ecd12b tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': task-1340742, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.913805] env[62066]: DEBUG oslo_concurrency.lockutils [None req-86820846-d757-4de7-918c-ad5501b560db tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] Lock "daffaf51-4c45-44aa-8fc2-4db066a09971" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 150.877s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 798.979019] env[62066]: DEBUG oslo_concurrency.lockutils [None req-816774ed-d63b-4dd8-94be-e4c70c8f10e7 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Releasing lock "refresh_cache-459f432b-e817-405f-9c3e-dfa5d957788c" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 798.979019] env[62066]: DEBUG nova.compute.manager [None req-816774ed-d63b-4dd8-94be-e4c70c8f10e7 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: 459f432b-e817-405f-9c3e-dfa5d957788c] Instance network_info: |[{"id": "cb74e58f-b481-4a0f-861b-96381e2d52dc", "address": "fa:16:3e:87:b9:2f", "network": {"id": "449dfe4e-bb49-43b6-9ba7-b57af74ebfb3", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-282029676-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0d4e33a99c7741fb8cdd97f4ec5dbbd8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db68bd64-5b56-49af-a075-13dcf85cb2e0", "external-id": "nsx-vlan-transportzone-590", "segmentation_id": 590, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcb74e58f-b4", "ovs_interfaceid": "cb74e58f-b481-4a0f-861b-96381e2d52dc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 798.979251] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-816774ed-d63b-4dd8-94be-e4c70c8f10e7 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: 459f432b-e817-405f-9c3e-dfa5d957788c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:87:b9:2f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'db68bd64-5b56-49af-a075-13dcf85cb2e0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'cb74e58f-b481-4a0f-861b-96381e2d52dc', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 798.987393] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-816774ed-d63b-4dd8-94be-e4c70c8f10e7 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Creating folder: Project (0d4e33a99c7741fb8cdd97f4ec5dbbd8). Parent ref: group-v285980. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 798.987702] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-637c6ba4-0219-4a9a-94b5-8a1659a268a4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.999976] env[62066]: DEBUG oslo_vmware.api [None req-dcd99ffb-af75-402a-906d-756c3b1954c8 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Task: {'id': task-1340741, 'name': CreateSnapshot_Task, 'duration_secs': 0.970666} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.000251] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-dcd99ffb-af75-402a-906d-756c3b1954c8 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] [instance: 1a4b0637-1a56-41ef-b89b-6b56d24ed206] Created Snapshot of the VM instance {{(pid=62066) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 799.001161] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-059f541c-dddc-4442-8c1d-1d331268f4e1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.004697] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-816774ed-d63b-4dd8-94be-e4c70c8f10e7 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Created folder: Project (0d4e33a99c7741fb8cdd97f4ec5dbbd8) in parent group-v285980. [ 799.004879] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-816774ed-d63b-4dd8-94be-e4c70c8f10e7 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Creating folder: Instances. Parent ref: group-v286018. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 799.005425] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0a5af6a6-026a-4a42-be62-87380e146d8e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.014462] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-816774ed-d63b-4dd8-94be-e4c70c8f10e7 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Created folder: Instances in parent group-v286018. [ 799.014722] env[62066]: DEBUG oslo.service.loopingcall [None req-816774ed-d63b-4dd8-94be-e4c70c8f10e7 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 799.014870] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 459f432b-e817-405f-9c3e-dfa5d957788c] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 799.016069] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e2893f07-de96-4d40-84e1-4f5e16ee5da8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.034247] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 799.034247] env[62066]: value = "task-1340745" [ 799.034247] env[62066]: _type = "Task" [ 799.034247] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.043642] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1340745, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.098621] env[62066]: DEBUG oslo_concurrency.lockutils [None req-febd07a3-de81-4924-888e-f9c7d8ba2329 tempest-ServerMetadataNegativeTestJSON-1549470887 tempest-ServerMetadataNegativeTestJSON-1549470887-project-member] Acquiring lock "refresh_cache-31d47299-83a3-4f27-aeb7-95c4cd36c5ac" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 799.098885] env[62066]: DEBUG oslo_concurrency.lockutils [None req-febd07a3-de81-4924-888e-f9c7d8ba2329 tempest-ServerMetadataNegativeTestJSON-1549470887 tempest-ServerMetadataNegativeTestJSON-1549470887-project-member] Acquired lock "refresh_cache-31d47299-83a3-4f27-aeb7-95c4cd36c5ac" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 799.098958] env[62066]: DEBUG nova.network.neutron [None req-febd07a3-de81-4924-888e-f9c7d8ba2329 tempest-ServerMetadataNegativeTestJSON-1549470887 tempest-ServerMetadataNegativeTestJSON-1549470887-project-member] [instance: 31d47299-83a3-4f27-aeb7-95c4cd36c5ac] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 799.105440] env[62066]: DEBUG nova.network.neutron [None req-df48af02-4c4b-4364-8d38-efc343bf8241 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 49f46244-34fa-48a1-95a2-8e95850f345d] Successfully created port: 213f129a-7b70-4807-ac40-621d02749750 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 799.271710] env[62066]: DEBUG nova.compute.manager [None req-df48af02-4c4b-4364-8d38-efc343bf8241 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 49f46244-34fa-48a1-95a2-8e95850f345d] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 799.366977] env[62066]: DEBUG oslo_vmware.api [None req-c70bcb7f-78ff-424d-8121-684c45ecd12b tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': task-1340742, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.533416} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.369580] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-c70bcb7f-78ff-424d-8121-684c45ecd12b tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk to [datastore2] 02fb3e92-5dd6-4b1e-a6e4-d60d3fc85b07/02fb3e92-5dd6-4b1e-a6e4-d60d3fc85b07.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 799.369798] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-c70bcb7f-78ff-424d-8121-684c45ecd12b tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 02fb3e92-5dd6-4b1e-a6e4-d60d3fc85b07] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 799.370287] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d7d8a762-43b2-4baa-b228-67122b03f9b3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.377155] env[62066]: DEBUG oslo_vmware.api [None req-c70bcb7f-78ff-424d-8121-684c45ecd12b tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Waiting for the task: (returnval){ [ 799.377155] env[62066]: value = "task-1340746" [ 799.377155] env[62066]: _type = "Task" [ 799.377155] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.386160] env[62066]: DEBUG oslo_vmware.api [None req-c70bcb7f-78ff-424d-8121-684c45ecd12b tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': task-1340746, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.416764] env[62066]: DEBUG nova.compute.manager [None req-3d67ea1c-5f75-4620-b677-6f1e0faf1bd6 tempest-ServersV294TestFqdnHostnames-843300434 tempest-ServersV294TestFqdnHostnames-843300434-project-member] [instance: 2709d0ad-5dd8-4e3c-b1e6-3d1e2e242380] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 799.521025] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-dcd99ffb-af75-402a-906d-756c3b1954c8 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] [instance: 1a4b0637-1a56-41ef-b89b-6b56d24ed206] Creating linked-clone VM from snapshot {{(pid=62066) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 799.521344] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-6be7e761-9062-447c-baef-fed9ee79d569 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.533230] env[62066]: DEBUG oslo_vmware.api [None req-dcd99ffb-af75-402a-906d-756c3b1954c8 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Waiting for the task: (returnval){ [ 799.533230] env[62066]: value = "task-1340747" [ 799.533230] env[62066]: _type = "Task" [ 799.533230] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.540731] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da5ce1e3-34d8-46b2-973a-0425954650aa {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.547024] env[62066]: DEBUG oslo_vmware.api [None req-dcd99ffb-af75-402a-906d-756c3b1954c8 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Task: {'id': task-1340747, 'name': CloneVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.550631] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1340745, 'name': CreateVM_Task, 'duration_secs': 0.354443} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.552274] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 459f432b-e817-405f-9c3e-dfa5d957788c] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 799.552966] env[62066]: DEBUG oslo_concurrency.lockutils [None req-816774ed-d63b-4dd8-94be-e4c70c8f10e7 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 799.553753] env[62066]: DEBUG oslo_concurrency.lockutils [None req-816774ed-d63b-4dd8-94be-e4c70c8f10e7 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Acquired lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 799.553753] env[62066]: DEBUG oslo_concurrency.lockutils [None req-816774ed-d63b-4dd8-94be-e4c70c8f10e7 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 799.554512] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-851cc7ea-6fe7-4f92-8b27-7c8e45e10437 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.557612] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dbfd39e3-9810-4622-9bc8-469d2b0b86d8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.562165] env[62066]: DEBUG oslo_vmware.api [None req-816774ed-d63b-4dd8-94be-e4c70c8f10e7 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Waiting for the task: (returnval){ [ 799.562165] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]527104d9-c18e-699d-8968-56f867ad62fc" [ 799.562165] env[62066]: _type = "Task" [ 799.562165] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.595236] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-876b1fb4-2cc2-4087-9a2d-03f15ee73d0f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.608947] env[62066]: DEBUG oslo_vmware.api [None req-816774ed-d63b-4dd8-94be-e4c70c8f10e7 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]527104d9-c18e-699d-8968-56f867ad62fc, 'name': SearchDatastore_Task, 'duration_secs': 0.038256} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.611826] env[62066]: DEBUG oslo_concurrency.lockutils [None req-816774ed-d63b-4dd8-94be-e4c70c8f10e7 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Releasing lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 799.612039] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-816774ed-d63b-4dd8-94be-e4c70c8f10e7 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: 459f432b-e817-405f-9c3e-dfa5d957788c] Processing image 50ff584c-3b50-4395-af07-3e66769bc9f7 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 799.612726] env[62066]: DEBUG oslo_concurrency.lockutils [None req-816774ed-d63b-4dd8-94be-e4c70c8f10e7 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 799.612880] env[62066]: DEBUG oslo_concurrency.lockutils [None req-816774ed-d63b-4dd8-94be-e4c70c8f10e7 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Acquired lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 799.613078] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-816774ed-d63b-4dd8-94be-e4c70c8f10e7 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 799.613489] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1f267c98-b8d6-4e96-8381-5907f4f003c6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.616710] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fff0c67-70b5-4ec7-975b-081b2980b699 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.633694] env[62066]: DEBUG nova.compute.provider_tree [None req-8034bd7f-2294-4f4c-b924-f35bec7ebaac tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 799.636880] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-816774ed-d63b-4dd8-94be-e4c70c8f10e7 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 799.637172] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-816774ed-d63b-4dd8-94be-e4c70c8f10e7 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 799.639016] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bcdc7066-7517-4a8a-8919-dd8d73db5052 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.641060] env[62066]: DEBUG nova.network.neutron [None req-febd07a3-de81-4924-888e-f9c7d8ba2329 tempest-ServerMetadataNegativeTestJSON-1549470887 tempest-ServerMetadataNegativeTestJSON-1549470887-project-member] [instance: 31d47299-83a3-4f27-aeb7-95c4cd36c5ac] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 799.646427] env[62066]: DEBUG oslo_vmware.api [None req-816774ed-d63b-4dd8-94be-e4c70c8f10e7 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Waiting for the task: (returnval){ [ 799.646427] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]52e510ac-f1ca-1311-b8ae-094cb1e7aeb9" [ 799.646427] env[62066]: _type = "Task" [ 799.646427] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.655403] env[62066]: DEBUG oslo_vmware.api [None req-816774ed-d63b-4dd8-94be-e4c70c8f10e7 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52e510ac-f1ca-1311-b8ae-094cb1e7aeb9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.775971] env[62066]: DEBUG nova.compute.manager [req-170d002b-1f49-47dd-8443-ad3de69f6baf req-ef457172-7257-41b7-a4e4-9fac2692fd20 service nova] [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] Received event network-changed-88380b24-b187-4b82-b92e-b93adeeab3eb {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 799.776283] env[62066]: DEBUG nova.compute.manager [req-170d002b-1f49-47dd-8443-ad3de69f6baf req-ef457172-7257-41b7-a4e4-9fac2692fd20 service nova] [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] Refreshing instance network info cache due to event network-changed-88380b24-b187-4b82-b92e-b93adeeab3eb. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 799.776432] env[62066]: DEBUG oslo_concurrency.lockutils [req-170d002b-1f49-47dd-8443-ad3de69f6baf req-ef457172-7257-41b7-a4e4-9fac2692fd20 service nova] Acquiring lock "refresh_cache-daffaf51-4c45-44aa-8fc2-4db066a09971" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 799.776572] env[62066]: DEBUG oslo_concurrency.lockutils [req-170d002b-1f49-47dd-8443-ad3de69f6baf req-ef457172-7257-41b7-a4e4-9fac2692fd20 service nova] Acquired lock "refresh_cache-daffaf51-4c45-44aa-8fc2-4db066a09971" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 799.776768] env[62066]: DEBUG nova.network.neutron [req-170d002b-1f49-47dd-8443-ad3de69f6baf req-ef457172-7257-41b7-a4e4-9fac2692fd20 service nova] [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] Refreshing network info cache for port 88380b24-b187-4b82-b92e-b93adeeab3eb {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 799.801831] env[62066]: DEBUG nova.network.neutron [None req-febd07a3-de81-4924-888e-f9c7d8ba2329 tempest-ServerMetadataNegativeTestJSON-1549470887 tempest-ServerMetadataNegativeTestJSON-1549470887-project-member] [instance: 31d47299-83a3-4f27-aeb7-95c4cd36c5ac] Updating instance_info_cache with network_info: [{"id": "4d836af3-8ac2-4b67-842b-c868c679bd5e", "address": "fa:16:3e:f3:45:37", "network": {"id": "6b1d3873-7d3c-42ef-a70c-df94aac1e6eb", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-689803675-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "47db3c06473c46a5a28865c0756ea745", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1323cb03-8367-485a-962e-131af8eba474", "external-id": "nsx-vlan-transportzone-41", "segmentation_id": 41, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4d836af3-8a", "ovs_interfaceid": "4d836af3-8ac2-4b67-842b-c868c679bd5e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 799.891011] env[62066]: DEBUG oslo_vmware.api [None req-c70bcb7f-78ff-424d-8121-684c45ecd12b tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': task-1340746, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070849} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.891099] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-c70bcb7f-78ff-424d-8121-684c45ecd12b tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 02fb3e92-5dd6-4b1e-a6e4-d60d3fc85b07] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 799.891943] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7d85b17-4ece-43f1-bfd5-09f24167999d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.915095] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-c70bcb7f-78ff-424d-8121-684c45ecd12b tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 02fb3e92-5dd6-4b1e-a6e4-d60d3fc85b07] Reconfiguring VM instance instance-0000003d to attach disk [datastore2] 02fb3e92-5dd6-4b1e-a6e4-d60d3fc85b07/02fb3e92-5dd6-4b1e-a6e4-d60d3fc85b07.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 799.915798] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ef78a9d6-9f5f-4b67-86c5-872051bbb79b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.939058] env[62066]: DEBUG oslo_vmware.api [None req-c70bcb7f-78ff-424d-8121-684c45ecd12b tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Waiting for the task: (returnval){ [ 799.939058] env[62066]: value = "task-1340748" [ 799.939058] env[62066]: _type = "Task" [ 799.939058] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.950555] env[62066]: DEBUG oslo_vmware.api [None req-c70bcb7f-78ff-424d-8121-684c45ecd12b tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': task-1340748, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.959102] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3d67ea1c-5f75-4620-b677-6f1e0faf1bd6 tempest-ServersV294TestFqdnHostnames-843300434 tempest-ServersV294TestFqdnHostnames-843300434-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 800.043351] env[62066]: DEBUG oslo_vmware.api [None req-dcd99ffb-af75-402a-906d-756c3b1954c8 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Task: {'id': task-1340747, 'name': CloneVM_Task} progress is 94%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.137131] env[62066]: DEBUG nova.scheduler.client.report [None req-8034bd7f-2294-4f4c-b924-f35bec7ebaac tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 800.163478] env[62066]: DEBUG oslo_vmware.api [None req-816774ed-d63b-4dd8-94be-e4c70c8f10e7 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52e510ac-f1ca-1311-b8ae-094cb1e7aeb9, 'name': SearchDatastore_Task, 'duration_secs': 0.052593} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 800.165805] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a51cacf3-5b75-4df7-b5ff-3e953750b8cd {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.173725] env[62066]: DEBUG oslo_vmware.api [None req-816774ed-d63b-4dd8-94be-e4c70c8f10e7 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Waiting for the task: (returnval){ [ 800.173725] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]52ac090f-0c20-f6c9-5298-d378b3be2cf8" [ 800.173725] env[62066]: _type = "Task" [ 800.173725] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 800.188611] env[62066]: DEBUG oslo_vmware.api [None req-816774ed-d63b-4dd8-94be-e4c70c8f10e7 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52ac090f-0c20-f6c9-5298-d378b3be2cf8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.288713] env[62066]: DEBUG nova.compute.manager [None req-df48af02-4c4b-4364-8d38-efc343bf8241 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 49f46244-34fa-48a1-95a2-8e95850f345d] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 800.306273] env[62066]: DEBUG oslo_concurrency.lockutils [None req-febd07a3-de81-4924-888e-f9c7d8ba2329 tempest-ServerMetadataNegativeTestJSON-1549470887 tempest-ServerMetadataNegativeTestJSON-1549470887-project-member] Releasing lock "refresh_cache-31d47299-83a3-4f27-aeb7-95c4cd36c5ac" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 800.306583] env[62066]: DEBUG nova.compute.manager [None req-febd07a3-de81-4924-888e-f9c7d8ba2329 tempest-ServerMetadataNegativeTestJSON-1549470887 tempest-ServerMetadataNegativeTestJSON-1549470887-project-member] [instance: 31d47299-83a3-4f27-aeb7-95c4cd36c5ac] Instance network_info: |[{"id": "4d836af3-8ac2-4b67-842b-c868c679bd5e", "address": "fa:16:3e:f3:45:37", "network": {"id": "6b1d3873-7d3c-42ef-a70c-df94aac1e6eb", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-689803675-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "47db3c06473c46a5a28865c0756ea745", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1323cb03-8367-485a-962e-131af8eba474", "external-id": "nsx-vlan-transportzone-41", "segmentation_id": 41, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4d836af3-8a", "ovs_interfaceid": "4d836af3-8ac2-4b67-842b-c868c679bd5e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 800.307248] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-febd07a3-de81-4924-888e-f9c7d8ba2329 tempest-ServerMetadataNegativeTestJSON-1549470887 tempest-ServerMetadataNegativeTestJSON-1549470887-project-member] [instance: 31d47299-83a3-4f27-aeb7-95c4cd36c5ac] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f3:45:37', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '1323cb03-8367-485a-962e-131af8eba474', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4d836af3-8ac2-4b67-842b-c868c679bd5e', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 800.314631] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-febd07a3-de81-4924-888e-f9c7d8ba2329 tempest-ServerMetadataNegativeTestJSON-1549470887 tempest-ServerMetadataNegativeTestJSON-1549470887-project-member] Creating folder: Project (47db3c06473c46a5a28865c0756ea745). Parent ref: group-v285980. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 800.317500] env[62066]: DEBUG nova.virt.hardware [None req-df48af02-4c4b-4364-8d38-efc343bf8241 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-23T13:40:41Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-23T13:40:23Z,direct_url=,disk_format='vmdk',id=50ff584c-3b50-4395-af07-3e66769bc9f7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='eb52f7069a374c61ae946f052007c6d9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-23T13:40:24Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 800.317911] env[62066]: DEBUG nova.virt.hardware [None req-df48af02-4c4b-4364-8d38-efc343bf8241 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 800.317911] env[62066]: DEBUG nova.virt.hardware [None req-df48af02-4c4b-4364-8d38-efc343bf8241 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 800.318067] env[62066]: DEBUG nova.virt.hardware [None req-df48af02-4c4b-4364-8d38-efc343bf8241 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 800.318222] env[62066]: DEBUG nova.virt.hardware [None req-df48af02-4c4b-4364-8d38-efc343bf8241 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 800.318326] env[62066]: DEBUG nova.virt.hardware [None req-df48af02-4c4b-4364-8d38-efc343bf8241 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 800.318533] env[62066]: DEBUG nova.virt.hardware [None req-df48af02-4c4b-4364-8d38-efc343bf8241 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 800.318685] env[62066]: DEBUG nova.virt.hardware [None req-df48af02-4c4b-4364-8d38-efc343bf8241 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 800.319066] env[62066]: DEBUG nova.virt.hardware [None req-df48af02-4c4b-4364-8d38-efc343bf8241 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 800.319066] env[62066]: DEBUG nova.virt.hardware [None req-df48af02-4c4b-4364-8d38-efc343bf8241 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 800.319259] env[62066]: DEBUG nova.virt.hardware [None req-df48af02-4c4b-4364-8d38-efc343bf8241 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 800.319622] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-68cd6e15-e7f0-49b4-b955-8760b4f31370 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.321857] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ec6769f-8a8c-4b88-9bb6-191e4ac1ea92 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.332280] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9394f69-0543-4010-9589-983aa5d0f5af {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.337314] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-febd07a3-de81-4924-888e-f9c7d8ba2329 tempest-ServerMetadataNegativeTestJSON-1549470887 tempest-ServerMetadataNegativeTestJSON-1549470887-project-member] Created folder: Project (47db3c06473c46a5a28865c0756ea745) in parent group-v285980. [ 800.337518] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-febd07a3-de81-4924-888e-f9c7d8ba2329 tempest-ServerMetadataNegativeTestJSON-1549470887 tempest-ServerMetadataNegativeTestJSON-1549470887-project-member] Creating folder: Instances. Parent ref: group-v286022. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 800.338123] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-da1bf522-b781-4f44-9189-385fe6b8228b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.355719] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-febd07a3-de81-4924-888e-f9c7d8ba2329 tempest-ServerMetadataNegativeTestJSON-1549470887 tempest-ServerMetadataNegativeTestJSON-1549470887-project-member] Created folder: Instances in parent group-v286022. [ 800.355979] env[62066]: DEBUG oslo.service.loopingcall [None req-febd07a3-de81-4924-888e-f9c7d8ba2329 tempest-ServerMetadataNegativeTestJSON-1549470887 tempest-ServerMetadataNegativeTestJSON-1549470887-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 800.356257] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 31d47299-83a3-4f27-aeb7-95c4cd36c5ac] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 800.356506] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2a7c15ea-c233-4cd8-bdaf-39057338a7ab {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.376336] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 800.376336] env[62066]: value = "task-1340751" [ 800.376336] env[62066]: _type = "Task" [ 800.376336] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 800.384348] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1340751, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.448827] env[62066]: DEBUG oslo_vmware.api [None req-c70bcb7f-78ff-424d-8121-684c45ecd12b tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': task-1340748, 'name': ReconfigVM_Task, 'duration_secs': 0.448027} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 800.449191] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-c70bcb7f-78ff-424d-8121-684c45ecd12b tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 02fb3e92-5dd6-4b1e-a6e4-d60d3fc85b07] Reconfigured VM instance instance-0000003d to attach disk [datastore2] 02fb3e92-5dd6-4b1e-a6e4-d60d3fc85b07/02fb3e92-5dd6-4b1e-a6e4-d60d3fc85b07.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 800.449890] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0afece47-67dc-4bff-959f-374263acf340 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.456796] env[62066]: DEBUG oslo_vmware.api [None req-c70bcb7f-78ff-424d-8121-684c45ecd12b tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Waiting for the task: (returnval){ [ 800.456796] env[62066]: value = "task-1340752" [ 800.456796] env[62066]: _type = "Task" [ 800.456796] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 800.465534] env[62066]: DEBUG oslo_vmware.api [None req-c70bcb7f-78ff-424d-8121-684c45ecd12b tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': task-1340752, 'name': Rename_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.544460] env[62066]: DEBUG oslo_vmware.api [None req-dcd99ffb-af75-402a-906d-756c3b1954c8 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Task: {'id': task-1340747, 'name': CloneVM_Task} progress is 94%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.565648] env[62066]: DEBUG nova.compute.manager [req-6befd109-6353-41c7-8799-a370f2382812 req-2e5058a8-b088-48b9-a99c-d1db2ade8358 service nova] [instance: 31d47299-83a3-4f27-aeb7-95c4cd36c5ac] Received event network-changed-4d836af3-8ac2-4b67-842b-c868c679bd5e {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 800.565920] env[62066]: DEBUG nova.compute.manager [req-6befd109-6353-41c7-8799-a370f2382812 req-2e5058a8-b088-48b9-a99c-d1db2ade8358 service nova] [instance: 31d47299-83a3-4f27-aeb7-95c4cd36c5ac] Refreshing instance network info cache due to event network-changed-4d836af3-8ac2-4b67-842b-c868c679bd5e. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 800.566191] env[62066]: DEBUG oslo_concurrency.lockutils [req-6befd109-6353-41c7-8799-a370f2382812 req-2e5058a8-b088-48b9-a99c-d1db2ade8358 service nova] Acquiring lock "refresh_cache-31d47299-83a3-4f27-aeb7-95c4cd36c5ac" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 800.566346] env[62066]: DEBUG oslo_concurrency.lockutils [req-6befd109-6353-41c7-8799-a370f2382812 req-2e5058a8-b088-48b9-a99c-d1db2ade8358 service nova] Acquired lock "refresh_cache-31d47299-83a3-4f27-aeb7-95c4cd36c5ac" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 800.566563] env[62066]: DEBUG nova.network.neutron [req-6befd109-6353-41c7-8799-a370f2382812 req-2e5058a8-b088-48b9-a99c-d1db2ade8358 service nova] [instance: 31d47299-83a3-4f27-aeb7-95c4cd36c5ac] Refreshing network info cache for port 4d836af3-8ac2-4b67-842b-c868c679bd5e {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 800.641697] env[62066]: DEBUG oslo_concurrency.lockutils [None req-8034bd7f-2294-4f4c-b924-f35bec7ebaac tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.385s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 800.646019] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f6cc809d-278e-4fb4-906b-f468fa41739d tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.725s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 800.646019] env[62066]: INFO nova.compute.claims [None req-f6cc809d-278e-4fb4-906b-f468fa41739d tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] [instance: 5f645eaa-6edc-4362-82ad-38c4c57b2be2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 800.684842] env[62066]: DEBUG oslo_vmware.api [None req-816774ed-d63b-4dd8-94be-e4c70c8f10e7 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52ac090f-0c20-f6c9-5298-d378b3be2cf8, 'name': SearchDatastore_Task, 'duration_secs': 0.018568} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 800.685804] env[62066]: DEBUG nova.network.neutron [req-170d002b-1f49-47dd-8443-ad3de69f6baf req-ef457172-7257-41b7-a4e4-9fac2692fd20 service nova] [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] Updated VIF entry in instance network info cache for port 88380b24-b187-4b82-b92e-b93adeeab3eb. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 800.686651] env[62066]: DEBUG nova.network.neutron [req-170d002b-1f49-47dd-8443-ad3de69f6baf req-ef457172-7257-41b7-a4e4-9fac2692fd20 service nova] [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] Updating instance_info_cache with network_info: [{"id": "88380b24-b187-4b82-b92e-b93adeeab3eb", "address": "fa:16:3e:5c:2a:4b", "network": {"id": "b08881d1-c195-4165-9400-8c2012516c9e", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-1637068010-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.160", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0b8c246df0e8404ca3743c7207cfc808", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e1c9bb98-73a9-48eb-856e-a541afe9b07b", "external-id": "nsx-vlan-transportzone-755", "segmentation_id": 755, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap88380b24-b1", "ovs_interfaceid": "88380b24-b187-4b82-b92e-b93adeeab3eb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 800.687528] env[62066]: DEBUG oslo_concurrency.lockutils [None req-816774ed-d63b-4dd8-94be-e4c70c8f10e7 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Releasing lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 800.687746] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-816774ed-d63b-4dd8-94be-e4c70c8f10e7 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk to [datastore2] 459f432b-e817-405f-9c3e-dfa5d957788c/459f432b-e817-405f-9c3e-dfa5d957788c.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 800.688249] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f799e259-7aeb-4387-87a1-a8835e9c8681 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.695894] env[62066]: DEBUG oslo_vmware.api [None req-816774ed-d63b-4dd8-94be-e4c70c8f10e7 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Waiting for the task: (returnval){ [ 800.695894] env[62066]: value = "task-1340753" [ 800.695894] env[62066]: _type = "Task" [ 800.695894] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 800.707187] env[62066]: DEBUG oslo_vmware.api [None req-816774ed-d63b-4dd8-94be-e4c70c8f10e7 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': task-1340753, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.789041] env[62066]: DEBUG nova.network.neutron [None req-df48af02-4c4b-4364-8d38-efc343bf8241 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 49f46244-34fa-48a1-95a2-8e95850f345d] Successfully updated port: 213f129a-7b70-4807-ac40-621d02749750 {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 800.888250] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1340751, 'name': CreateVM_Task, 'duration_secs': 0.457165} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 800.888250] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 31d47299-83a3-4f27-aeb7-95c4cd36c5ac] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 800.888250] env[62066]: DEBUG oslo_concurrency.lockutils [None req-febd07a3-de81-4924-888e-f9c7d8ba2329 tempest-ServerMetadataNegativeTestJSON-1549470887 tempest-ServerMetadataNegativeTestJSON-1549470887-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 800.888250] env[62066]: DEBUG oslo_concurrency.lockutils [None req-febd07a3-de81-4924-888e-f9c7d8ba2329 tempest-ServerMetadataNegativeTestJSON-1549470887 tempest-ServerMetadataNegativeTestJSON-1549470887-project-member] Acquired lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 800.888708] env[62066]: DEBUG oslo_concurrency.lockutils [None req-febd07a3-de81-4924-888e-f9c7d8ba2329 tempest-ServerMetadataNegativeTestJSON-1549470887 tempest-ServerMetadataNegativeTestJSON-1549470887-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 800.888883] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6b9c95ba-ea51-4719-b0c2-1374c977b883 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.893936] env[62066]: DEBUG oslo_vmware.api [None req-febd07a3-de81-4924-888e-f9c7d8ba2329 tempest-ServerMetadataNegativeTestJSON-1549470887 tempest-ServerMetadataNegativeTestJSON-1549470887-project-member] Waiting for the task: (returnval){ [ 800.893936] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]52c6a1c7-4cf0-ea0a-da1f-9234ba5981f6" [ 800.893936] env[62066]: _type = "Task" [ 800.893936] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 800.902511] env[62066]: DEBUG oslo_vmware.api [None req-febd07a3-de81-4924-888e-f9c7d8ba2329 tempest-ServerMetadataNegativeTestJSON-1549470887 tempest-ServerMetadataNegativeTestJSON-1549470887-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52c6a1c7-4cf0-ea0a-da1f-9234ba5981f6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.966257] env[62066]: DEBUG oslo_vmware.api [None req-c70bcb7f-78ff-424d-8121-684c45ecd12b tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': task-1340752, 'name': Rename_Task, 'duration_secs': 0.146803} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 800.966549] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-c70bcb7f-78ff-424d-8121-684c45ecd12b tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 02fb3e92-5dd6-4b1e-a6e4-d60d3fc85b07] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 800.966798] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d605e8dd-b8fb-408a-a7ab-72617a004ef0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.974152] env[62066]: DEBUG oslo_vmware.api [None req-c70bcb7f-78ff-424d-8121-684c45ecd12b tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Waiting for the task: (returnval){ [ 800.974152] env[62066]: value = "task-1340754" [ 800.974152] env[62066]: _type = "Task" [ 800.974152] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 800.982266] env[62066]: DEBUG oslo_vmware.api [None req-c70bcb7f-78ff-424d-8121-684c45ecd12b tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': task-1340754, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.046152] env[62066]: DEBUG oslo_vmware.api [None req-dcd99ffb-af75-402a-906d-756c3b1954c8 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Task: {'id': task-1340747, 'name': CloneVM_Task, 'duration_secs': 1.381057} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.046417] env[62066]: INFO nova.virt.vmwareapi.vmops [None req-dcd99ffb-af75-402a-906d-756c3b1954c8 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] [instance: 1a4b0637-1a56-41ef-b89b-6b56d24ed206] Created linked-clone VM from snapshot [ 801.047318] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42678b34-5134-4b73-ba7e-9a4282769571 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.056813] env[62066]: DEBUG nova.virt.vmwareapi.images [None req-dcd99ffb-af75-402a-906d-756c3b1954c8 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] [instance: 1a4b0637-1a56-41ef-b89b-6b56d24ed206] Uploading image ce0c3df1-323d-4736-9213-367dbe18cd1e {{(pid=62066) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 801.081033] env[62066]: DEBUG oslo_vmware.rw_handles [None req-dcd99ffb-af75-402a-906d-756c3b1954c8 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 801.081033] env[62066]: value = "vm-286021" [ 801.081033] env[62066]: _type = "VirtualMachine" [ 801.081033] env[62066]: }. {{(pid=62066) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 801.081434] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-57b5e68b-725f-4afd-98b4-f7586d47aeab {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.089922] env[62066]: DEBUG oslo_vmware.rw_handles [None req-dcd99ffb-af75-402a-906d-756c3b1954c8 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Lease: (returnval){ [ 801.089922] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]52cda7e5-a8a6-1fbe-18d8-8ddc8312c7ce" [ 801.089922] env[62066]: _type = "HttpNfcLease" [ 801.089922] env[62066]: } obtained for exporting VM: (result){ [ 801.089922] env[62066]: value = "vm-286021" [ 801.089922] env[62066]: _type = "VirtualMachine" [ 801.089922] env[62066]: }. {{(pid=62066) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 801.090653] env[62066]: DEBUG oslo_vmware.api [None req-dcd99ffb-af75-402a-906d-756c3b1954c8 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Waiting for the lease: (returnval){ [ 801.090653] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]52cda7e5-a8a6-1fbe-18d8-8ddc8312c7ce" [ 801.090653] env[62066]: _type = "HttpNfcLease" [ 801.090653] env[62066]: } to be ready. {{(pid=62066) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 801.097941] env[62066]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 801.097941] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]52cda7e5-a8a6-1fbe-18d8-8ddc8312c7ce" [ 801.097941] env[62066]: _type = "HttpNfcLease" [ 801.097941] env[62066]: } is initializing. {{(pid=62066) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 801.150045] env[62066]: DEBUG oslo_concurrency.lockutils [None req-8034bd7f-2294-4f4c-b924-f35bec7ebaac tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] Acquiring lock "efdc3578-4775-4f17-b796-232953e9d2a2" by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 801.150045] env[62066]: DEBUG oslo_concurrency.lockutils [None req-8034bd7f-2294-4f4c-b924-f35bec7ebaac tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] Lock "efdc3578-4775-4f17-b796-232953e9d2a2" acquired by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 801.188875] env[62066]: DEBUG oslo_concurrency.lockutils [req-170d002b-1f49-47dd-8443-ad3de69f6baf req-ef457172-7257-41b7-a4e4-9fac2692fd20 service nova] Releasing lock "refresh_cache-daffaf51-4c45-44aa-8fc2-4db066a09971" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 801.206751] env[62066]: DEBUG oslo_vmware.api [None req-816774ed-d63b-4dd8-94be-e4c70c8f10e7 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': task-1340753, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.291457] env[62066]: DEBUG oslo_concurrency.lockutils [None req-df48af02-4c4b-4364-8d38-efc343bf8241 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Acquiring lock "refresh_cache-49f46244-34fa-48a1-95a2-8e95850f345d" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 801.291651] env[62066]: DEBUG oslo_concurrency.lockutils [None req-df48af02-4c4b-4364-8d38-efc343bf8241 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Acquired lock "refresh_cache-49f46244-34fa-48a1-95a2-8e95850f345d" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 801.291776] env[62066]: DEBUG nova.network.neutron [None req-df48af02-4c4b-4364-8d38-efc343bf8241 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 49f46244-34fa-48a1-95a2-8e95850f345d] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 801.327422] env[62066]: DEBUG nova.network.neutron [req-6befd109-6353-41c7-8799-a370f2382812 req-2e5058a8-b088-48b9-a99c-d1db2ade8358 service nova] [instance: 31d47299-83a3-4f27-aeb7-95c4cd36c5ac] Updated VIF entry in instance network info cache for port 4d836af3-8ac2-4b67-842b-c868c679bd5e. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 801.327776] env[62066]: DEBUG nova.network.neutron [req-6befd109-6353-41c7-8799-a370f2382812 req-2e5058a8-b088-48b9-a99c-d1db2ade8358 service nova] [instance: 31d47299-83a3-4f27-aeb7-95c4cd36c5ac] Updating instance_info_cache with network_info: [{"id": "4d836af3-8ac2-4b67-842b-c868c679bd5e", "address": "fa:16:3e:f3:45:37", "network": {"id": "6b1d3873-7d3c-42ef-a70c-df94aac1e6eb", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-689803675-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "47db3c06473c46a5a28865c0756ea745", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1323cb03-8367-485a-962e-131af8eba474", "external-id": "nsx-vlan-transportzone-41", "segmentation_id": 41, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4d836af3-8a", "ovs_interfaceid": "4d836af3-8ac2-4b67-842b-c868c679bd5e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 801.404071] env[62066]: DEBUG oslo_vmware.api [None req-febd07a3-de81-4924-888e-f9c7d8ba2329 tempest-ServerMetadataNegativeTestJSON-1549470887 tempest-ServerMetadataNegativeTestJSON-1549470887-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52c6a1c7-4cf0-ea0a-da1f-9234ba5981f6, 'name': SearchDatastore_Task, 'duration_secs': 0.015187} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.404640] env[62066]: DEBUG oslo_concurrency.lockutils [None req-febd07a3-de81-4924-888e-f9c7d8ba2329 tempest-ServerMetadataNegativeTestJSON-1549470887 tempest-ServerMetadataNegativeTestJSON-1549470887-project-member] Releasing lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 801.404872] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-febd07a3-de81-4924-888e-f9c7d8ba2329 tempest-ServerMetadataNegativeTestJSON-1549470887 tempest-ServerMetadataNegativeTestJSON-1549470887-project-member] [instance: 31d47299-83a3-4f27-aeb7-95c4cd36c5ac] Processing image 50ff584c-3b50-4395-af07-3e66769bc9f7 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 801.405111] env[62066]: DEBUG oslo_concurrency.lockutils [None req-febd07a3-de81-4924-888e-f9c7d8ba2329 tempest-ServerMetadataNegativeTestJSON-1549470887 tempest-ServerMetadataNegativeTestJSON-1549470887-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 801.405260] env[62066]: DEBUG oslo_concurrency.lockutils [None req-febd07a3-de81-4924-888e-f9c7d8ba2329 tempest-ServerMetadataNegativeTestJSON-1549470887 tempest-ServerMetadataNegativeTestJSON-1549470887-project-member] Acquired lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 801.405438] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-febd07a3-de81-4924-888e-f9c7d8ba2329 tempest-ServerMetadataNegativeTestJSON-1549470887 tempest-ServerMetadataNegativeTestJSON-1549470887-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 801.405687] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-05aa3081-d81f-4e63-aaba-8e337a77b03b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.417413] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-febd07a3-de81-4924-888e-f9c7d8ba2329 tempest-ServerMetadataNegativeTestJSON-1549470887 tempest-ServerMetadataNegativeTestJSON-1549470887-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 801.417584] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-febd07a3-de81-4924-888e-f9c7d8ba2329 tempest-ServerMetadataNegativeTestJSON-1549470887 tempest-ServerMetadataNegativeTestJSON-1549470887-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 801.418288] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f3a68741-6d87-41e7-8458-76097970351f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.422766] env[62066]: DEBUG oslo_vmware.api [None req-febd07a3-de81-4924-888e-f9c7d8ba2329 tempest-ServerMetadataNegativeTestJSON-1549470887 tempest-ServerMetadataNegativeTestJSON-1549470887-project-member] Waiting for the task: (returnval){ [ 801.422766] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]526faa2a-a5dc-f83f-2a01-384beac1a380" [ 801.422766] env[62066]: _type = "Task" [ 801.422766] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.429579] env[62066]: DEBUG oslo_vmware.api [None req-febd07a3-de81-4924-888e-f9c7d8ba2329 tempest-ServerMetadataNegativeTestJSON-1549470887 tempest-ServerMetadataNegativeTestJSON-1549470887-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]526faa2a-a5dc-f83f-2a01-384beac1a380, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.484759] env[62066]: DEBUG oslo_vmware.api [None req-c70bcb7f-78ff-424d-8121-684c45ecd12b tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': task-1340754, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.598850] env[62066]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 801.598850] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]52cda7e5-a8a6-1fbe-18d8-8ddc8312c7ce" [ 801.598850] env[62066]: _type = "HttpNfcLease" [ 801.598850] env[62066]: } is ready. {{(pid=62066) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 801.599189] env[62066]: DEBUG oslo_vmware.rw_handles [None req-dcd99ffb-af75-402a-906d-756c3b1954c8 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 801.599189] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]52cda7e5-a8a6-1fbe-18d8-8ddc8312c7ce" [ 801.599189] env[62066]: _type = "HttpNfcLease" [ 801.599189] env[62066]: }. {{(pid=62066) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 801.599892] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b0d4c44-d3ae-4999-8c50-8b6cba65b806 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.608055] env[62066]: DEBUG oslo_vmware.rw_handles [None req-dcd99ffb-af75-402a-906d-756c3b1954c8 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/522b566b-f748-bd1a-5c5d-57241d2d64f6/disk-0.vmdk from lease info. {{(pid=62066) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 801.608231] env[62066]: DEBUG oslo_vmware.rw_handles [None req-dcd99ffb-af75-402a-906d-756c3b1954c8 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/522b566b-f748-bd1a-5c5d-57241d2d64f6/disk-0.vmdk for reading. {{(pid=62066) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 801.664710] env[62066]: DEBUG oslo_concurrency.lockutils [None req-8034bd7f-2294-4f4c-b924-f35bec7ebaac tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] Lock "efdc3578-4775-4f17-b796-232953e9d2a2" "released" by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" :: held 0.515s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 801.665323] env[62066]: DEBUG nova.compute.manager [None req-8034bd7f-2294-4f4c-b924-f35bec7ebaac tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] [instance: 738ebb2f-7cc6-4d0f-871c-1428c9c21384] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 801.704906] env[62066]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-56a9fa90-cc26-4c54-ba96-bfe0d478e8ba {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.710921] env[62066]: DEBUG oslo_vmware.api [None req-816774ed-d63b-4dd8-94be-e4c70c8f10e7 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': task-1340753, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.660566} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.713893] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-816774ed-d63b-4dd8-94be-e4c70c8f10e7 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk to [datastore2] 459f432b-e817-405f-9c3e-dfa5d957788c/459f432b-e817-405f-9c3e-dfa5d957788c.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 801.714145] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-816774ed-d63b-4dd8-94be-e4c70c8f10e7 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: 459f432b-e817-405f-9c3e-dfa5d957788c] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 801.714812] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b26edfd6-4393-4409-a8f1-519ff9783a7b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.723615] env[62066]: DEBUG oslo_vmware.api [None req-816774ed-d63b-4dd8-94be-e4c70c8f10e7 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Waiting for the task: (returnval){ [ 801.723615] env[62066]: value = "task-1340756" [ 801.723615] env[62066]: _type = "Task" [ 801.723615] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.735816] env[62066]: DEBUG oslo_vmware.api [None req-816774ed-d63b-4dd8-94be-e4c70c8f10e7 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': task-1340756, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.825326] env[62066]: DEBUG nova.network.neutron [None req-df48af02-4c4b-4364-8d38-efc343bf8241 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 49f46244-34fa-48a1-95a2-8e95850f345d] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 801.830429] env[62066]: DEBUG oslo_concurrency.lockutils [req-6befd109-6353-41c7-8799-a370f2382812 req-2e5058a8-b088-48b9-a99c-d1db2ade8358 service nova] Releasing lock "refresh_cache-31d47299-83a3-4f27-aeb7-95c4cd36c5ac" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 801.934026] env[62066]: DEBUG oslo_vmware.api [None req-febd07a3-de81-4924-888e-f9c7d8ba2329 tempest-ServerMetadataNegativeTestJSON-1549470887 tempest-ServerMetadataNegativeTestJSON-1549470887-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]526faa2a-a5dc-f83f-2a01-384beac1a380, 'name': SearchDatastore_Task, 'duration_secs': 0.057289} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.935142] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-18d4e95e-3c7c-4340-9ffa-f0df9f2a9be5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.947668] env[62066]: DEBUG oslo_vmware.api [None req-febd07a3-de81-4924-888e-f9c7d8ba2329 tempest-ServerMetadataNegativeTestJSON-1549470887 tempest-ServerMetadataNegativeTestJSON-1549470887-project-member] Waiting for the task: (returnval){ [ 801.947668] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]52d2d19d-438f-5169-f62a-56503da9404e" [ 801.947668] env[62066]: _type = "Task" [ 801.947668] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.955838] env[62066]: DEBUG oslo_vmware.api [None req-febd07a3-de81-4924-888e-f9c7d8ba2329 tempest-ServerMetadataNegativeTestJSON-1549470887 tempest-ServerMetadataNegativeTestJSON-1549470887-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52d2d19d-438f-5169-f62a-56503da9404e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.964451] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2ba9e95-0bbf-41b5-9b1a-fedd1befb705 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.969423] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f29bd40c-8e77-478e-a945-402205ed60d0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.004473] env[62066]: DEBUG nova.network.neutron [None req-df48af02-4c4b-4364-8d38-efc343bf8241 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 49f46244-34fa-48a1-95a2-8e95850f345d] Updating instance_info_cache with network_info: [{"id": "213f129a-7b70-4807-ac40-621d02749750", "address": "fa:16:3e:cb:64:95", "network": {"id": "a3852490-7a1f-499c-813a-46cbf52318da", "bridge": "br-int", "label": "tempest-ImagesTestJSON-735840736-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "75f0352c852947369474c9e05766c584", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "50171613-b419-45e3-9ada-fcb6cd921428", "external-id": "nsx-vlan-transportzone-914", "segmentation_id": 914, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap213f129a-7b", "ovs_interfaceid": "213f129a-7b70-4807-ac40-621d02749750", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 802.006647] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b4db12c-e51c-4d5b-b5ab-47065cf39966 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.013689] env[62066]: DEBUG oslo_vmware.api [None req-c70bcb7f-78ff-424d-8121-684c45ecd12b tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': task-1340754, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.018624] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-215ce0e4-8da9-45b8-b36a-a8a05db28c44 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.035015] env[62066]: DEBUG nova.compute.provider_tree [None req-f6cc809d-278e-4fb4-906b-f468fa41739d tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 802.173240] env[62066]: DEBUG nova.compute.utils [None req-8034bd7f-2294-4f4c-b924-f35bec7ebaac tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 802.175301] env[62066]: DEBUG nova.compute.manager [None req-8034bd7f-2294-4f4c-b924-f35bec7ebaac tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] [instance: 738ebb2f-7cc6-4d0f-871c-1428c9c21384] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 802.175629] env[62066]: DEBUG nova.network.neutron [None req-8034bd7f-2294-4f4c-b924-f35bec7ebaac tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] [instance: 738ebb2f-7cc6-4d0f-871c-1428c9c21384] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 802.219328] env[62066]: DEBUG nova.policy [None req-8034bd7f-2294-4f4c-b924-f35bec7ebaac tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ce308adb0ebb4ee8a5d77732ced20e39', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '636da8edfb3a4ad58312b97b1cb23a97', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 802.234643] env[62066]: DEBUG oslo_vmware.api [None req-816774ed-d63b-4dd8-94be-e4c70c8f10e7 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': task-1340756, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067689} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.235191] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-816774ed-d63b-4dd8-94be-e4c70c8f10e7 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: 459f432b-e817-405f-9c3e-dfa5d957788c] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 802.236259] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ea25fbd-edc8-4595-b517-de84494fb934 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.260198] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-816774ed-d63b-4dd8-94be-e4c70c8f10e7 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: 459f432b-e817-405f-9c3e-dfa5d957788c] Reconfiguring VM instance instance-0000003e to attach disk [datastore2] 459f432b-e817-405f-9c3e-dfa5d957788c/459f432b-e817-405f-9c3e-dfa5d957788c.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 802.261339] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e796964d-1bba-4f69-a301-ddd02ca7c7d9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.286641] env[62066]: DEBUG oslo_vmware.api [None req-816774ed-d63b-4dd8-94be-e4c70c8f10e7 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Waiting for the task: (returnval){ [ 802.286641] env[62066]: value = "task-1340757" [ 802.286641] env[62066]: _type = "Task" [ 802.286641] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.295110] env[62066]: DEBUG oslo_vmware.api [None req-816774ed-d63b-4dd8-94be-e4c70c8f10e7 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': task-1340757, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.462852] env[62066]: DEBUG oslo_vmware.api [None req-febd07a3-de81-4924-888e-f9c7d8ba2329 tempest-ServerMetadataNegativeTestJSON-1549470887 tempest-ServerMetadataNegativeTestJSON-1549470887-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52d2d19d-438f-5169-f62a-56503da9404e, 'name': SearchDatastore_Task, 'duration_secs': 0.014152} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.463321] env[62066]: DEBUG oslo_concurrency.lockutils [None req-febd07a3-de81-4924-888e-f9c7d8ba2329 tempest-ServerMetadataNegativeTestJSON-1549470887 tempest-ServerMetadataNegativeTestJSON-1549470887-project-member] Releasing lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 802.463705] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-febd07a3-de81-4924-888e-f9c7d8ba2329 tempest-ServerMetadataNegativeTestJSON-1549470887 tempest-ServerMetadataNegativeTestJSON-1549470887-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk to [datastore2] 31d47299-83a3-4f27-aeb7-95c4cd36c5ac/31d47299-83a3-4f27-aeb7-95c4cd36c5ac.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 802.464059] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9f75805d-da45-43bf-ae1c-5cd91f62531b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.472328] env[62066]: DEBUG oslo_vmware.api [None req-febd07a3-de81-4924-888e-f9c7d8ba2329 tempest-ServerMetadataNegativeTestJSON-1549470887 tempest-ServerMetadataNegativeTestJSON-1549470887-project-member] Waiting for the task: (returnval){ [ 802.472328] env[62066]: value = "task-1340758" [ 802.472328] env[62066]: _type = "Task" [ 802.472328] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.485246] env[62066]: DEBUG oslo_vmware.api [None req-febd07a3-de81-4924-888e-f9c7d8ba2329 tempest-ServerMetadataNegativeTestJSON-1549470887 tempest-ServerMetadataNegativeTestJSON-1549470887-project-member] Task: {'id': task-1340758, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.488726] env[62066]: DEBUG oslo_vmware.api [None req-c70bcb7f-78ff-424d-8121-684c45ecd12b tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': task-1340754, 'name': PowerOnVM_Task, 'duration_secs': 1.187773} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.489572] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-c70bcb7f-78ff-424d-8121-684c45ecd12b tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 02fb3e92-5dd6-4b1e-a6e4-d60d3fc85b07] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 802.490250] env[62066]: INFO nova.compute.manager [None req-c70bcb7f-78ff-424d-8121-684c45ecd12b tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 02fb3e92-5dd6-4b1e-a6e4-d60d3fc85b07] Took 9.40 seconds to spawn the instance on the hypervisor. [ 802.491202] env[62066]: DEBUG nova.compute.manager [None req-c70bcb7f-78ff-424d-8121-684c45ecd12b tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 02fb3e92-5dd6-4b1e-a6e4-d60d3fc85b07] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 802.494079] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30119208-47ac-4472-ad31-e4a1d034c558 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.514362] env[62066]: DEBUG oslo_concurrency.lockutils [None req-df48af02-4c4b-4364-8d38-efc343bf8241 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Releasing lock "refresh_cache-49f46244-34fa-48a1-95a2-8e95850f345d" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 802.514697] env[62066]: DEBUG nova.compute.manager [None req-df48af02-4c4b-4364-8d38-efc343bf8241 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 49f46244-34fa-48a1-95a2-8e95850f345d] Instance network_info: |[{"id": "213f129a-7b70-4807-ac40-621d02749750", "address": "fa:16:3e:cb:64:95", "network": {"id": "a3852490-7a1f-499c-813a-46cbf52318da", "bridge": "br-int", "label": "tempest-ImagesTestJSON-735840736-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "75f0352c852947369474c9e05766c584", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "50171613-b419-45e3-9ada-fcb6cd921428", "external-id": "nsx-vlan-transportzone-914", "segmentation_id": 914, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap213f129a-7b", "ovs_interfaceid": "213f129a-7b70-4807-ac40-621d02749750", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 802.515183] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-df48af02-4c4b-4364-8d38-efc343bf8241 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 49f46244-34fa-48a1-95a2-8e95850f345d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:cb:64:95', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '50171613-b419-45e3-9ada-fcb6cd921428', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '213f129a-7b70-4807-ac40-621d02749750', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 802.525398] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-df48af02-4c4b-4364-8d38-efc343bf8241 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Creating folder: Project (75f0352c852947369474c9e05766c584). Parent ref: group-v285980. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 802.528329] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b956defa-a34f-4617-b134-bb5149921620 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.536167] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-df48af02-4c4b-4364-8d38-efc343bf8241 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Created folder: Project (75f0352c852947369474c9e05766c584) in parent group-v285980. [ 802.536395] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-df48af02-4c4b-4364-8d38-efc343bf8241 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Creating folder: Instances. Parent ref: group-v286025. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 802.537096] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-01b37256-5b90-4004-9525-34338a060945 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.540818] env[62066]: DEBUG nova.scheduler.client.report [None req-f6cc809d-278e-4fb4-906b-f468fa41739d tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 802.546037] env[62066]: DEBUG nova.network.neutron [None req-8034bd7f-2294-4f4c-b924-f35bec7ebaac tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] [instance: 738ebb2f-7cc6-4d0f-871c-1428c9c21384] Successfully created port: fa65affa-3b23-4d64-a714-83d41caa3293 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 802.555247] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-df48af02-4c4b-4364-8d38-efc343bf8241 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Created folder: Instances in parent group-v286025. [ 802.555247] env[62066]: DEBUG oslo.service.loopingcall [None req-df48af02-4c4b-4364-8d38-efc343bf8241 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 802.555402] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 49f46244-34fa-48a1-95a2-8e95850f345d] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 802.556520] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1818314c-fc63-49e8-84ab-0f905e994e4d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.584931] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 802.584931] env[62066]: value = "task-1340761" [ 802.584931] env[62066]: _type = "Task" [ 802.584931] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.595238] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1340761, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.682156] env[62066]: DEBUG nova.compute.manager [None req-8034bd7f-2294-4f4c-b924-f35bec7ebaac tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] [instance: 738ebb2f-7cc6-4d0f-871c-1428c9c21384] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 802.684954] env[62066]: DEBUG nova.compute.manager [req-253eccc3-9577-4e26-ad62-66d658ec5f7f req-05f5e28c-0c57-42dc-9c45-45941054fb0d service nova] [instance: 49f46244-34fa-48a1-95a2-8e95850f345d] Received event network-vif-plugged-213f129a-7b70-4807-ac40-621d02749750 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 802.685304] env[62066]: DEBUG oslo_concurrency.lockutils [req-253eccc3-9577-4e26-ad62-66d658ec5f7f req-05f5e28c-0c57-42dc-9c45-45941054fb0d service nova] Acquiring lock "49f46244-34fa-48a1-95a2-8e95850f345d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 802.685631] env[62066]: DEBUG oslo_concurrency.lockutils [req-253eccc3-9577-4e26-ad62-66d658ec5f7f req-05f5e28c-0c57-42dc-9c45-45941054fb0d service nova] Lock "49f46244-34fa-48a1-95a2-8e95850f345d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 802.685914] env[62066]: DEBUG oslo_concurrency.lockutils [req-253eccc3-9577-4e26-ad62-66d658ec5f7f req-05f5e28c-0c57-42dc-9c45-45941054fb0d service nova] Lock "49f46244-34fa-48a1-95a2-8e95850f345d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 802.686610] env[62066]: DEBUG nova.compute.manager [req-253eccc3-9577-4e26-ad62-66d658ec5f7f req-05f5e28c-0c57-42dc-9c45-45941054fb0d service nova] [instance: 49f46244-34fa-48a1-95a2-8e95850f345d] No waiting events found dispatching network-vif-plugged-213f129a-7b70-4807-ac40-621d02749750 {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 802.687278] env[62066]: WARNING nova.compute.manager [req-253eccc3-9577-4e26-ad62-66d658ec5f7f req-05f5e28c-0c57-42dc-9c45-45941054fb0d service nova] [instance: 49f46244-34fa-48a1-95a2-8e95850f345d] Received unexpected event network-vif-plugged-213f129a-7b70-4807-ac40-621d02749750 for instance with vm_state building and task_state spawning. [ 802.687573] env[62066]: DEBUG nova.compute.manager [req-253eccc3-9577-4e26-ad62-66d658ec5f7f req-05f5e28c-0c57-42dc-9c45-45941054fb0d service nova] [instance: 49f46244-34fa-48a1-95a2-8e95850f345d] Received event network-changed-213f129a-7b70-4807-ac40-621d02749750 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 802.687930] env[62066]: DEBUG nova.compute.manager [req-253eccc3-9577-4e26-ad62-66d658ec5f7f req-05f5e28c-0c57-42dc-9c45-45941054fb0d service nova] [instance: 49f46244-34fa-48a1-95a2-8e95850f345d] Refreshing instance network info cache due to event network-changed-213f129a-7b70-4807-ac40-621d02749750. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 802.688242] env[62066]: DEBUG oslo_concurrency.lockutils [req-253eccc3-9577-4e26-ad62-66d658ec5f7f req-05f5e28c-0c57-42dc-9c45-45941054fb0d service nova] Acquiring lock "refresh_cache-49f46244-34fa-48a1-95a2-8e95850f345d" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 802.688541] env[62066]: DEBUG oslo_concurrency.lockutils [req-253eccc3-9577-4e26-ad62-66d658ec5f7f req-05f5e28c-0c57-42dc-9c45-45941054fb0d service nova] Acquired lock "refresh_cache-49f46244-34fa-48a1-95a2-8e95850f345d" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 802.688824] env[62066]: DEBUG nova.network.neutron [req-253eccc3-9577-4e26-ad62-66d658ec5f7f req-05f5e28c-0c57-42dc-9c45-45941054fb0d service nova] [instance: 49f46244-34fa-48a1-95a2-8e95850f345d] Refreshing network info cache for port 213f129a-7b70-4807-ac40-621d02749750 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 802.798201] env[62066]: DEBUG oslo_vmware.api [None req-816774ed-d63b-4dd8-94be-e4c70c8f10e7 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': task-1340757, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.985286] env[62066]: DEBUG oslo_vmware.api [None req-febd07a3-de81-4924-888e-f9c7d8ba2329 tempest-ServerMetadataNegativeTestJSON-1549470887 tempest-ServerMetadataNegativeTestJSON-1549470887-project-member] Task: {'id': task-1340758, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.013918] env[62066]: INFO nova.compute.manager [None req-c70bcb7f-78ff-424d-8121-684c45ecd12b tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 02fb3e92-5dd6-4b1e-a6e4-d60d3fc85b07] Took 28.86 seconds to build instance. [ 803.049263] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f6cc809d-278e-4fb4-906b-f468fa41739d tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.406s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 803.049965] env[62066]: DEBUG nova.compute.manager [None req-f6cc809d-278e-4fb4-906b-f468fa41739d tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] [instance: 5f645eaa-6edc-4362-82ad-38c4c57b2be2] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 803.053605] env[62066]: DEBUG oslo_concurrency.lockutils [None req-90f5cbab-7c0d-4b69-92aa-09bf24d9f7da tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.797s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 803.055773] env[62066]: INFO nova.compute.claims [None req-90f5cbab-7c0d-4b69-92aa-09bf24d9f7da tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] [instance: 050040cb-6ee3-4ad2-960a-fcebb53ac394] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 803.104078] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1340761, 'name': CreateVM_Task} progress is 99%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.298309] env[62066]: DEBUG oslo_vmware.api [None req-816774ed-d63b-4dd8-94be-e4c70c8f10e7 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': task-1340757, 'name': ReconfigVM_Task, 'duration_secs': 0.701238} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.299095] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-816774ed-d63b-4dd8-94be-e4c70c8f10e7 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: 459f432b-e817-405f-9c3e-dfa5d957788c] Reconfigured VM instance instance-0000003e to attach disk [datastore2] 459f432b-e817-405f-9c3e-dfa5d957788c/459f432b-e817-405f-9c3e-dfa5d957788c.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 803.299828] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-75284730-d491-4f4e-a905-7ad7b36c29f5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.306391] env[62066]: DEBUG oslo_vmware.api [None req-816774ed-d63b-4dd8-94be-e4c70c8f10e7 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Waiting for the task: (returnval){ [ 803.306391] env[62066]: value = "task-1340762" [ 803.306391] env[62066]: _type = "Task" [ 803.306391] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.315261] env[62066]: DEBUG oslo_vmware.api [None req-816774ed-d63b-4dd8-94be-e4c70c8f10e7 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': task-1340762, 'name': Rename_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.420242] env[62066]: DEBUG nova.network.neutron [req-253eccc3-9577-4e26-ad62-66d658ec5f7f req-05f5e28c-0c57-42dc-9c45-45941054fb0d service nova] [instance: 49f46244-34fa-48a1-95a2-8e95850f345d] Updated VIF entry in instance network info cache for port 213f129a-7b70-4807-ac40-621d02749750. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 803.420751] env[62066]: DEBUG nova.network.neutron [req-253eccc3-9577-4e26-ad62-66d658ec5f7f req-05f5e28c-0c57-42dc-9c45-45941054fb0d service nova] [instance: 49f46244-34fa-48a1-95a2-8e95850f345d] Updating instance_info_cache with network_info: [{"id": "213f129a-7b70-4807-ac40-621d02749750", "address": "fa:16:3e:cb:64:95", "network": {"id": "a3852490-7a1f-499c-813a-46cbf52318da", "bridge": "br-int", "label": "tempest-ImagesTestJSON-735840736-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "75f0352c852947369474c9e05766c584", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "50171613-b419-45e3-9ada-fcb6cd921428", "external-id": "nsx-vlan-transportzone-914", "segmentation_id": 914, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap213f129a-7b", "ovs_interfaceid": "213f129a-7b70-4807-ac40-621d02749750", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 803.483676] env[62066]: DEBUG oslo_vmware.api [None req-febd07a3-de81-4924-888e-f9c7d8ba2329 tempest-ServerMetadataNegativeTestJSON-1549470887 tempest-ServerMetadataNegativeTestJSON-1549470887-project-member] Task: {'id': task-1340758, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.554024} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.483960] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-febd07a3-de81-4924-888e-f9c7d8ba2329 tempest-ServerMetadataNegativeTestJSON-1549470887 tempest-ServerMetadataNegativeTestJSON-1549470887-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk to [datastore2] 31d47299-83a3-4f27-aeb7-95c4cd36c5ac/31d47299-83a3-4f27-aeb7-95c4cd36c5ac.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 803.484267] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-febd07a3-de81-4924-888e-f9c7d8ba2329 tempest-ServerMetadataNegativeTestJSON-1549470887 tempest-ServerMetadataNegativeTestJSON-1549470887-project-member] [instance: 31d47299-83a3-4f27-aeb7-95c4cd36c5ac] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 803.484537] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ba7971da-9a2b-4909-855e-6edd176f044b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.490877] env[62066]: DEBUG oslo_vmware.api [None req-febd07a3-de81-4924-888e-f9c7d8ba2329 tempest-ServerMetadataNegativeTestJSON-1549470887 tempest-ServerMetadataNegativeTestJSON-1549470887-project-member] Waiting for the task: (returnval){ [ 803.490877] env[62066]: value = "task-1340763" [ 803.490877] env[62066]: _type = "Task" [ 803.490877] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.503251] env[62066]: DEBUG oslo_vmware.api [None req-febd07a3-de81-4924-888e-f9c7d8ba2329 tempest-ServerMetadataNegativeTestJSON-1549470887 tempest-ServerMetadataNegativeTestJSON-1549470887-project-member] Task: {'id': task-1340763, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.515984] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c70bcb7f-78ff-424d-8121-684c45ecd12b tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Lock "02fb3e92-5dd6-4b1e-a6e4-d60d3fc85b07" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 132.832s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 803.555106] env[62066]: DEBUG nova.compute.utils [None req-f6cc809d-278e-4fb4-906b-f468fa41739d tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 803.556849] env[62066]: DEBUG nova.compute.manager [None req-f6cc809d-278e-4fb4-906b-f468fa41739d tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] [instance: 5f645eaa-6edc-4362-82ad-38c4c57b2be2] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 803.557653] env[62066]: DEBUG nova.network.neutron [None req-f6cc809d-278e-4fb4-906b-f468fa41739d tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] [instance: 5f645eaa-6edc-4362-82ad-38c4c57b2be2] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 803.601033] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1340761, 'name': CreateVM_Task, 'duration_secs': 0.592014} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.601033] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 49f46244-34fa-48a1-95a2-8e95850f345d] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 803.601033] env[62066]: DEBUG oslo_concurrency.lockutils [None req-df48af02-4c4b-4364-8d38-efc343bf8241 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 803.601033] env[62066]: DEBUG oslo_concurrency.lockutils [None req-df48af02-4c4b-4364-8d38-efc343bf8241 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Acquired lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 803.601033] env[62066]: DEBUG oslo_concurrency.lockutils [None req-df48af02-4c4b-4364-8d38-efc343bf8241 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 803.601033] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-20983b59-6c7b-4a15-858a-4328303dd7b0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.605803] env[62066]: DEBUG nova.policy [None req-f6cc809d-278e-4fb4-906b-f468fa41739d tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '16c8fcc7c0b140f48716a8a00da46a3b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a9b239f4cb3242d0bd1dcd98a2a3bfc0', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 803.611480] env[62066]: DEBUG oslo_vmware.api [None req-df48af02-4c4b-4364-8d38-efc343bf8241 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Waiting for the task: (returnval){ [ 803.611480] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]52db1a10-9a5e-592f-bfec-ff751d402440" [ 803.611480] env[62066]: _type = "Task" [ 803.611480] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.620867] env[62066]: DEBUG oslo_vmware.api [None req-df48af02-4c4b-4364-8d38-efc343bf8241 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52db1a10-9a5e-592f-bfec-ff751d402440, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.703178] env[62066]: DEBUG nova.compute.manager [None req-8034bd7f-2294-4f4c-b924-f35bec7ebaac tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] [instance: 738ebb2f-7cc6-4d0f-871c-1428c9c21384] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 803.736496] env[62066]: DEBUG nova.virt.hardware [None req-8034bd7f-2294-4f4c-b924-f35bec7ebaac tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-23T13:40:41Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-23T13:40:23Z,direct_url=,disk_format='vmdk',id=50ff584c-3b50-4395-af07-3e66769bc9f7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='eb52f7069a374c61ae946f052007c6d9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-23T13:40:24Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 803.736750] env[62066]: DEBUG nova.virt.hardware [None req-8034bd7f-2294-4f4c-b924-f35bec7ebaac tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 803.736911] env[62066]: DEBUG nova.virt.hardware [None req-8034bd7f-2294-4f4c-b924-f35bec7ebaac tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 803.737124] env[62066]: DEBUG nova.virt.hardware [None req-8034bd7f-2294-4f4c-b924-f35bec7ebaac tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 803.737359] env[62066]: DEBUG nova.virt.hardware [None req-8034bd7f-2294-4f4c-b924-f35bec7ebaac tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 803.737473] env[62066]: DEBUG nova.virt.hardware [None req-8034bd7f-2294-4f4c-b924-f35bec7ebaac tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 803.737691] env[62066]: DEBUG nova.virt.hardware [None req-8034bd7f-2294-4f4c-b924-f35bec7ebaac tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 803.737879] env[62066]: DEBUG nova.virt.hardware [None req-8034bd7f-2294-4f4c-b924-f35bec7ebaac tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 803.738151] env[62066]: DEBUG nova.virt.hardware [None req-8034bd7f-2294-4f4c-b924-f35bec7ebaac tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 803.738398] env[62066]: DEBUG nova.virt.hardware [None req-8034bd7f-2294-4f4c-b924-f35bec7ebaac tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 803.738617] env[62066]: DEBUG nova.virt.hardware [None req-8034bd7f-2294-4f4c-b924-f35bec7ebaac tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 803.739777] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7eee59fd-c653-4cf0-a4eb-6adba77c46ec {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.748622] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98b44c69-bec8-4b00-9d6f-687520cf9836 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.820408] env[62066]: DEBUG oslo_vmware.api [None req-816774ed-d63b-4dd8-94be-e4c70c8f10e7 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': task-1340762, 'name': Rename_Task, 'duration_secs': 0.173494} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.820408] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-816774ed-d63b-4dd8-94be-e4c70c8f10e7 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: 459f432b-e817-405f-9c3e-dfa5d957788c] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 803.820408] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-421d2a22-ca77-4244-bee8-bb805192fba3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.823812] env[62066]: DEBUG oslo_vmware.api [None req-816774ed-d63b-4dd8-94be-e4c70c8f10e7 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Waiting for the task: (returnval){ [ 803.823812] env[62066]: value = "task-1340764" [ 803.823812] env[62066]: _type = "Task" [ 803.823812] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.832564] env[62066]: DEBUG oslo_vmware.api [None req-816774ed-d63b-4dd8-94be-e4c70c8f10e7 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': task-1340764, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.924347] env[62066]: DEBUG oslo_concurrency.lockutils [req-253eccc3-9577-4e26-ad62-66d658ec5f7f req-05f5e28c-0c57-42dc-9c45-45941054fb0d service nova] Releasing lock "refresh_cache-49f46244-34fa-48a1-95a2-8e95850f345d" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 804.001452] env[62066]: DEBUG oslo_vmware.api [None req-febd07a3-de81-4924-888e-f9c7d8ba2329 tempest-ServerMetadataNegativeTestJSON-1549470887 tempest-ServerMetadataNegativeTestJSON-1549470887-project-member] Task: {'id': task-1340763, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065456} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.001764] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-febd07a3-de81-4924-888e-f9c7d8ba2329 tempest-ServerMetadataNegativeTestJSON-1549470887 tempest-ServerMetadataNegativeTestJSON-1549470887-project-member] [instance: 31d47299-83a3-4f27-aeb7-95c4cd36c5ac] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 804.002605] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34dbe065-d110-4d77-9a3c-34b739460c7f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.026044] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-febd07a3-de81-4924-888e-f9c7d8ba2329 tempest-ServerMetadataNegativeTestJSON-1549470887 tempest-ServerMetadataNegativeTestJSON-1549470887-project-member] [instance: 31d47299-83a3-4f27-aeb7-95c4cd36c5ac] Reconfiguring VM instance instance-0000003f to attach disk [datastore2] 31d47299-83a3-4f27-aeb7-95c4cd36c5ac/31d47299-83a3-4f27-aeb7-95c4cd36c5ac.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 804.026514] env[62066]: DEBUG nova.compute.manager [None req-b8836db2-293e-442c-80f2-42d8bc4c2918 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 88daaaa6-f385-4161-bc74-9fffabb1145d] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 804.029664] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1bd235cb-3555-4915-842d-4cc435b1b517 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.051101] env[62066]: DEBUG oslo_vmware.api [None req-febd07a3-de81-4924-888e-f9c7d8ba2329 tempest-ServerMetadataNegativeTestJSON-1549470887 tempest-ServerMetadataNegativeTestJSON-1549470887-project-member] Waiting for the task: (returnval){ [ 804.051101] env[62066]: value = "task-1340765" [ 804.051101] env[62066]: _type = "Task" [ 804.051101] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.057433] env[62066]: DEBUG oslo_vmware.api [None req-febd07a3-de81-4924-888e-f9c7d8ba2329 tempest-ServerMetadataNegativeTestJSON-1549470887 tempest-ServerMetadataNegativeTestJSON-1549470887-project-member] Task: {'id': task-1340765, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.060163] env[62066]: DEBUG nova.compute.manager [None req-f6cc809d-278e-4fb4-906b-f468fa41739d tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] [instance: 5f645eaa-6edc-4362-82ad-38c4c57b2be2] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 804.125371] env[62066]: DEBUG oslo_vmware.api [None req-df48af02-4c4b-4364-8d38-efc343bf8241 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52db1a10-9a5e-592f-bfec-ff751d402440, 'name': SearchDatastore_Task, 'duration_secs': 0.018768} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.128056] env[62066]: DEBUG oslo_concurrency.lockutils [None req-df48af02-4c4b-4364-8d38-efc343bf8241 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Releasing lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 804.128335] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-df48af02-4c4b-4364-8d38-efc343bf8241 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 49f46244-34fa-48a1-95a2-8e95850f345d] Processing image 50ff584c-3b50-4395-af07-3e66769bc9f7 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 804.128619] env[62066]: DEBUG oslo_concurrency.lockutils [None req-df48af02-4c4b-4364-8d38-efc343bf8241 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 804.128814] env[62066]: DEBUG oslo_concurrency.lockutils [None req-df48af02-4c4b-4364-8d38-efc343bf8241 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Acquired lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 804.129038] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-df48af02-4c4b-4364-8d38-efc343bf8241 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 804.129818] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f47fed1b-928f-48dc-9847-27d5346068f6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.139844] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-df48af02-4c4b-4364-8d38-efc343bf8241 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 804.140134] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-df48af02-4c4b-4364-8d38-efc343bf8241 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 804.140941] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cc5e6418-6d2b-4e97-b4a8-5ad401dbe271 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.147754] env[62066]: DEBUG oslo_vmware.api [None req-df48af02-4c4b-4364-8d38-efc343bf8241 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Waiting for the task: (returnval){ [ 804.147754] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]52060559-ac9e-c666-488d-b49182a586f0" [ 804.147754] env[62066]: _type = "Task" [ 804.147754] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.155321] env[62066]: DEBUG oslo_vmware.api [None req-df48af02-4c4b-4364-8d38-efc343bf8241 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52060559-ac9e-c666-488d-b49182a586f0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.235385] env[62066]: DEBUG nova.network.neutron [None req-f6cc809d-278e-4fb4-906b-f468fa41739d tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] [instance: 5f645eaa-6edc-4362-82ad-38c4c57b2be2] Successfully created port: 65fcb910-75be-4a68-8cec-476f7cc7cd00 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 804.269209] env[62066]: DEBUG nova.network.neutron [None req-8034bd7f-2294-4f4c-b924-f35bec7ebaac tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] [instance: 738ebb2f-7cc6-4d0f-871c-1428c9c21384] Successfully updated port: fa65affa-3b23-4d64-a714-83d41caa3293 {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 804.336950] env[62066]: DEBUG oslo_vmware.api [None req-816774ed-d63b-4dd8-94be-e4c70c8f10e7 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': task-1340764, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.369596] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6750c710-e598-4ad2-b4e4-e7abf6469fb7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.377797] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6beaa98-34db-4e7a-833c-b9974cabb6c7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.408134] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad5963f3-234f-4301-93bf-19d7ebefa0c5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.415637] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4397a4ee-66e9-44c5-a5f8-95b6ee93dffb {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.429314] env[62066]: DEBUG nova.compute.provider_tree [None req-90f5cbab-7c0d-4b69-92aa-09bf24d9f7da tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 804.560609] env[62066]: DEBUG oslo_vmware.api [None req-febd07a3-de81-4924-888e-f9c7d8ba2329 tempest-ServerMetadataNegativeTestJSON-1549470887 tempest-ServerMetadataNegativeTestJSON-1549470887-project-member] Task: {'id': task-1340765, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.561640] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b8836db2-293e-442c-80f2-42d8bc4c2918 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 804.661526] env[62066]: DEBUG oslo_vmware.api [None req-df48af02-4c4b-4364-8d38-efc343bf8241 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52060559-ac9e-c666-488d-b49182a586f0, 'name': SearchDatastore_Task, 'duration_secs': 0.016244} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.662438] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f9e8fdfb-8d12-466b-ad81-f09abbaca2fc {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.667965] env[62066]: DEBUG oslo_vmware.api [None req-df48af02-4c4b-4364-8d38-efc343bf8241 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Waiting for the task: (returnval){ [ 804.667965] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]52a843d8-6d50-2a9d-43a5-af38ba5ad7b6" [ 804.667965] env[62066]: _type = "Task" [ 804.667965] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.676530] env[62066]: DEBUG oslo_vmware.api [None req-df48af02-4c4b-4364-8d38-efc343bf8241 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52a843d8-6d50-2a9d-43a5-af38ba5ad7b6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.708925] env[62066]: DEBUG nova.compute.manager [req-8e5703a6-8a53-4f03-8f39-11223f0b88b5 req-fa8bd9e8-f0f5-4c15-95ec-67983d6b3d97 service nova] [instance: 738ebb2f-7cc6-4d0f-871c-1428c9c21384] Received event network-vif-plugged-fa65affa-3b23-4d64-a714-83d41caa3293 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 804.709115] env[62066]: DEBUG oslo_concurrency.lockutils [req-8e5703a6-8a53-4f03-8f39-11223f0b88b5 req-fa8bd9e8-f0f5-4c15-95ec-67983d6b3d97 service nova] Acquiring lock "738ebb2f-7cc6-4d0f-871c-1428c9c21384-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 804.709388] env[62066]: DEBUG oslo_concurrency.lockutils [req-8e5703a6-8a53-4f03-8f39-11223f0b88b5 req-fa8bd9e8-f0f5-4c15-95ec-67983d6b3d97 service nova] Lock "738ebb2f-7cc6-4d0f-871c-1428c9c21384-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 804.709578] env[62066]: DEBUG oslo_concurrency.lockutils [req-8e5703a6-8a53-4f03-8f39-11223f0b88b5 req-fa8bd9e8-f0f5-4c15-95ec-67983d6b3d97 service nova] Lock "738ebb2f-7cc6-4d0f-871c-1428c9c21384-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 804.709761] env[62066]: DEBUG nova.compute.manager [req-8e5703a6-8a53-4f03-8f39-11223f0b88b5 req-fa8bd9e8-f0f5-4c15-95ec-67983d6b3d97 service nova] [instance: 738ebb2f-7cc6-4d0f-871c-1428c9c21384] No waiting events found dispatching network-vif-plugged-fa65affa-3b23-4d64-a714-83d41caa3293 {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 804.709954] env[62066]: WARNING nova.compute.manager [req-8e5703a6-8a53-4f03-8f39-11223f0b88b5 req-fa8bd9e8-f0f5-4c15-95ec-67983d6b3d97 service nova] [instance: 738ebb2f-7cc6-4d0f-871c-1428c9c21384] Received unexpected event network-vif-plugged-fa65affa-3b23-4d64-a714-83d41caa3293 for instance with vm_state building and task_state spawning. [ 804.710354] env[62066]: DEBUG nova.compute.manager [req-8e5703a6-8a53-4f03-8f39-11223f0b88b5 req-fa8bd9e8-f0f5-4c15-95ec-67983d6b3d97 service nova] [instance: 738ebb2f-7cc6-4d0f-871c-1428c9c21384] Received event network-changed-fa65affa-3b23-4d64-a714-83d41caa3293 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 804.710564] env[62066]: DEBUG nova.compute.manager [req-8e5703a6-8a53-4f03-8f39-11223f0b88b5 req-fa8bd9e8-f0f5-4c15-95ec-67983d6b3d97 service nova] [instance: 738ebb2f-7cc6-4d0f-871c-1428c9c21384] Refreshing instance network info cache due to event network-changed-fa65affa-3b23-4d64-a714-83d41caa3293. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 804.710762] env[62066]: DEBUG oslo_concurrency.lockutils [req-8e5703a6-8a53-4f03-8f39-11223f0b88b5 req-fa8bd9e8-f0f5-4c15-95ec-67983d6b3d97 service nova] Acquiring lock "refresh_cache-738ebb2f-7cc6-4d0f-871c-1428c9c21384" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 804.710902] env[62066]: DEBUG oslo_concurrency.lockutils [req-8e5703a6-8a53-4f03-8f39-11223f0b88b5 req-fa8bd9e8-f0f5-4c15-95ec-67983d6b3d97 service nova] Acquired lock "refresh_cache-738ebb2f-7cc6-4d0f-871c-1428c9c21384" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 804.711078] env[62066]: DEBUG nova.network.neutron [req-8e5703a6-8a53-4f03-8f39-11223f0b88b5 req-fa8bd9e8-f0f5-4c15-95ec-67983d6b3d97 service nova] [instance: 738ebb2f-7cc6-4d0f-871c-1428c9c21384] Refreshing network info cache for port fa65affa-3b23-4d64-a714-83d41caa3293 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 804.771414] env[62066]: DEBUG oslo_concurrency.lockutils [None req-8034bd7f-2294-4f4c-b924-f35bec7ebaac tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] Acquiring lock "refresh_cache-738ebb2f-7cc6-4d0f-871c-1428c9c21384" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 804.834774] env[62066]: DEBUG oslo_vmware.api [None req-816774ed-d63b-4dd8-94be-e4c70c8f10e7 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': task-1340764, 'name': PowerOnVM_Task, 'duration_secs': 0.805328} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.835247] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-816774ed-d63b-4dd8-94be-e4c70c8f10e7 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: 459f432b-e817-405f-9c3e-dfa5d957788c] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 804.835542] env[62066]: INFO nova.compute.manager [None req-816774ed-d63b-4dd8-94be-e4c70c8f10e7 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: 459f432b-e817-405f-9c3e-dfa5d957788c] Took 9.26 seconds to spawn the instance on the hypervisor. [ 804.835739] env[62066]: DEBUG nova.compute.manager [None req-816774ed-d63b-4dd8-94be-e4c70c8f10e7 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: 459f432b-e817-405f-9c3e-dfa5d957788c] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 804.836632] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ab1ab41-e85a-42cf-b29e-4fc92cdaf1c8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.932828] env[62066]: DEBUG nova.scheduler.client.report [None req-90f5cbab-7c0d-4b69-92aa-09bf24d9f7da tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 804.987820] env[62066]: DEBUG oslo_concurrency.lockutils [None req-02667002-03be-4979-8517-c36490492214 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Acquiring lock "86a92b61-f3e9-48f9-8ee2-756669d558ef" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 804.987991] env[62066]: DEBUG oslo_concurrency.lockutils [None req-02667002-03be-4979-8517-c36490492214 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Lock "86a92b61-f3e9-48f9-8ee2-756669d558ef" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 805.061482] env[62066]: DEBUG oslo_vmware.api [None req-febd07a3-de81-4924-888e-f9c7d8ba2329 tempest-ServerMetadataNegativeTestJSON-1549470887 tempest-ServerMetadataNegativeTestJSON-1549470887-project-member] Task: {'id': task-1340765, 'name': ReconfigVM_Task, 'duration_secs': 0.649947} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.061971] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-febd07a3-de81-4924-888e-f9c7d8ba2329 tempest-ServerMetadataNegativeTestJSON-1549470887 tempest-ServerMetadataNegativeTestJSON-1549470887-project-member] [instance: 31d47299-83a3-4f27-aeb7-95c4cd36c5ac] Reconfigured VM instance instance-0000003f to attach disk [datastore2] 31d47299-83a3-4f27-aeb7-95c4cd36c5ac/31d47299-83a3-4f27-aeb7-95c4cd36c5ac.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 805.062599] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e74361d1-de2d-465c-a9fc-e0968ea31352 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.069556] env[62066]: DEBUG oslo_vmware.api [None req-febd07a3-de81-4924-888e-f9c7d8ba2329 tempest-ServerMetadataNegativeTestJSON-1549470887 tempest-ServerMetadataNegativeTestJSON-1549470887-project-member] Waiting for the task: (returnval){ [ 805.069556] env[62066]: value = "task-1340766" [ 805.069556] env[62066]: _type = "Task" [ 805.069556] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.075126] env[62066]: DEBUG nova.compute.manager [None req-f6cc809d-278e-4fb4-906b-f468fa41739d tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] [instance: 5f645eaa-6edc-4362-82ad-38c4c57b2be2] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 805.080688] env[62066]: DEBUG oslo_vmware.api [None req-febd07a3-de81-4924-888e-f9c7d8ba2329 tempest-ServerMetadataNegativeTestJSON-1549470887 tempest-ServerMetadataNegativeTestJSON-1549470887-project-member] Task: {'id': task-1340766, 'name': Rename_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.100746] env[62066]: DEBUG nova.virt.hardware [None req-f6cc809d-278e-4fb4-906b-f468fa41739d tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-23T13:40:41Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-23T13:40:23Z,direct_url=,disk_format='vmdk',id=50ff584c-3b50-4395-af07-3e66769bc9f7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='eb52f7069a374c61ae946f052007c6d9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-23T13:40:24Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 805.101025] env[62066]: DEBUG nova.virt.hardware [None req-f6cc809d-278e-4fb4-906b-f468fa41739d tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 805.101197] env[62066]: DEBUG nova.virt.hardware [None req-f6cc809d-278e-4fb4-906b-f468fa41739d tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 805.101388] env[62066]: DEBUG nova.virt.hardware [None req-f6cc809d-278e-4fb4-906b-f468fa41739d tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 805.101562] env[62066]: DEBUG nova.virt.hardware [None req-f6cc809d-278e-4fb4-906b-f468fa41739d tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 805.101739] env[62066]: DEBUG nova.virt.hardware [None req-f6cc809d-278e-4fb4-906b-f468fa41739d tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 805.101960] env[62066]: DEBUG nova.virt.hardware [None req-f6cc809d-278e-4fb4-906b-f468fa41739d tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 805.102140] env[62066]: DEBUG nova.virt.hardware [None req-f6cc809d-278e-4fb4-906b-f468fa41739d tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 805.102352] env[62066]: DEBUG nova.virt.hardware [None req-f6cc809d-278e-4fb4-906b-f468fa41739d tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 805.102520] env[62066]: DEBUG nova.virt.hardware [None req-f6cc809d-278e-4fb4-906b-f468fa41739d tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 805.102699] env[62066]: DEBUG nova.virt.hardware [None req-f6cc809d-278e-4fb4-906b-f468fa41739d tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 805.103600] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4b27366-99af-4d01-a5e3-c47a2f6b83fc {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.111506] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15bc9dbc-1505-4851-b372-bd7abecc3923 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.177943] env[62066]: DEBUG oslo_vmware.api [None req-df48af02-4c4b-4364-8d38-efc343bf8241 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52a843d8-6d50-2a9d-43a5-af38ba5ad7b6, 'name': SearchDatastore_Task, 'duration_secs': 0.023632} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.178234] env[62066]: DEBUG oslo_concurrency.lockutils [None req-df48af02-4c4b-4364-8d38-efc343bf8241 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Releasing lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 805.178500] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-df48af02-4c4b-4364-8d38-efc343bf8241 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk to [datastore2] 49f46244-34fa-48a1-95a2-8e95850f345d/49f46244-34fa-48a1-95a2-8e95850f345d.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 805.178761] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1963c4a3-9cb2-43a1-ab82-5c2da01ca06f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.185606] env[62066]: DEBUG oslo_vmware.api [None req-df48af02-4c4b-4364-8d38-efc343bf8241 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Waiting for the task: (returnval){ [ 805.185606] env[62066]: value = "task-1340767" [ 805.185606] env[62066]: _type = "Task" [ 805.185606] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.193565] env[62066]: DEBUG oslo_vmware.api [None req-df48af02-4c4b-4364-8d38-efc343bf8241 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': task-1340767, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.249027] env[62066]: DEBUG nova.network.neutron [req-8e5703a6-8a53-4f03-8f39-11223f0b88b5 req-fa8bd9e8-f0f5-4c15-95ec-67983d6b3d97 service nova] [instance: 738ebb2f-7cc6-4d0f-871c-1428c9c21384] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 805.354058] env[62066]: DEBUG nova.network.neutron [req-8e5703a6-8a53-4f03-8f39-11223f0b88b5 req-fa8bd9e8-f0f5-4c15-95ec-67983d6b3d97 service nova] [instance: 738ebb2f-7cc6-4d0f-871c-1428c9c21384] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 805.358477] env[62066]: INFO nova.compute.manager [None req-816774ed-d63b-4dd8-94be-e4c70c8f10e7 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: 459f432b-e817-405f-9c3e-dfa5d957788c] Took 26.97 seconds to build instance. [ 805.437924] env[62066]: DEBUG oslo_concurrency.lockutils [None req-90f5cbab-7c0d-4b69-92aa-09bf24d9f7da tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.384s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 805.438658] env[62066]: DEBUG nova.compute.manager [None req-90f5cbab-7c0d-4b69-92aa-09bf24d9f7da tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] [instance: 050040cb-6ee3-4ad2-960a-fcebb53ac394] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 805.441820] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0188668d-25e1-420a-be60-fb5771349282 tempest-ServerAddressesTestJSON-1053293248 tempest-ServerAddressesTestJSON-1053293248-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.731s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 805.443303] env[62066]: INFO nova.compute.claims [None req-0188668d-25e1-420a-be60-fb5771349282 tempest-ServerAddressesTestJSON-1053293248 tempest-ServerAddressesTestJSON-1053293248-project-member] [instance: 49bc1350-0095-406b-bc68-005eb6b681a8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 805.580262] env[62066]: DEBUG oslo_vmware.api [None req-febd07a3-de81-4924-888e-f9c7d8ba2329 tempest-ServerMetadataNegativeTestJSON-1549470887 tempest-ServerMetadataNegativeTestJSON-1549470887-project-member] Task: {'id': task-1340766, 'name': Rename_Task, 'duration_secs': 0.27107} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.580573] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-febd07a3-de81-4924-888e-f9c7d8ba2329 tempest-ServerMetadataNegativeTestJSON-1549470887 tempest-ServerMetadataNegativeTestJSON-1549470887-project-member] [instance: 31d47299-83a3-4f27-aeb7-95c4cd36c5ac] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 805.580921] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-94137960-c80d-4d50-a092-4381009f1ab4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.588574] env[62066]: DEBUG oslo_vmware.api [None req-febd07a3-de81-4924-888e-f9c7d8ba2329 tempest-ServerMetadataNegativeTestJSON-1549470887 tempest-ServerMetadataNegativeTestJSON-1549470887-project-member] Waiting for the task: (returnval){ [ 805.588574] env[62066]: value = "task-1340768" [ 805.588574] env[62066]: _type = "Task" [ 805.588574] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.597053] env[62066]: DEBUG oslo_vmware.api [None req-febd07a3-de81-4924-888e-f9c7d8ba2329 tempest-ServerMetadataNegativeTestJSON-1549470887 tempest-ServerMetadataNegativeTestJSON-1549470887-project-member] Task: {'id': task-1340768, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.702211] env[62066]: DEBUG oslo_vmware.api [None req-df48af02-4c4b-4364-8d38-efc343bf8241 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': task-1340767, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.858704] env[62066]: DEBUG oslo_concurrency.lockutils [req-8e5703a6-8a53-4f03-8f39-11223f0b88b5 req-fa8bd9e8-f0f5-4c15-95ec-67983d6b3d97 service nova] Releasing lock "refresh_cache-738ebb2f-7cc6-4d0f-871c-1428c9c21384" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 805.859469] env[62066]: DEBUG oslo_concurrency.lockutils [None req-8034bd7f-2294-4f4c-b924-f35bec7ebaac tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] Acquired lock "refresh_cache-738ebb2f-7cc6-4d0f-871c-1428c9c21384" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 805.859838] env[62066]: DEBUG nova.network.neutron [None req-8034bd7f-2294-4f4c-b924-f35bec7ebaac tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] [instance: 738ebb2f-7cc6-4d0f-871c-1428c9c21384] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 805.861279] env[62066]: DEBUG oslo_concurrency.lockutils [None req-816774ed-d63b-4dd8-94be-e4c70c8f10e7 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Lock "459f432b-e817-405f-9c3e-dfa5d957788c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 128.250s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 805.950023] env[62066]: DEBUG nova.compute.utils [None req-90f5cbab-7c0d-4b69-92aa-09bf24d9f7da tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 805.954218] env[62066]: DEBUG nova.compute.manager [None req-90f5cbab-7c0d-4b69-92aa-09bf24d9f7da tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] [instance: 050040cb-6ee3-4ad2-960a-fcebb53ac394] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 805.954218] env[62066]: DEBUG nova.network.neutron [None req-90f5cbab-7c0d-4b69-92aa-09bf24d9f7da tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] [instance: 050040cb-6ee3-4ad2-960a-fcebb53ac394] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 806.055131] env[62066]: DEBUG nova.policy [None req-90f5cbab-7c0d-4b69-92aa-09bf24d9f7da tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3fe859c5caeb4338a2cd8c6f8cd0653e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b69df8557ba0480893da58e1e8bf591d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 806.102384] env[62066]: DEBUG oslo_vmware.api [None req-febd07a3-de81-4924-888e-f9c7d8ba2329 tempest-ServerMetadataNegativeTestJSON-1549470887 tempest-ServerMetadataNegativeTestJSON-1549470887-project-member] Task: {'id': task-1340768, 'name': PowerOnVM_Task} progress is 1%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.196885] env[62066]: DEBUG oslo_vmware.api [None req-df48af02-4c4b-4364-8d38-efc343bf8241 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': task-1340767, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.693486} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.196885] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-df48af02-4c4b-4364-8d38-efc343bf8241 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk to [datastore2] 49f46244-34fa-48a1-95a2-8e95850f345d/49f46244-34fa-48a1-95a2-8e95850f345d.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 806.197025] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-df48af02-4c4b-4364-8d38-efc343bf8241 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 49f46244-34fa-48a1-95a2-8e95850f345d] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 806.197310] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-491c9e9c-1f2d-4a8c-8e46-290c6230409f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.203983] env[62066]: DEBUG oslo_vmware.api [None req-df48af02-4c4b-4364-8d38-efc343bf8241 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Waiting for the task: (returnval){ [ 806.203983] env[62066]: value = "task-1340769" [ 806.203983] env[62066]: _type = "Task" [ 806.203983] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.208724] env[62066]: DEBUG nova.network.neutron [None req-f6cc809d-278e-4fb4-906b-f468fa41739d tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] [instance: 5f645eaa-6edc-4362-82ad-38c4c57b2be2] Successfully updated port: 65fcb910-75be-4a68-8cec-476f7cc7cd00 {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 806.214955] env[62066]: DEBUG oslo_vmware.api [None req-df48af02-4c4b-4364-8d38-efc343bf8241 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': task-1340769, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.366851] env[62066]: DEBUG nova.compute.manager [None req-d2edd17d-ac3d-4b81-90f1-278c62ed416c tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] [instance: de2e2be7-efdb-45a8-842a-640ab9deb1d9] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 806.411961] env[62066]: DEBUG nova.network.neutron [None req-8034bd7f-2294-4f4c-b924-f35bec7ebaac tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] [instance: 738ebb2f-7cc6-4d0f-871c-1428c9c21384] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 806.454160] env[62066]: DEBUG nova.compute.manager [None req-90f5cbab-7c0d-4b69-92aa-09bf24d9f7da tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] [instance: 050040cb-6ee3-4ad2-960a-fcebb53ac394] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 806.594678] env[62066]: DEBUG nova.network.neutron [None req-90f5cbab-7c0d-4b69-92aa-09bf24d9f7da tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] [instance: 050040cb-6ee3-4ad2-960a-fcebb53ac394] Successfully created port: 72521b90-587a-4f59-b744-919ef3087539 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 806.601817] env[62066]: DEBUG oslo_vmware.api [None req-febd07a3-de81-4924-888e-f9c7d8ba2329 tempest-ServerMetadataNegativeTestJSON-1549470887 tempest-ServerMetadataNegativeTestJSON-1549470887-project-member] Task: {'id': task-1340768, 'name': PowerOnVM_Task} progress is 64%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.652489] env[62066]: DEBUG nova.network.neutron [None req-8034bd7f-2294-4f4c-b924-f35bec7ebaac tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] [instance: 738ebb2f-7cc6-4d0f-871c-1428c9c21384] Updating instance_info_cache with network_info: [{"id": "fa65affa-3b23-4d64-a714-83d41caa3293", "address": "fa:16:3e:96:ce:2d", "network": {"id": "c9965d43-8841-4019-9ef2-6fc9276004c2", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-1873908750-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "636da8edfb3a4ad58312b97b1cb23a97", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4307c18-b235-43cd-bcd5-e226012d8ee9", "external-id": "nsx-vlan-transportzone-867", "segmentation_id": 867, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfa65affa-3b", "ovs_interfaceid": "fa65affa-3b23-4d64-a714-83d41caa3293", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 806.716345] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f6cc809d-278e-4fb4-906b-f468fa41739d tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Acquiring lock "refresh_cache-5f645eaa-6edc-4362-82ad-38c4c57b2be2" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 806.716345] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f6cc809d-278e-4fb4-906b-f468fa41739d tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Acquired lock "refresh_cache-5f645eaa-6edc-4362-82ad-38c4c57b2be2" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 806.716345] env[62066]: DEBUG nova.network.neutron [None req-f6cc809d-278e-4fb4-906b-f468fa41739d tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] [instance: 5f645eaa-6edc-4362-82ad-38c4c57b2be2] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 806.716721] env[62066]: DEBUG oslo_vmware.api [None req-df48af02-4c4b-4364-8d38-efc343bf8241 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': task-1340769, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064062} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.720444] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-df48af02-4c4b-4364-8d38-efc343bf8241 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 49f46244-34fa-48a1-95a2-8e95850f345d] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 806.722437] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1b10d3a-8dd4-4d9f-909e-fd83cbf2db35 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.751094] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-df48af02-4c4b-4364-8d38-efc343bf8241 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 49f46244-34fa-48a1-95a2-8e95850f345d] Reconfiguring VM instance instance-00000040 to attach disk [datastore2] 49f46244-34fa-48a1-95a2-8e95850f345d/49f46244-34fa-48a1-95a2-8e95850f345d.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 806.755905] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9b3fa90b-dc05-4536-8d6f-64a35d2c078d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.772305] env[62066]: DEBUG nova.compute.manager [req-44a82cb1-be80-400c-948b-4f6e4ad5ff62 req-9fc98555-3ab1-42a0-92d0-ed33959d16c1 service nova] [instance: 5f645eaa-6edc-4362-82ad-38c4c57b2be2] Received event network-vif-plugged-65fcb910-75be-4a68-8cec-476f7cc7cd00 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 806.772939] env[62066]: DEBUG oslo_concurrency.lockutils [req-44a82cb1-be80-400c-948b-4f6e4ad5ff62 req-9fc98555-3ab1-42a0-92d0-ed33959d16c1 service nova] Acquiring lock "5f645eaa-6edc-4362-82ad-38c4c57b2be2-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 806.772939] env[62066]: DEBUG oslo_concurrency.lockutils [req-44a82cb1-be80-400c-948b-4f6e4ad5ff62 req-9fc98555-3ab1-42a0-92d0-ed33959d16c1 service nova] Lock "5f645eaa-6edc-4362-82ad-38c4c57b2be2-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 806.772939] env[62066]: DEBUG oslo_concurrency.lockutils [req-44a82cb1-be80-400c-948b-4f6e4ad5ff62 req-9fc98555-3ab1-42a0-92d0-ed33959d16c1 service nova] Lock "5f645eaa-6edc-4362-82ad-38c4c57b2be2-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 806.773362] env[62066]: DEBUG nova.compute.manager [req-44a82cb1-be80-400c-948b-4f6e4ad5ff62 req-9fc98555-3ab1-42a0-92d0-ed33959d16c1 service nova] [instance: 5f645eaa-6edc-4362-82ad-38c4c57b2be2] No waiting events found dispatching network-vif-plugged-65fcb910-75be-4a68-8cec-476f7cc7cd00 {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 806.773362] env[62066]: WARNING nova.compute.manager [req-44a82cb1-be80-400c-948b-4f6e4ad5ff62 req-9fc98555-3ab1-42a0-92d0-ed33959d16c1 service nova] [instance: 5f645eaa-6edc-4362-82ad-38c4c57b2be2] Received unexpected event network-vif-plugged-65fcb910-75be-4a68-8cec-476f7cc7cd00 for instance with vm_state building and task_state spawning. [ 806.773530] env[62066]: DEBUG nova.compute.manager [req-44a82cb1-be80-400c-948b-4f6e4ad5ff62 req-9fc98555-3ab1-42a0-92d0-ed33959d16c1 service nova] [instance: 5f645eaa-6edc-4362-82ad-38c4c57b2be2] Received event network-changed-65fcb910-75be-4a68-8cec-476f7cc7cd00 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 806.773644] env[62066]: DEBUG nova.compute.manager [req-44a82cb1-be80-400c-948b-4f6e4ad5ff62 req-9fc98555-3ab1-42a0-92d0-ed33959d16c1 service nova] [instance: 5f645eaa-6edc-4362-82ad-38c4c57b2be2] Refreshing instance network info cache due to event network-changed-65fcb910-75be-4a68-8cec-476f7cc7cd00. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 806.773845] env[62066]: DEBUG oslo_concurrency.lockutils [req-44a82cb1-be80-400c-948b-4f6e4ad5ff62 req-9fc98555-3ab1-42a0-92d0-ed33959d16c1 service nova] Acquiring lock "refresh_cache-5f645eaa-6edc-4362-82ad-38c4c57b2be2" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 806.780945] env[62066]: DEBUG oslo_vmware.api [None req-df48af02-4c4b-4364-8d38-efc343bf8241 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Waiting for the task: (returnval){ [ 806.780945] env[62066]: value = "task-1340770" [ 806.780945] env[62066]: _type = "Task" [ 806.780945] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.792757] env[62066]: DEBUG oslo_vmware.api [None req-df48af02-4c4b-4364-8d38-efc343bf8241 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': task-1340770, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.801896] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-527d97e0-9a46-4747-a430-3569f222f443 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.810937] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b58565b1-36d8-4480-89b5-085d1325df88 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.842896] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba68d68a-ad7b-4e65-aecb-70a003ed7b4d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.851451] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c8831d0-0398-403d-9f91-503f39de1f7c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.872226] env[62066]: DEBUG nova.compute.provider_tree [None req-0188668d-25e1-420a-be60-fb5771349282 tempest-ServerAddressesTestJSON-1053293248 tempest-ServerAddressesTestJSON-1053293248-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 806.904584] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d2edd17d-ac3d-4b81-90f1-278c62ed416c tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 807.100172] env[62066]: DEBUG oslo_vmware.api [None req-febd07a3-de81-4924-888e-f9c7d8ba2329 tempest-ServerMetadataNegativeTestJSON-1549470887 tempest-ServerMetadataNegativeTestJSON-1549470887-project-member] Task: {'id': task-1340768, 'name': PowerOnVM_Task} progress is 91%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.124123] env[62066]: DEBUG nova.network.neutron [None req-90f5cbab-7c0d-4b69-92aa-09bf24d9f7da tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] [instance: 050040cb-6ee3-4ad2-960a-fcebb53ac394] Successfully created port: d329136d-bd14-4bee-b09a-39ef97845251 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 807.158681] env[62066]: DEBUG oslo_concurrency.lockutils [None req-8034bd7f-2294-4f4c-b924-f35bec7ebaac tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] Releasing lock "refresh_cache-738ebb2f-7cc6-4d0f-871c-1428c9c21384" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 807.159035] env[62066]: DEBUG nova.compute.manager [None req-8034bd7f-2294-4f4c-b924-f35bec7ebaac tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] [instance: 738ebb2f-7cc6-4d0f-871c-1428c9c21384] Instance network_info: |[{"id": "fa65affa-3b23-4d64-a714-83d41caa3293", "address": "fa:16:3e:96:ce:2d", "network": {"id": "c9965d43-8841-4019-9ef2-6fc9276004c2", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-1873908750-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "636da8edfb3a4ad58312b97b1cb23a97", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4307c18-b235-43cd-bcd5-e226012d8ee9", "external-id": "nsx-vlan-transportzone-867", "segmentation_id": 867, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfa65affa-3b", "ovs_interfaceid": "fa65affa-3b23-4d64-a714-83d41caa3293", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 807.159565] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-8034bd7f-2294-4f4c-b924-f35bec7ebaac tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] [instance: 738ebb2f-7cc6-4d0f-871c-1428c9c21384] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:96:ce:2d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a4307c18-b235-43cd-bcd5-e226012d8ee9', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'fa65affa-3b23-4d64-a714-83d41caa3293', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 807.171494] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-8034bd7f-2294-4f4c-b924-f35bec7ebaac tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] Creating folder: Project (636da8edfb3a4ad58312b97b1cb23a97). Parent ref: group-v285980. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 807.171849] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-74932b63-d494-47f1-8c9e-6f8187be3e7a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.182378] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-8034bd7f-2294-4f4c-b924-f35bec7ebaac tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] Created folder: Project (636da8edfb3a4ad58312b97b1cb23a97) in parent group-v285980. [ 807.182705] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-8034bd7f-2294-4f4c-b924-f35bec7ebaac tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] Creating folder: Instances. Parent ref: group-v286028. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 807.182852] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b8968b2d-1329-4925-a694-16f0383ea846 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.193021] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-8034bd7f-2294-4f4c-b924-f35bec7ebaac tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] Created folder: Instances in parent group-v286028. [ 807.193021] env[62066]: DEBUG oslo.service.loopingcall [None req-8034bd7f-2294-4f4c-b924-f35bec7ebaac tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 807.193021] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 738ebb2f-7cc6-4d0f-871c-1428c9c21384] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 807.193021] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6d06df95-3015-4471-92df-0978f737194d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.212888] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 807.212888] env[62066]: value = "task-1340773" [ 807.212888] env[62066]: _type = "Task" [ 807.212888] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.223046] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1340773, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.258823] env[62066]: DEBUG nova.network.neutron [None req-f6cc809d-278e-4fb4-906b-f468fa41739d tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] [instance: 5f645eaa-6edc-4362-82ad-38c4c57b2be2] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 807.296532] env[62066]: DEBUG oslo_vmware.api [None req-df48af02-4c4b-4364-8d38-efc343bf8241 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': task-1340770, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.379927] env[62066]: DEBUG nova.scheduler.client.report [None req-0188668d-25e1-420a-be60-fb5771349282 tempest-ServerAddressesTestJSON-1053293248 tempest-ServerAddressesTestJSON-1053293248-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 807.468150] env[62066]: DEBUG nova.compute.manager [None req-90f5cbab-7c0d-4b69-92aa-09bf24d9f7da tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] [instance: 050040cb-6ee3-4ad2-960a-fcebb53ac394] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 807.498024] env[62066]: DEBUG nova.virt.hardware [None req-90f5cbab-7c0d-4b69-92aa-09bf24d9f7da tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-23T13:40:41Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-23T13:40:23Z,direct_url=,disk_format='vmdk',id=50ff584c-3b50-4395-af07-3e66769bc9f7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='eb52f7069a374c61ae946f052007c6d9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-23T13:40:24Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 807.498024] env[62066]: DEBUG nova.virt.hardware [None req-90f5cbab-7c0d-4b69-92aa-09bf24d9f7da tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 807.498024] env[62066]: DEBUG nova.virt.hardware [None req-90f5cbab-7c0d-4b69-92aa-09bf24d9f7da tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 807.498024] env[62066]: DEBUG nova.virt.hardware [None req-90f5cbab-7c0d-4b69-92aa-09bf24d9f7da tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 807.498271] env[62066]: DEBUG nova.virt.hardware [None req-90f5cbab-7c0d-4b69-92aa-09bf24d9f7da tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 807.498271] env[62066]: DEBUG nova.virt.hardware [None req-90f5cbab-7c0d-4b69-92aa-09bf24d9f7da tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 807.498387] env[62066]: DEBUG nova.virt.hardware [None req-90f5cbab-7c0d-4b69-92aa-09bf24d9f7da tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 807.498667] env[62066]: DEBUG nova.virt.hardware [None req-90f5cbab-7c0d-4b69-92aa-09bf24d9f7da tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 807.498988] env[62066]: DEBUG nova.virt.hardware [None req-90f5cbab-7c0d-4b69-92aa-09bf24d9f7da tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 807.499439] env[62066]: DEBUG nova.virt.hardware [None req-90f5cbab-7c0d-4b69-92aa-09bf24d9f7da tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 807.499881] env[62066]: DEBUG nova.virt.hardware [None req-90f5cbab-7c0d-4b69-92aa-09bf24d9f7da tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 807.501082] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07710b8e-21c8-4ae7-b845-1b78c7b39025 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.507023] env[62066]: DEBUG nova.network.neutron [None req-f6cc809d-278e-4fb4-906b-f468fa41739d tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] [instance: 5f645eaa-6edc-4362-82ad-38c4c57b2be2] Updating instance_info_cache with network_info: [{"id": "65fcb910-75be-4a68-8cec-476f7cc7cd00", "address": "fa:16:3e:20:29:72", "network": {"id": "b0f4607a-9d59-457e-aabc-5851c90b3366", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-1154549021-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a9b239f4cb3242d0bd1dcd98a2a3bfc0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c02dd284-ab80-451c-93eb-48c8360acb9c", "external-id": "nsx-vlan-transportzone-818", "segmentation_id": 818, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap65fcb910-75", "ovs_interfaceid": "65fcb910-75be-4a68-8cec-476f7cc7cd00", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 807.513969] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d97eda32-d189-4845-90c7-f665e9591fd0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.601556] env[62066]: DEBUG oslo_vmware.api [None req-febd07a3-de81-4924-888e-f9c7d8ba2329 tempest-ServerMetadataNegativeTestJSON-1549470887 tempest-ServerMetadataNegativeTestJSON-1549470887-project-member] Task: {'id': task-1340768, 'name': PowerOnVM_Task, 'duration_secs': 1.582495} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.601923] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-febd07a3-de81-4924-888e-f9c7d8ba2329 tempest-ServerMetadataNegativeTestJSON-1549470887 tempest-ServerMetadataNegativeTestJSON-1549470887-project-member] [instance: 31d47299-83a3-4f27-aeb7-95c4cd36c5ac] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 807.602188] env[62066]: INFO nova.compute.manager [None req-febd07a3-de81-4924-888e-f9c7d8ba2329 tempest-ServerMetadataNegativeTestJSON-1549470887 tempest-ServerMetadataNegativeTestJSON-1549470887-project-member] [instance: 31d47299-83a3-4f27-aeb7-95c4cd36c5ac] Took 9.67 seconds to spawn the instance on the hypervisor. [ 807.602422] env[62066]: DEBUG nova.compute.manager [None req-febd07a3-de81-4924-888e-f9c7d8ba2329 tempest-ServerMetadataNegativeTestJSON-1549470887 tempest-ServerMetadataNegativeTestJSON-1549470887-project-member] [instance: 31d47299-83a3-4f27-aeb7-95c4cd36c5ac] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 807.603300] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81152009-8c7c-4df9-958f-e9e759f7e407 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.723613] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1340773, 'name': CreateVM_Task, 'duration_secs': 0.4129} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.723899] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 738ebb2f-7cc6-4d0f-871c-1428c9c21384] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 807.725486] env[62066]: DEBUG oslo_vmware.service [None req-8034bd7f-2294-4f4c-b924-f35bec7ebaac tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db576a48-2052-4a72-ba09-0500657c09e9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.731236] env[62066]: DEBUG oslo_concurrency.lockutils [None req-8034bd7f-2294-4f4c-b924-f35bec7ebaac tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 807.731537] env[62066]: DEBUG oslo_concurrency.lockutils [None req-8034bd7f-2294-4f4c-b924-f35bec7ebaac tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] Acquired lock "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 807.731966] env[62066]: DEBUG oslo_concurrency.lockutils [None req-8034bd7f-2294-4f4c-b924-f35bec7ebaac tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 807.732305] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-eb9855c0-d549-4369-be37-c1be39ee06f6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.737115] env[62066]: DEBUG oslo_vmware.api [None req-8034bd7f-2294-4f4c-b924-f35bec7ebaac tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] Waiting for the task: (returnval){ [ 807.737115] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]52725f02-9693-f004-3b6f-da6c848b7608" [ 807.737115] env[62066]: _type = "Task" [ 807.737115] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.744746] env[62066]: DEBUG oslo_vmware.api [None req-8034bd7f-2294-4f4c-b924-f35bec7ebaac tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52725f02-9693-f004-3b6f-da6c848b7608, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.797361] env[62066]: DEBUG oslo_vmware.api [None req-df48af02-4c4b-4364-8d38-efc343bf8241 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': task-1340770, 'name': ReconfigVM_Task, 'duration_secs': 0.620008} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.797890] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-df48af02-4c4b-4364-8d38-efc343bf8241 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 49f46244-34fa-48a1-95a2-8e95850f345d] Reconfigured VM instance instance-00000040 to attach disk [datastore2] 49f46244-34fa-48a1-95a2-8e95850f345d/49f46244-34fa-48a1-95a2-8e95850f345d.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 807.798851] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b8711d73-f2be-49f1-a7fd-a7cd832326e3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.808218] env[62066]: DEBUG oslo_vmware.api [None req-df48af02-4c4b-4364-8d38-efc343bf8241 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Waiting for the task: (returnval){ [ 807.808218] env[62066]: value = "task-1340774" [ 807.808218] env[62066]: _type = "Task" [ 807.808218] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.819460] env[62066]: DEBUG oslo_vmware.api [None req-df48af02-4c4b-4364-8d38-efc343bf8241 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': task-1340774, 'name': Rename_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.888721] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0188668d-25e1-420a-be60-fb5771349282 tempest-ServerAddressesTestJSON-1053293248 tempest-ServerAddressesTestJSON-1053293248-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.446s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 807.888721] env[62066]: DEBUG nova.compute.manager [None req-0188668d-25e1-420a-be60-fb5771349282 tempest-ServerAddressesTestJSON-1053293248 tempest-ServerAddressesTestJSON-1053293248-project-member] [instance: 49bc1350-0095-406b-bc68-005eb6b681a8] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 807.892845] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ba3a3138-4a4e-4a10-bbee-93dd3a372fa2 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 14.149s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 807.892952] env[62066]: DEBUG nova.objects.instance [None req-ba3a3138-4a4e-4a10-bbee-93dd3a372fa2 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Lazy-loading 'resources' on Instance uuid 0ebba3b2-f82d-4c1d-b01a-f75c3559c117 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 808.012161] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f6cc809d-278e-4fb4-906b-f468fa41739d tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Releasing lock "refresh_cache-5f645eaa-6edc-4362-82ad-38c4c57b2be2" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 808.012161] env[62066]: DEBUG nova.compute.manager [None req-f6cc809d-278e-4fb4-906b-f468fa41739d tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] [instance: 5f645eaa-6edc-4362-82ad-38c4c57b2be2] Instance network_info: |[{"id": "65fcb910-75be-4a68-8cec-476f7cc7cd00", "address": "fa:16:3e:20:29:72", "network": {"id": "b0f4607a-9d59-457e-aabc-5851c90b3366", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-1154549021-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a9b239f4cb3242d0bd1dcd98a2a3bfc0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c02dd284-ab80-451c-93eb-48c8360acb9c", "external-id": "nsx-vlan-transportzone-818", "segmentation_id": 818, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap65fcb910-75", "ovs_interfaceid": "65fcb910-75be-4a68-8cec-476f7cc7cd00", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 808.012358] env[62066]: DEBUG oslo_concurrency.lockutils [req-44a82cb1-be80-400c-948b-4f6e4ad5ff62 req-9fc98555-3ab1-42a0-92d0-ed33959d16c1 service nova] Acquired lock "refresh_cache-5f645eaa-6edc-4362-82ad-38c4c57b2be2" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 808.012589] env[62066]: DEBUG nova.network.neutron [req-44a82cb1-be80-400c-948b-4f6e4ad5ff62 req-9fc98555-3ab1-42a0-92d0-ed33959d16c1 service nova] [instance: 5f645eaa-6edc-4362-82ad-38c4c57b2be2] Refreshing network info cache for port 65fcb910-75be-4a68-8cec-476f7cc7cd00 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 808.014304] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-f6cc809d-278e-4fb4-906b-f468fa41739d tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] [instance: 5f645eaa-6edc-4362-82ad-38c4c57b2be2] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:20:29:72', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c02dd284-ab80-451c-93eb-48c8360acb9c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '65fcb910-75be-4a68-8cec-476f7cc7cd00', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 808.027906] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-f6cc809d-278e-4fb4-906b-f468fa41739d tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Creating folder: Project (a9b239f4cb3242d0bd1dcd98a2a3bfc0). Parent ref: group-v285980. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 808.029074] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4886b89a-2f76-4341-8017-823d1dec8fff {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.041108] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-f6cc809d-278e-4fb4-906b-f468fa41739d tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Created folder: Project (a9b239f4cb3242d0bd1dcd98a2a3bfc0) in parent group-v285980. [ 808.041273] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-f6cc809d-278e-4fb4-906b-f468fa41739d tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Creating folder: Instances. Parent ref: group-v286031. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 808.041513] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-36d5616e-c64f-425d-b58b-c2305ec72fe9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.049596] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-f6cc809d-278e-4fb4-906b-f468fa41739d tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Created folder: Instances in parent group-v286031. [ 808.050015] env[62066]: DEBUG oslo.service.loopingcall [None req-f6cc809d-278e-4fb4-906b-f468fa41739d tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 808.050941] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5f645eaa-6edc-4362-82ad-38c4c57b2be2] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 808.050941] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d3e586fa-1a41-4da5-9518-222092838f99 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.070452] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 808.070452] env[62066]: value = "task-1340777" [ 808.070452] env[62066]: _type = "Task" [ 808.070452] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.078790] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1340777, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.122180] env[62066]: INFO nova.compute.manager [None req-febd07a3-de81-4924-888e-f9c7d8ba2329 tempest-ServerMetadataNegativeTestJSON-1549470887 tempest-ServerMetadataNegativeTestJSON-1549470887-project-member] [instance: 31d47299-83a3-4f27-aeb7-95c4cd36c5ac] Took 27.91 seconds to build instance. [ 808.248601] env[62066]: DEBUG oslo_concurrency.lockutils [None req-8034bd7f-2294-4f4c-b924-f35bec7ebaac tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] Releasing lock "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 808.248896] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-8034bd7f-2294-4f4c-b924-f35bec7ebaac tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] [instance: 738ebb2f-7cc6-4d0f-871c-1428c9c21384] Processing image 50ff584c-3b50-4395-af07-3e66769bc9f7 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 808.249133] env[62066]: DEBUG oslo_concurrency.lockutils [None req-8034bd7f-2294-4f4c-b924-f35bec7ebaac tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 808.249371] env[62066]: DEBUG oslo_concurrency.lockutils [None req-8034bd7f-2294-4f4c-b924-f35bec7ebaac tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] Acquired lock "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 808.249564] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-8034bd7f-2294-4f4c-b924-f35bec7ebaac tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 808.249821] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-feeac24a-cb34-4055-9060-ae507fda9812 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.258816] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-8034bd7f-2294-4f4c-b924-f35bec7ebaac tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 808.259074] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-8034bd7f-2294-4f4c-b924-f35bec7ebaac tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 808.259889] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9755dfde-9527-452b-87db-699536d75ed6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.266224] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cf48aabb-0f80-41e0-9082-55e519cd5840 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.271643] env[62066]: DEBUG oslo_vmware.api [None req-8034bd7f-2294-4f4c-b924-f35bec7ebaac tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] Waiting for the task: (returnval){ [ 808.271643] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]52601e50-8b7e-b12c-69e7-35ad75bba7a3" [ 808.271643] env[62066]: _type = "Task" [ 808.271643] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.279461] env[62066]: DEBUG oslo_vmware.api [None req-8034bd7f-2294-4f4c-b924-f35bec7ebaac tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52601e50-8b7e-b12c-69e7-35ad75bba7a3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.318542] env[62066]: DEBUG oslo_vmware.api [None req-df48af02-4c4b-4364-8d38-efc343bf8241 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': task-1340774, 'name': Rename_Task, 'duration_secs': 0.315449} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.318869] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-df48af02-4c4b-4364-8d38-efc343bf8241 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 49f46244-34fa-48a1-95a2-8e95850f345d] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 808.319145] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-fee05035-5a26-415b-92d8-6436a0408034 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.325466] env[62066]: DEBUG oslo_vmware.api [None req-df48af02-4c4b-4364-8d38-efc343bf8241 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Waiting for the task: (returnval){ [ 808.325466] env[62066]: value = "task-1340778" [ 808.325466] env[62066]: _type = "Task" [ 808.325466] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.333539] env[62066]: DEBUG oslo_vmware.api [None req-df48af02-4c4b-4364-8d38-efc343bf8241 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': task-1340778, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.397257] env[62066]: DEBUG nova.compute.utils [None req-0188668d-25e1-420a-be60-fb5771349282 tempest-ServerAddressesTestJSON-1053293248 tempest-ServerAddressesTestJSON-1053293248-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 808.401106] env[62066]: DEBUG nova.compute.manager [None req-0188668d-25e1-420a-be60-fb5771349282 tempest-ServerAddressesTestJSON-1053293248 tempest-ServerAddressesTestJSON-1053293248-project-member] [instance: 49bc1350-0095-406b-bc68-005eb6b681a8] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 808.401313] env[62066]: DEBUG nova.network.neutron [None req-0188668d-25e1-420a-be60-fb5771349282 tempest-ServerAddressesTestJSON-1053293248 tempest-ServerAddressesTestJSON-1053293248-project-member] [instance: 49bc1350-0095-406b-bc68-005eb6b681a8] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 808.454422] env[62066]: DEBUG nova.policy [None req-0188668d-25e1-420a-be60-fb5771349282 tempest-ServerAddressesTestJSON-1053293248 tempest-ServerAddressesTestJSON-1053293248-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8d6a7e477826481c980e70bca5cd9c42', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '39a464e941734bfaad54e66d920290fa', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 808.581158] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1340777, 'name': CreateVM_Task} progress is 25%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.624501] env[62066]: DEBUG oslo_concurrency.lockutils [None req-febd07a3-de81-4924-888e-f9c7d8ba2329 tempest-ServerMetadataNegativeTestJSON-1549470887 tempest-ServerMetadataNegativeTestJSON-1549470887-project-member] Lock "31d47299-83a3-4f27-aeb7-95c4cd36c5ac" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 125.747s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 808.673348] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-236ba4d2-c5e4-4000-ba11-4afe21a9dc78 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.682806] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5b99596-eeb4-4838-ad9d-0020febcb27e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.715975] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc12d78f-ef30-4559-b3de-f8e1f830015a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.723276] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a7bd349-af9c-43fe-9c32-5755cd23b321 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.737194] env[62066]: DEBUG nova.compute.provider_tree [None req-ba3a3138-4a4e-4a10-bbee-93dd3a372fa2 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 808.783511] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-8034bd7f-2294-4f4c-b924-f35bec7ebaac tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] [instance: 738ebb2f-7cc6-4d0f-871c-1428c9c21384] Preparing fetch location {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 808.783511] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-8034bd7f-2294-4f4c-b924-f35bec7ebaac tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] Creating directory with path [datastore1] vmware_temp/0ed0c1aa-638f-496c-bf5c-978d07f863bd/50ff584c-3b50-4395-af07-3e66769bc9f7 {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 808.783511] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-43662350-6534-47ef-8b64-ea4d1d73a588 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.831451] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-8034bd7f-2294-4f4c-b924-f35bec7ebaac tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] Created directory with path [datastore1] vmware_temp/0ed0c1aa-638f-496c-bf5c-978d07f863bd/50ff584c-3b50-4395-af07-3e66769bc9f7 {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 808.831772] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-8034bd7f-2294-4f4c-b924-f35bec7ebaac tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] [instance: 738ebb2f-7cc6-4d0f-871c-1428c9c21384] Fetch image to [datastore1] vmware_temp/0ed0c1aa-638f-496c-bf5c-978d07f863bd/50ff584c-3b50-4395-af07-3e66769bc9f7/tmp-sparse.vmdk {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 808.831957] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-8034bd7f-2294-4f4c-b924-f35bec7ebaac tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] [instance: 738ebb2f-7cc6-4d0f-871c-1428c9c21384] Downloading image file data 50ff584c-3b50-4395-af07-3e66769bc9f7 to [datastore1] vmware_temp/0ed0c1aa-638f-496c-bf5c-978d07f863bd/50ff584c-3b50-4395-af07-3e66769bc9f7/tmp-sparse.vmdk on the data store datastore1 {{(pid=62066) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 808.835890] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b933385c-f82c-4490-8912-f4253ac2dc58 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.839151] env[62066]: DEBUG oslo_vmware.api [None req-df48af02-4c4b-4364-8d38-efc343bf8241 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': task-1340778, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.844207] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6d6d7ee-3935-4cad-bc55-ecee15981009 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.857765] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de5d36af-2130-4f4a-9df5-19801381abd6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.893145] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e5f755b-63c7-4b0e-b006-7267726625df {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.898614] env[62066]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-5d70911e-3d25-4fcd-905d-989350e18244 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.904703] env[62066]: DEBUG nova.compute.manager [None req-0188668d-25e1-420a-be60-fb5771349282 tempest-ServerAddressesTestJSON-1053293248 tempest-ServerAddressesTestJSON-1053293248-project-member] [instance: 49bc1350-0095-406b-bc68-005eb6b681a8] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 808.997965] env[62066]: DEBUG nova.virt.vmwareapi.images [None req-8034bd7f-2294-4f4c-b924-f35bec7ebaac tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] [instance: 738ebb2f-7cc6-4d0f-871c-1428c9c21384] Downloading image file data 50ff584c-3b50-4395-af07-3e66769bc9f7 to the data store datastore1 {{(pid=62066) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 809.018371] env[62066]: DEBUG nova.network.neutron [None req-0188668d-25e1-420a-be60-fb5771349282 tempest-ServerAddressesTestJSON-1053293248 tempest-ServerAddressesTestJSON-1053293248-project-member] [instance: 49bc1350-0095-406b-bc68-005eb6b681a8] Successfully created port: f2c406e7-6447-47bf-9ee1-dba97dc7e5bc {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 809.041906] env[62066]: DEBUG nova.compute.manager [req-9a26e71e-3d22-4894-818b-2f391fb3f853 req-87474802-b203-4371-b39a-bab06781da8e service nova] [instance: 050040cb-6ee3-4ad2-960a-fcebb53ac394] Received event network-vif-plugged-72521b90-587a-4f59-b744-919ef3087539 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 809.042154] env[62066]: DEBUG oslo_concurrency.lockutils [req-9a26e71e-3d22-4894-818b-2f391fb3f853 req-87474802-b203-4371-b39a-bab06781da8e service nova] Acquiring lock "050040cb-6ee3-4ad2-960a-fcebb53ac394-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 809.042376] env[62066]: DEBUG oslo_concurrency.lockutils [req-9a26e71e-3d22-4894-818b-2f391fb3f853 req-87474802-b203-4371-b39a-bab06781da8e service nova] Lock "050040cb-6ee3-4ad2-960a-fcebb53ac394-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 809.042554] env[62066]: DEBUG oslo_concurrency.lockutils [req-9a26e71e-3d22-4894-818b-2f391fb3f853 req-87474802-b203-4371-b39a-bab06781da8e service nova] Lock "050040cb-6ee3-4ad2-960a-fcebb53ac394-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 809.042723] env[62066]: DEBUG nova.compute.manager [req-9a26e71e-3d22-4894-818b-2f391fb3f853 req-87474802-b203-4371-b39a-bab06781da8e service nova] [instance: 050040cb-6ee3-4ad2-960a-fcebb53ac394] No waiting events found dispatching network-vif-plugged-72521b90-587a-4f59-b744-919ef3087539 {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 809.042890] env[62066]: WARNING nova.compute.manager [req-9a26e71e-3d22-4894-818b-2f391fb3f853 req-87474802-b203-4371-b39a-bab06781da8e service nova] [instance: 050040cb-6ee3-4ad2-960a-fcebb53ac394] Received unexpected event network-vif-plugged-72521b90-587a-4f59-b744-919ef3087539 for instance with vm_state building and task_state spawning. [ 809.051853] env[62066]: DEBUG nova.network.neutron [req-44a82cb1-be80-400c-948b-4f6e4ad5ff62 req-9fc98555-3ab1-42a0-92d0-ed33959d16c1 service nova] [instance: 5f645eaa-6edc-4362-82ad-38c4c57b2be2] Updated VIF entry in instance network info cache for port 65fcb910-75be-4a68-8cec-476f7cc7cd00. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 809.052232] env[62066]: DEBUG nova.network.neutron [req-44a82cb1-be80-400c-948b-4f6e4ad5ff62 req-9fc98555-3ab1-42a0-92d0-ed33959d16c1 service nova] [instance: 5f645eaa-6edc-4362-82ad-38c4c57b2be2] Updating instance_info_cache with network_info: [{"id": "65fcb910-75be-4a68-8cec-476f7cc7cd00", "address": "fa:16:3e:20:29:72", "network": {"id": "b0f4607a-9d59-457e-aabc-5851c90b3366", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-1154549021-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a9b239f4cb3242d0bd1dcd98a2a3bfc0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c02dd284-ab80-451c-93eb-48c8360acb9c", "external-id": "nsx-vlan-transportzone-818", "segmentation_id": 818, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap65fcb910-75", "ovs_interfaceid": "65fcb910-75be-4a68-8cec-476f7cc7cd00", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 809.064755] env[62066]: DEBUG oslo_vmware.rw_handles [None req-8034bd7f-2294-4f4c-b924-f35bec7ebaac tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c1n1.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/0ed0c1aa-638f-496c-bf5c-978d07f863bd/50ff584c-3b50-4395-af07-3e66769bc9f7/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=62066) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 809.129395] env[62066]: DEBUG nova.compute.manager [None req-acef35f4-b906-4efb-ae97-e8cdc80e8e5a tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] [instance: 5e96def2-0cbd-4bd9-93f4-6a365a0142b0] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 809.147374] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1340777, 'name': CreateVM_Task, 'duration_secs': 0.765692} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.147647] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5f645eaa-6edc-4362-82ad-38c4c57b2be2] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 809.148670] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f6cc809d-278e-4fb4-906b-f468fa41739d tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 809.148939] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f6cc809d-278e-4fb4-906b-f468fa41739d tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Acquired lock "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 809.149463] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f6cc809d-278e-4fb4-906b-f468fa41739d tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 809.149852] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-03304586-d264-4ad0-b247-fb002aa284a5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.157344] env[62066]: DEBUG oslo_vmware.api [None req-f6cc809d-278e-4fb4-906b-f468fa41739d tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Waiting for the task: (returnval){ [ 809.157344] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]52370633-6437-c5cd-13c6-af70882196f0" [ 809.157344] env[62066]: _type = "Task" [ 809.157344] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.167733] env[62066]: DEBUG oslo_vmware.api [None req-f6cc809d-278e-4fb4-906b-f468fa41739d tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52370633-6437-c5cd-13c6-af70882196f0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.212161] env[62066]: DEBUG nova.network.neutron [None req-90f5cbab-7c0d-4b69-92aa-09bf24d9f7da tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] [instance: 050040cb-6ee3-4ad2-960a-fcebb53ac394] Successfully updated port: 72521b90-587a-4f59-b744-919ef3087539 {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 809.240860] env[62066]: DEBUG nova.scheduler.client.report [None req-ba3a3138-4a4e-4a10-bbee-93dd3a372fa2 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 809.290792] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d041ede3-45cc-4fa6-b4cc-7d603e4097a9 tempest-ServerMetadataNegativeTestJSON-1549470887 tempest-ServerMetadataNegativeTestJSON-1549470887-project-member] Acquiring lock "31d47299-83a3-4f27-aeb7-95c4cd36c5ac" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 809.291150] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d041ede3-45cc-4fa6-b4cc-7d603e4097a9 tempest-ServerMetadataNegativeTestJSON-1549470887 tempest-ServerMetadataNegativeTestJSON-1549470887-project-member] Lock "31d47299-83a3-4f27-aeb7-95c4cd36c5ac" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 809.291451] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d041ede3-45cc-4fa6-b4cc-7d603e4097a9 tempest-ServerMetadataNegativeTestJSON-1549470887 tempest-ServerMetadataNegativeTestJSON-1549470887-project-member] Acquiring lock "31d47299-83a3-4f27-aeb7-95c4cd36c5ac-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 809.291693] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d041ede3-45cc-4fa6-b4cc-7d603e4097a9 tempest-ServerMetadataNegativeTestJSON-1549470887 tempest-ServerMetadataNegativeTestJSON-1549470887-project-member] Lock "31d47299-83a3-4f27-aeb7-95c4cd36c5ac-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 809.291871] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d041ede3-45cc-4fa6-b4cc-7d603e4097a9 tempest-ServerMetadataNegativeTestJSON-1549470887 tempest-ServerMetadataNegativeTestJSON-1549470887-project-member] Lock "31d47299-83a3-4f27-aeb7-95c4cd36c5ac-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 809.294424] env[62066]: INFO nova.compute.manager [None req-d041ede3-45cc-4fa6-b4cc-7d603e4097a9 tempest-ServerMetadataNegativeTestJSON-1549470887 tempest-ServerMetadataNegativeTestJSON-1549470887-project-member] [instance: 31d47299-83a3-4f27-aeb7-95c4cd36c5ac] Terminating instance [ 809.296809] env[62066]: DEBUG nova.compute.manager [None req-d041ede3-45cc-4fa6-b4cc-7d603e4097a9 tempest-ServerMetadataNegativeTestJSON-1549470887 tempest-ServerMetadataNegativeTestJSON-1549470887-project-member] [instance: 31d47299-83a3-4f27-aeb7-95c4cd36c5ac] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 809.297015] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-d041ede3-45cc-4fa6-b4cc-7d603e4097a9 tempest-ServerMetadataNegativeTestJSON-1549470887 tempest-ServerMetadataNegativeTestJSON-1549470887-project-member] [instance: 31d47299-83a3-4f27-aeb7-95c4cd36c5ac] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 809.297973] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ab1caec-139a-4e11-b846-3d0286cb9784 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.307876] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-d041ede3-45cc-4fa6-b4cc-7d603e4097a9 tempest-ServerMetadataNegativeTestJSON-1549470887 tempest-ServerMetadataNegativeTestJSON-1549470887-project-member] [instance: 31d47299-83a3-4f27-aeb7-95c4cd36c5ac] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 809.310660] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b1b71743-4cda-47a4-999a-ac04f48160fc {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.319413] env[62066]: DEBUG oslo_vmware.api [None req-d041ede3-45cc-4fa6-b4cc-7d603e4097a9 tempest-ServerMetadataNegativeTestJSON-1549470887 tempest-ServerMetadataNegativeTestJSON-1549470887-project-member] Waiting for the task: (returnval){ [ 809.319413] env[62066]: value = "task-1340779" [ 809.319413] env[62066]: _type = "Task" [ 809.319413] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.333953] env[62066]: DEBUG oslo_vmware.api [None req-d041ede3-45cc-4fa6-b4cc-7d603e4097a9 tempest-ServerMetadataNegativeTestJSON-1549470887 tempest-ServerMetadataNegativeTestJSON-1549470887-project-member] Task: {'id': task-1340779, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.342464] env[62066]: DEBUG oslo_vmware.api [None req-df48af02-4c4b-4364-8d38-efc343bf8241 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': task-1340778, 'name': PowerOnVM_Task, 'duration_secs': 0.810945} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.342625] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-df48af02-4c4b-4364-8d38-efc343bf8241 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 49f46244-34fa-48a1-95a2-8e95850f345d] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 809.342753] env[62066]: INFO nova.compute.manager [None req-df48af02-4c4b-4364-8d38-efc343bf8241 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 49f46244-34fa-48a1-95a2-8e95850f345d] Took 9.05 seconds to spawn the instance on the hypervisor. [ 809.342995] env[62066]: DEBUG nova.compute.manager [None req-df48af02-4c4b-4364-8d38-efc343bf8241 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 49f46244-34fa-48a1-95a2-8e95850f345d] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 809.343866] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9a88378-0b40-44c8-82e7-8e77a3e88add {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.555066] env[62066]: DEBUG oslo_concurrency.lockutils [req-44a82cb1-be80-400c-948b-4f6e4ad5ff62 req-9fc98555-3ab1-42a0-92d0-ed33959d16c1 service nova] Releasing lock "refresh_cache-5f645eaa-6edc-4362-82ad-38c4c57b2be2" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 809.555443] env[62066]: DEBUG nova.compute.manager [req-44a82cb1-be80-400c-948b-4f6e4ad5ff62 req-9fc98555-3ab1-42a0-92d0-ed33959d16c1 service nova] [instance: 459f432b-e817-405f-9c3e-dfa5d957788c] Received event network-changed-cb74e58f-b481-4a0f-861b-96381e2d52dc {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 809.556074] env[62066]: DEBUG nova.compute.manager [req-44a82cb1-be80-400c-948b-4f6e4ad5ff62 req-9fc98555-3ab1-42a0-92d0-ed33959d16c1 service nova] [instance: 459f432b-e817-405f-9c3e-dfa5d957788c] Refreshing instance network info cache due to event network-changed-cb74e58f-b481-4a0f-861b-96381e2d52dc. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 809.556074] env[62066]: DEBUG oslo_concurrency.lockutils [req-44a82cb1-be80-400c-948b-4f6e4ad5ff62 req-9fc98555-3ab1-42a0-92d0-ed33959d16c1 service nova] Acquiring lock "refresh_cache-459f432b-e817-405f-9c3e-dfa5d957788c" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 809.556236] env[62066]: DEBUG oslo_concurrency.lockutils [req-44a82cb1-be80-400c-948b-4f6e4ad5ff62 req-9fc98555-3ab1-42a0-92d0-ed33959d16c1 service nova] Acquired lock "refresh_cache-459f432b-e817-405f-9c3e-dfa5d957788c" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 809.556345] env[62066]: DEBUG nova.network.neutron [req-44a82cb1-be80-400c-948b-4f6e4ad5ff62 req-9fc98555-3ab1-42a0-92d0-ed33959d16c1 service nova] [instance: 459f432b-e817-405f-9c3e-dfa5d957788c] Refreshing network info cache for port cb74e58f-b481-4a0f-861b-96381e2d52dc {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 809.665209] env[62066]: DEBUG oslo_concurrency.lockutils [None req-acef35f4-b906-4efb-ae97-e8cdc80e8e5a tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 809.672526] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f6cc809d-278e-4fb4-906b-f468fa41739d tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Releasing lock "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 809.672797] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-f6cc809d-278e-4fb4-906b-f468fa41739d tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] [instance: 5f645eaa-6edc-4362-82ad-38c4c57b2be2] Processing image 50ff584c-3b50-4395-af07-3e66769bc9f7 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 809.672987] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f6cc809d-278e-4fb4-906b-f468fa41739d tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 809.749412] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ba3a3138-4a4e-4a10-bbee-93dd3a372fa2 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.856s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 809.751727] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e5dc4f2e-1d02-4ba9-acb0-6628ba7125f6 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 15.973s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 809.752108] env[62066]: DEBUG nova.objects.instance [None req-e5dc4f2e-1d02-4ba9-acb0-6628ba7125f6 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Lazy-loading 'resources' on Instance uuid 98fb270c-f3f8-4375-8b5c-c2279305d476 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 809.777600] env[62066]: INFO nova.scheduler.client.report [None req-ba3a3138-4a4e-4a10-bbee-93dd3a372fa2 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Deleted allocations for instance 0ebba3b2-f82d-4c1d-b01a-f75c3559c117 [ 809.830193] env[62066]: DEBUG oslo_vmware.api [None req-d041ede3-45cc-4fa6-b4cc-7d603e4097a9 tempest-ServerMetadataNegativeTestJSON-1549470887 tempest-ServerMetadataNegativeTestJSON-1549470887-project-member] Task: {'id': task-1340779, 'name': PowerOffVM_Task, 'duration_secs': 0.21952} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.830542] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-d041ede3-45cc-4fa6-b4cc-7d603e4097a9 tempest-ServerMetadataNegativeTestJSON-1549470887 tempest-ServerMetadataNegativeTestJSON-1549470887-project-member] [instance: 31d47299-83a3-4f27-aeb7-95c4cd36c5ac] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 809.830734] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-d041ede3-45cc-4fa6-b4cc-7d603e4097a9 tempest-ServerMetadataNegativeTestJSON-1549470887 tempest-ServerMetadataNegativeTestJSON-1549470887-project-member] [instance: 31d47299-83a3-4f27-aeb7-95c4cd36c5ac] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 809.832052] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-76f2b132-a197-4a96-a091-4e51c0207ef4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.862936] env[62066]: INFO nova.compute.manager [None req-df48af02-4c4b-4364-8d38-efc343bf8241 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 49f46244-34fa-48a1-95a2-8e95850f345d] Took 24.93 seconds to build instance. [ 809.897259] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-d041ede3-45cc-4fa6-b4cc-7d603e4097a9 tempest-ServerMetadataNegativeTestJSON-1549470887 tempest-ServerMetadataNegativeTestJSON-1549470887-project-member] [instance: 31d47299-83a3-4f27-aeb7-95c4cd36c5ac] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 809.897463] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-d041ede3-45cc-4fa6-b4cc-7d603e4097a9 tempest-ServerMetadataNegativeTestJSON-1549470887 tempest-ServerMetadataNegativeTestJSON-1549470887-project-member] [instance: 31d47299-83a3-4f27-aeb7-95c4cd36c5ac] Deleting contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 809.897661] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-d041ede3-45cc-4fa6-b4cc-7d603e4097a9 tempest-ServerMetadataNegativeTestJSON-1549470887 tempest-ServerMetadataNegativeTestJSON-1549470887-project-member] Deleting the datastore file [datastore2] 31d47299-83a3-4f27-aeb7-95c4cd36c5ac {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 809.898266] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-43e1d225-8f4b-4644-aef9-849ac97501ef {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.904776] env[62066]: DEBUG oslo_vmware.api [None req-d041ede3-45cc-4fa6-b4cc-7d603e4097a9 tempest-ServerMetadataNegativeTestJSON-1549470887 tempest-ServerMetadataNegativeTestJSON-1549470887-project-member] Waiting for the task: (returnval){ [ 809.904776] env[62066]: value = "task-1340781" [ 809.904776] env[62066]: _type = "Task" [ 809.904776] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.913805] env[62066]: DEBUG oslo_vmware.api [None req-d041ede3-45cc-4fa6-b4cc-7d603e4097a9 tempest-ServerMetadataNegativeTestJSON-1549470887 tempest-ServerMetadataNegativeTestJSON-1549470887-project-member] Task: {'id': task-1340781, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.918541] env[62066]: DEBUG nova.compute.manager [None req-0188668d-25e1-420a-be60-fb5771349282 tempest-ServerAddressesTestJSON-1053293248 tempest-ServerAddressesTestJSON-1053293248-project-member] [instance: 49bc1350-0095-406b-bc68-005eb6b681a8] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 809.950080] env[62066]: DEBUG nova.virt.hardware [None req-0188668d-25e1-420a-be60-fb5771349282 tempest-ServerAddressesTestJSON-1053293248 tempest-ServerAddressesTestJSON-1053293248-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-23T13:40:41Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-23T13:40:23Z,direct_url=,disk_format='vmdk',id=50ff584c-3b50-4395-af07-3e66769bc9f7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='eb52f7069a374c61ae946f052007c6d9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-23T13:40:24Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 809.950352] env[62066]: DEBUG nova.virt.hardware [None req-0188668d-25e1-420a-be60-fb5771349282 tempest-ServerAddressesTestJSON-1053293248 tempest-ServerAddressesTestJSON-1053293248-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 809.950520] env[62066]: DEBUG nova.virt.hardware [None req-0188668d-25e1-420a-be60-fb5771349282 tempest-ServerAddressesTestJSON-1053293248 tempest-ServerAddressesTestJSON-1053293248-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 809.951232] env[62066]: DEBUG nova.virt.hardware [None req-0188668d-25e1-420a-be60-fb5771349282 tempest-ServerAddressesTestJSON-1053293248 tempest-ServerAddressesTestJSON-1053293248-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 809.951422] env[62066]: DEBUG nova.virt.hardware [None req-0188668d-25e1-420a-be60-fb5771349282 tempest-ServerAddressesTestJSON-1053293248 tempest-ServerAddressesTestJSON-1053293248-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 809.951559] env[62066]: DEBUG nova.virt.hardware [None req-0188668d-25e1-420a-be60-fb5771349282 tempest-ServerAddressesTestJSON-1053293248 tempest-ServerAddressesTestJSON-1053293248-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 809.951906] env[62066]: DEBUG nova.virt.hardware [None req-0188668d-25e1-420a-be60-fb5771349282 tempest-ServerAddressesTestJSON-1053293248 tempest-ServerAddressesTestJSON-1053293248-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 809.952260] env[62066]: DEBUG nova.virt.hardware [None req-0188668d-25e1-420a-be60-fb5771349282 tempest-ServerAddressesTestJSON-1053293248 tempest-ServerAddressesTestJSON-1053293248-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 809.952516] env[62066]: DEBUG nova.virt.hardware [None req-0188668d-25e1-420a-be60-fb5771349282 tempest-ServerAddressesTestJSON-1053293248 tempest-ServerAddressesTestJSON-1053293248-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 809.952666] env[62066]: DEBUG nova.virt.hardware [None req-0188668d-25e1-420a-be60-fb5771349282 tempest-ServerAddressesTestJSON-1053293248 tempest-ServerAddressesTestJSON-1053293248-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 809.952845] env[62066]: DEBUG nova.virt.hardware [None req-0188668d-25e1-420a-be60-fb5771349282 tempest-ServerAddressesTestJSON-1053293248 tempest-ServerAddressesTestJSON-1053293248-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 809.954222] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23aa5e52-4a71-4998-841a-7941863e8fe8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.972767] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bee302df-dc47-4bce-86aa-9ddd53892185 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.978520] env[62066]: DEBUG oslo_vmware.rw_handles [None req-8034bd7f-2294-4f4c-b924-f35bec7ebaac tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] Completed reading data from the image iterator. {{(pid=62066) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 809.978822] env[62066]: DEBUG oslo_vmware.rw_handles [None req-8034bd7f-2294-4f4c-b924-f35bec7ebaac tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] Closing write handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/0ed0c1aa-638f-496c-bf5c-978d07f863bd/50ff584c-3b50-4395-af07-3e66769bc9f7/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=62066) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 810.030706] env[62066]: DEBUG nova.virt.vmwareapi.images [None req-8034bd7f-2294-4f4c-b924-f35bec7ebaac tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] [instance: 738ebb2f-7cc6-4d0f-871c-1428c9c21384] Downloaded image file data 50ff584c-3b50-4395-af07-3e66769bc9f7 to vmware_temp/0ed0c1aa-638f-496c-bf5c-978d07f863bd/50ff584c-3b50-4395-af07-3e66769bc9f7/tmp-sparse.vmdk on the data store datastore1 {{(pid=62066) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 810.032938] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-8034bd7f-2294-4f4c-b924-f35bec7ebaac tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] [instance: 738ebb2f-7cc6-4d0f-871c-1428c9c21384] Caching image {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 810.033437] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-8034bd7f-2294-4f4c-b924-f35bec7ebaac tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] Copying Virtual Disk [datastore1] vmware_temp/0ed0c1aa-638f-496c-bf5c-978d07f863bd/50ff584c-3b50-4395-af07-3e66769bc9f7/tmp-sparse.vmdk to [datastore1] vmware_temp/0ed0c1aa-638f-496c-bf5c-978d07f863bd/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 810.033751] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-bf409546-43bd-425c-8094-54af34762c99 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.042315] env[62066]: DEBUG oslo_vmware.api [None req-8034bd7f-2294-4f4c-b924-f35bec7ebaac tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] Waiting for the task: (returnval){ [ 810.042315] env[62066]: value = "task-1340782" [ 810.042315] env[62066]: _type = "Task" [ 810.042315] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.051649] env[62066]: DEBUG oslo_vmware.api [None req-8034bd7f-2294-4f4c-b924-f35bec7ebaac tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] Task: {'id': task-1340782, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.181087] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a74b0d8e-9150-40ce-bf03-13c1b20297cf {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.189934] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-284c9eb2-a426-469a-a7ee-df88949c108c tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 49f46244-34fa-48a1-95a2-8e95850f345d] Suspending the VM {{(pid=62066) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 810.190255] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-5b2afc66-e12a-4b82-8bf2-0c410871c202 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.197172] env[62066]: DEBUG oslo_vmware.api [None req-284c9eb2-a426-469a-a7ee-df88949c108c tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Waiting for the task: (returnval){ [ 810.197172] env[62066]: value = "task-1340783" [ 810.197172] env[62066]: _type = "Task" [ 810.197172] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.207102] env[62066]: DEBUG oslo_vmware.api [None req-284c9eb2-a426-469a-a7ee-df88949c108c tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': task-1340783, 'name': SuspendVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.292376] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ba3a3138-4a4e-4a10-bbee-93dd3a372fa2 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Lock "0ebba3b2-f82d-4c1d-b01a-f75c3559c117" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 20.026s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 810.338347] env[62066]: DEBUG nova.network.neutron [req-44a82cb1-be80-400c-948b-4f6e4ad5ff62 req-9fc98555-3ab1-42a0-92d0-ed33959d16c1 service nova] [instance: 459f432b-e817-405f-9c3e-dfa5d957788c] Updated VIF entry in instance network info cache for port cb74e58f-b481-4a0f-861b-96381e2d52dc. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 810.338544] env[62066]: DEBUG nova.network.neutron [req-44a82cb1-be80-400c-948b-4f6e4ad5ff62 req-9fc98555-3ab1-42a0-92d0-ed33959d16c1 service nova] [instance: 459f432b-e817-405f-9c3e-dfa5d957788c] Updating instance_info_cache with network_info: [{"id": "cb74e58f-b481-4a0f-861b-96381e2d52dc", "address": "fa:16:3e:87:b9:2f", "network": {"id": "449dfe4e-bb49-43b6-9ba7-b57af74ebfb3", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-282029676-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.159", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0d4e33a99c7741fb8cdd97f4ec5dbbd8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db68bd64-5b56-49af-a075-13dcf85cb2e0", "external-id": "nsx-vlan-transportzone-590", "segmentation_id": 590, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcb74e58f-b4", "ovs_interfaceid": "cb74e58f-b481-4a0f-861b-96381e2d52dc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 810.365511] env[62066]: DEBUG oslo_concurrency.lockutils [None req-df48af02-4c4b-4364-8d38-efc343bf8241 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Lock "49f46244-34fa-48a1-95a2-8e95850f345d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 124.274s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 810.419905] env[62066]: DEBUG oslo_vmware.api [None req-d041ede3-45cc-4fa6-b4cc-7d603e4097a9 tempest-ServerMetadataNegativeTestJSON-1549470887 tempest-ServerMetadataNegativeTestJSON-1549470887-project-member] Task: {'id': task-1340781, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.253797} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.420364] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-d041ede3-45cc-4fa6-b4cc-7d603e4097a9 tempest-ServerMetadataNegativeTestJSON-1549470887 tempest-ServerMetadataNegativeTestJSON-1549470887-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 810.420641] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-d041ede3-45cc-4fa6-b4cc-7d603e4097a9 tempest-ServerMetadataNegativeTestJSON-1549470887 tempest-ServerMetadataNegativeTestJSON-1549470887-project-member] [instance: 31d47299-83a3-4f27-aeb7-95c4cd36c5ac] Deleted contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 810.420933] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-d041ede3-45cc-4fa6-b4cc-7d603e4097a9 tempest-ServerMetadataNegativeTestJSON-1549470887 tempest-ServerMetadataNegativeTestJSON-1549470887-project-member] [instance: 31d47299-83a3-4f27-aeb7-95c4cd36c5ac] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 810.421238] env[62066]: INFO nova.compute.manager [None req-d041ede3-45cc-4fa6-b4cc-7d603e4097a9 tempest-ServerMetadataNegativeTestJSON-1549470887 tempest-ServerMetadataNegativeTestJSON-1549470887-project-member] [instance: 31d47299-83a3-4f27-aeb7-95c4cd36c5ac] Took 1.12 seconds to destroy the instance on the hypervisor. [ 810.421581] env[62066]: DEBUG oslo.service.loopingcall [None req-d041ede3-45cc-4fa6-b4cc-7d603e4097a9 tempest-ServerMetadataNegativeTestJSON-1549470887 tempest-ServerMetadataNegativeTestJSON-1549470887-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 810.421831] env[62066]: DEBUG nova.compute.manager [-] [instance: 31d47299-83a3-4f27-aeb7-95c4cd36c5ac] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 810.421961] env[62066]: DEBUG nova.network.neutron [-] [instance: 31d47299-83a3-4f27-aeb7-95c4cd36c5ac] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 810.554292] env[62066]: DEBUG oslo_vmware.api [None req-8034bd7f-2294-4f4c-b924-f35bec7ebaac tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] Task: {'id': task-1340782, 'name': CopyVirtualDisk_Task} progress is 35%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.561409] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be05f710-e695-4f0a-a668-4b53a83d5bff {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.568999] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6748e469-c2a8-4e46-9a7d-1843fb86b03d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.605607] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1d2e7ce-723b-4eeb-b2af-cfe9412a2068 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.613973] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b85b1f5-a1b7-494c-9138-f285b5353d97 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.637588] env[62066]: DEBUG nova.compute.provider_tree [None req-e5dc4f2e-1d02-4ba9-acb0-6628ba7125f6 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 810.710693] env[62066]: DEBUG oslo_vmware.api [None req-284c9eb2-a426-469a-a7ee-df88949c108c tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': task-1340783, 'name': SuspendVM_Task} progress is 58%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.847027] env[62066]: DEBUG oslo_concurrency.lockutils [req-44a82cb1-be80-400c-948b-4f6e4ad5ff62 req-9fc98555-3ab1-42a0-92d0-ed33959d16c1 service nova] Releasing lock "refresh_cache-459f432b-e817-405f-9c3e-dfa5d957788c" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 810.870571] env[62066]: DEBUG nova.compute.manager [None req-02667002-03be-4979-8517-c36490492214 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 86a92b61-f3e9-48f9-8ee2-756669d558ef] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 810.906168] env[62066]: DEBUG nova.network.neutron [None req-0188668d-25e1-420a-be60-fb5771349282 tempest-ServerAddressesTestJSON-1053293248 tempest-ServerAddressesTestJSON-1053293248-project-member] [instance: 49bc1350-0095-406b-bc68-005eb6b681a8] Successfully updated port: f2c406e7-6447-47bf-9ee1-dba97dc7e5bc {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 811.054334] env[62066]: DEBUG oslo_vmware.api [None req-8034bd7f-2294-4f4c-b924-f35bec7ebaac tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] Task: {'id': task-1340782, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.952803} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.055374] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-8034bd7f-2294-4f4c-b924-f35bec7ebaac tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] Copied Virtual Disk [datastore1] vmware_temp/0ed0c1aa-638f-496c-bf5c-978d07f863bd/50ff584c-3b50-4395-af07-3e66769bc9f7/tmp-sparse.vmdk to [datastore1] vmware_temp/0ed0c1aa-638f-496c-bf5c-978d07f863bd/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 811.055374] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-8034bd7f-2294-4f4c-b924-f35bec7ebaac tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] Deleting the datastore file [datastore1] vmware_temp/0ed0c1aa-638f-496c-bf5c-978d07f863bd/50ff584c-3b50-4395-af07-3e66769bc9f7/tmp-sparse.vmdk {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 811.055374] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f8ede30c-2efc-4881-85c3-0f2bda2a4e75 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.061483] env[62066]: DEBUG oslo_vmware.api [None req-8034bd7f-2294-4f4c-b924-f35bec7ebaac tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] Waiting for the task: (returnval){ [ 811.061483] env[62066]: value = "task-1340784" [ 811.061483] env[62066]: _type = "Task" [ 811.061483] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.069773] env[62066]: DEBUG oslo_vmware.api [None req-8034bd7f-2294-4f4c-b924-f35bec7ebaac tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] Task: {'id': task-1340784, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.078357] env[62066]: DEBUG nova.compute.manager [req-8ec05625-d8f6-406c-86a3-26a23ea2f260 req-b5e9ac18-ce2a-416f-8697-6cb363f24dde service nova] [instance: 050040cb-6ee3-4ad2-960a-fcebb53ac394] Received event network-changed-72521b90-587a-4f59-b744-919ef3087539 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 811.078524] env[62066]: DEBUG nova.compute.manager [req-8ec05625-d8f6-406c-86a3-26a23ea2f260 req-b5e9ac18-ce2a-416f-8697-6cb363f24dde service nova] [instance: 050040cb-6ee3-4ad2-960a-fcebb53ac394] Refreshing instance network info cache due to event network-changed-72521b90-587a-4f59-b744-919ef3087539. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 811.078774] env[62066]: DEBUG oslo_concurrency.lockutils [req-8ec05625-d8f6-406c-86a3-26a23ea2f260 req-b5e9ac18-ce2a-416f-8697-6cb363f24dde service nova] Acquiring lock "refresh_cache-050040cb-6ee3-4ad2-960a-fcebb53ac394" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 811.078896] env[62066]: DEBUG oslo_concurrency.lockutils [req-8ec05625-d8f6-406c-86a3-26a23ea2f260 req-b5e9ac18-ce2a-416f-8697-6cb363f24dde service nova] Acquired lock "refresh_cache-050040cb-6ee3-4ad2-960a-fcebb53ac394" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 811.079262] env[62066]: DEBUG nova.network.neutron [req-8ec05625-d8f6-406c-86a3-26a23ea2f260 req-b5e9ac18-ce2a-416f-8697-6cb363f24dde service nova] [instance: 050040cb-6ee3-4ad2-960a-fcebb53ac394] Refreshing network info cache for port 72521b90-587a-4f59-b744-919ef3087539 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 811.142430] env[62066]: DEBUG nova.scheduler.client.report [None req-e5dc4f2e-1d02-4ba9-acb0-6628ba7125f6 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 811.205946] env[62066]: DEBUG oslo_vmware.api [None req-284c9eb2-a426-469a-a7ee-df88949c108c tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': task-1340783, 'name': SuspendVM_Task, 'duration_secs': 0.82202} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.206253] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-284c9eb2-a426-469a-a7ee-df88949c108c tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 49f46244-34fa-48a1-95a2-8e95850f345d] Suspended the VM {{(pid=62066) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 811.206462] env[62066]: DEBUG nova.compute.manager [None req-284c9eb2-a426-469a-a7ee-df88949c108c tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 49f46244-34fa-48a1-95a2-8e95850f345d] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 811.207343] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bacfa665-4459-4203-b6ce-e07ab85525f6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.291648] env[62066]: DEBUG nova.network.neutron [None req-90f5cbab-7c0d-4b69-92aa-09bf24d9f7da tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] [instance: 050040cb-6ee3-4ad2-960a-fcebb53ac394] Successfully updated port: d329136d-bd14-4bee-b09a-39ef97845251 {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 811.388269] env[62066]: DEBUG nova.network.neutron [-] [instance: 31d47299-83a3-4f27-aeb7-95c4cd36c5ac] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 811.395106] env[62066]: DEBUG oslo_concurrency.lockutils [None req-02667002-03be-4979-8517-c36490492214 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 811.409607] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0188668d-25e1-420a-be60-fb5771349282 tempest-ServerAddressesTestJSON-1053293248 tempest-ServerAddressesTestJSON-1053293248-project-member] Acquiring lock "refresh_cache-49bc1350-0095-406b-bc68-005eb6b681a8" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 811.409751] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0188668d-25e1-420a-be60-fb5771349282 tempest-ServerAddressesTestJSON-1053293248 tempest-ServerAddressesTestJSON-1053293248-project-member] Acquired lock "refresh_cache-49bc1350-0095-406b-bc68-005eb6b681a8" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 811.409900] env[62066]: DEBUG nova.network.neutron [None req-0188668d-25e1-420a-be60-fb5771349282 tempest-ServerAddressesTestJSON-1053293248 tempest-ServerAddressesTestJSON-1053293248-project-member] [instance: 49bc1350-0095-406b-bc68-005eb6b681a8] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 811.446375] env[62066]: DEBUG oslo_vmware.rw_handles [None req-dcd99ffb-af75-402a-906d-756c3b1954c8 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/522b566b-f748-bd1a-5c5d-57241d2d64f6/disk-0.vmdk. {{(pid=62066) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 811.447280] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5e125c7-56ac-41d5-82ca-6ad4ee381324 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.453910] env[62066]: DEBUG oslo_vmware.rw_handles [None req-dcd99ffb-af75-402a-906d-756c3b1954c8 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/522b566b-f748-bd1a-5c5d-57241d2d64f6/disk-0.vmdk is in state: ready. {{(pid=62066) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 811.453910] env[62066]: ERROR oslo_vmware.rw_handles [None req-dcd99ffb-af75-402a-906d-756c3b1954c8 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/522b566b-f748-bd1a-5c5d-57241d2d64f6/disk-0.vmdk due to incomplete transfer. [ 811.454148] env[62066]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-8e4fc797-a33b-4f5b-8004-e54e156ff7e5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.461252] env[62066]: DEBUG oslo_vmware.rw_handles [None req-dcd99ffb-af75-402a-906d-756c3b1954c8 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/522b566b-f748-bd1a-5c5d-57241d2d64f6/disk-0.vmdk. {{(pid=62066) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 811.461451] env[62066]: DEBUG nova.virt.vmwareapi.images [None req-dcd99ffb-af75-402a-906d-756c3b1954c8 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] [instance: 1a4b0637-1a56-41ef-b89b-6b56d24ed206] Uploaded image ce0c3df1-323d-4736-9213-367dbe18cd1e to the Glance image server {{(pid=62066) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 811.463618] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-dcd99ffb-af75-402a-906d-756c3b1954c8 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] [instance: 1a4b0637-1a56-41ef-b89b-6b56d24ed206] Destroying the VM {{(pid=62066) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 811.463867] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-e5c0ae38-8120-4e2a-a444-ee37888800ca {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.470789] env[62066]: DEBUG oslo_vmware.api [None req-dcd99ffb-af75-402a-906d-756c3b1954c8 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Waiting for the task: (returnval){ [ 811.470789] env[62066]: value = "task-1340785" [ 811.470789] env[62066]: _type = "Task" [ 811.470789] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.478190] env[62066]: DEBUG oslo_vmware.api [None req-dcd99ffb-af75-402a-906d-756c3b1954c8 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Task: {'id': task-1340785, 'name': Destroy_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.573846] env[62066]: DEBUG oslo_vmware.api [None req-8034bd7f-2294-4f4c-b924-f35bec7ebaac tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] Task: {'id': task-1340784, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.024504} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.574139] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-8034bd7f-2294-4f4c-b924-f35bec7ebaac tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 811.574361] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-8034bd7f-2294-4f4c-b924-f35bec7ebaac tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] Moving file from [datastore1] vmware_temp/0ed0c1aa-638f-496c-bf5c-978d07f863bd/50ff584c-3b50-4395-af07-3e66769bc9f7 to [datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7. {{(pid=62066) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 811.574626] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-b28eca3a-af65-4758-a051-1f266e55408c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.582981] env[62066]: DEBUG oslo_vmware.api [None req-8034bd7f-2294-4f4c-b924-f35bec7ebaac tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] Waiting for the task: (returnval){ [ 811.582981] env[62066]: value = "task-1340786" [ 811.582981] env[62066]: _type = "Task" [ 811.582981] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.591257] env[62066]: DEBUG oslo_vmware.api [None req-8034bd7f-2294-4f4c-b924-f35bec7ebaac tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] Task: {'id': task-1340786, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.615693] env[62066]: DEBUG nova.network.neutron [req-8ec05625-d8f6-406c-86a3-26a23ea2f260 req-b5e9ac18-ce2a-416f-8697-6cb363f24dde service nova] [instance: 050040cb-6ee3-4ad2-960a-fcebb53ac394] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 811.647395] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e5dc4f2e-1d02-4ba9-acb0-6628ba7125f6 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.895s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 811.649506] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9e052131-3359-4888-9d78-89440b639dcc tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.559s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 811.651092] env[62066]: INFO nova.compute.claims [None req-9e052131-3359-4888-9d78-89440b639dcc tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] [instance: 8ff6fbb9-c90f-498d-9a85-d220a8c2f794] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 811.666549] env[62066]: INFO nova.scheduler.client.report [None req-e5dc4f2e-1d02-4ba9-acb0-6628ba7125f6 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Deleted allocations for instance 98fb270c-f3f8-4375-8b5c-c2279305d476 [ 811.698104] env[62066]: DEBUG nova.network.neutron [req-8ec05625-d8f6-406c-86a3-26a23ea2f260 req-b5e9ac18-ce2a-416f-8697-6cb363f24dde service nova] [instance: 050040cb-6ee3-4ad2-960a-fcebb53ac394] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 811.794687] env[62066]: DEBUG oslo_concurrency.lockutils [None req-90f5cbab-7c0d-4b69-92aa-09bf24d9f7da tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] Acquiring lock "refresh_cache-050040cb-6ee3-4ad2-960a-fcebb53ac394" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 811.891642] env[62066]: INFO nova.compute.manager [-] [instance: 31d47299-83a3-4f27-aeb7-95c4cd36c5ac] Took 1.47 seconds to deallocate network for instance. [ 811.941950] env[62066]: DEBUG nova.network.neutron [None req-0188668d-25e1-420a-be60-fb5771349282 tempest-ServerAddressesTestJSON-1053293248 tempest-ServerAddressesTestJSON-1053293248-project-member] [instance: 49bc1350-0095-406b-bc68-005eb6b681a8] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 811.983095] env[62066]: DEBUG oslo_vmware.api [None req-dcd99ffb-af75-402a-906d-756c3b1954c8 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Task: {'id': task-1340785, 'name': Destroy_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.088227] env[62066]: DEBUG nova.network.neutron [None req-0188668d-25e1-420a-be60-fb5771349282 tempest-ServerAddressesTestJSON-1053293248 tempest-ServerAddressesTestJSON-1053293248-project-member] [instance: 49bc1350-0095-406b-bc68-005eb6b681a8] Updating instance_info_cache with network_info: [{"id": "f2c406e7-6447-47bf-9ee1-dba97dc7e5bc", "address": "fa:16:3e:3d:e5:24", "network": {"id": "9b50c52c-3d6f-48fc-a350-c669a9dbe974", "bridge": "br-int", "label": "tempest-ServerAddressesTestJSON-1706648399-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "39a464e941734bfaad54e66d920290fa", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b1f3e6c3-5584-4852-9017-476ab8ac4946", "external-id": "nsx-vlan-transportzone-304", "segmentation_id": 304, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf2c406e7-64", "ovs_interfaceid": "f2c406e7-6447-47bf-9ee1-dba97dc7e5bc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 812.095082] env[62066]: DEBUG oslo_vmware.api [None req-8034bd7f-2294-4f4c-b924-f35bec7ebaac tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] Task: {'id': task-1340786, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.093571} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.095354] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-8034bd7f-2294-4f4c-b924-f35bec7ebaac tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] File moved {{(pid=62066) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 812.095556] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-8034bd7f-2294-4f4c-b924-f35bec7ebaac tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] [instance: 738ebb2f-7cc6-4d0f-871c-1428c9c21384] Cleaning up location [datastore1] vmware_temp/0ed0c1aa-638f-496c-bf5c-978d07f863bd {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 812.095725] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-8034bd7f-2294-4f4c-b924-f35bec7ebaac tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] Deleting the datastore file [datastore1] vmware_temp/0ed0c1aa-638f-496c-bf5c-978d07f863bd {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 812.095985] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-17be4232-b5e3-4ade-8b40-e718a1e04eeb {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.103643] env[62066]: DEBUG oslo_vmware.api [None req-8034bd7f-2294-4f4c-b924-f35bec7ebaac tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] Waiting for the task: (returnval){ [ 812.103643] env[62066]: value = "task-1340787" [ 812.103643] env[62066]: _type = "Task" [ 812.103643] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.112786] env[62066]: DEBUG oslo_vmware.api [None req-8034bd7f-2294-4f4c-b924-f35bec7ebaac tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] Task: {'id': task-1340787, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.173695] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e5dc4f2e-1d02-4ba9-acb0-6628ba7125f6 tempest-MultipleCreateTestJSON-1404049077 tempest-MultipleCreateTestJSON-1404049077-project-member] Lock "98fb270c-f3f8-4375-8b5c-c2279305d476" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 21.815s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 812.200964] env[62066]: DEBUG oslo_concurrency.lockutils [req-8ec05625-d8f6-406c-86a3-26a23ea2f260 req-b5e9ac18-ce2a-416f-8697-6cb363f24dde service nova] Releasing lock "refresh_cache-050040cb-6ee3-4ad2-960a-fcebb53ac394" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 812.201274] env[62066]: DEBUG nova.compute.manager [req-8ec05625-d8f6-406c-86a3-26a23ea2f260 req-b5e9ac18-ce2a-416f-8697-6cb363f24dde service nova] [instance: 49bc1350-0095-406b-bc68-005eb6b681a8] Received event network-vif-plugged-f2c406e7-6447-47bf-9ee1-dba97dc7e5bc {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 812.201480] env[62066]: DEBUG oslo_concurrency.lockutils [req-8ec05625-d8f6-406c-86a3-26a23ea2f260 req-b5e9ac18-ce2a-416f-8697-6cb363f24dde service nova] Acquiring lock "49bc1350-0095-406b-bc68-005eb6b681a8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 812.201751] env[62066]: DEBUG oslo_concurrency.lockutils [req-8ec05625-d8f6-406c-86a3-26a23ea2f260 req-b5e9ac18-ce2a-416f-8697-6cb363f24dde service nova] Lock "49bc1350-0095-406b-bc68-005eb6b681a8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 812.201990] env[62066]: DEBUG oslo_concurrency.lockutils [req-8ec05625-d8f6-406c-86a3-26a23ea2f260 req-b5e9ac18-ce2a-416f-8697-6cb363f24dde service nova] Lock "49bc1350-0095-406b-bc68-005eb6b681a8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 812.202194] env[62066]: DEBUG nova.compute.manager [req-8ec05625-d8f6-406c-86a3-26a23ea2f260 req-b5e9ac18-ce2a-416f-8697-6cb363f24dde service nova] [instance: 49bc1350-0095-406b-bc68-005eb6b681a8] No waiting events found dispatching network-vif-plugged-f2c406e7-6447-47bf-9ee1-dba97dc7e5bc {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 812.202385] env[62066]: WARNING nova.compute.manager [req-8ec05625-d8f6-406c-86a3-26a23ea2f260 req-b5e9ac18-ce2a-416f-8697-6cb363f24dde service nova] [instance: 49bc1350-0095-406b-bc68-005eb6b681a8] Received unexpected event network-vif-plugged-f2c406e7-6447-47bf-9ee1-dba97dc7e5bc for instance with vm_state building and task_state spawning. [ 812.202536] env[62066]: DEBUG nova.compute.manager [req-8ec05625-d8f6-406c-86a3-26a23ea2f260 req-b5e9ac18-ce2a-416f-8697-6cb363f24dde service nova] [instance: 31d47299-83a3-4f27-aeb7-95c4cd36c5ac] Received event network-vif-deleted-4d836af3-8ac2-4b67-842b-c868c679bd5e {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 812.202703] env[62066]: INFO nova.compute.manager [req-8ec05625-d8f6-406c-86a3-26a23ea2f260 req-b5e9ac18-ce2a-416f-8697-6cb363f24dde service nova] [instance: 31d47299-83a3-4f27-aeb7-95c4cd36c5ac] Neutron deleted interface 4d836af3-8ac2-4b67-842b-c868c679bd5e; detaching it from the instance and deleting it from the info cache [ 812.202873] env[62066]: DEBUG nova.network.neutron [req-8ec05625-d8f6-406c-86a3-26a23ea2f260 req-b5e9ac18-ce2a-416f-8697-6cb363f24dde service nova] [instance: 31d47299-83a3-4f27-aeb7-95c4cd36c5ac] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 812.203981] env[62066]: DEBUG oslo_concurrency.lockutils [None req-90f5cbab-7c0d-4b69-92aa-09bf24d9f7da tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] Acquired lock "refresh_cache-050040cb-6ee3-4ad2-960a-fcebb53ac394" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 812.204137] env[62066]: DEBUG nova.network.neutron [None req-90f5cbab-7c0d-4b69-92aa-09bf24d9f7da tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] [instance: 050040cb-6ee3-4ad2-960a-fcebb53ac394] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 812.399033] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d041ede3-45cc-4fa6-b4cc-7d603e4097a9 tempest-ServerMetadataNegativeTestJSON-1549470887 tempest-ServerMetadataNegativeTestJSON-1549470887-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 812.483993] env[62066]: DEBUG oslo_vmware.api [None req-dcd99ffb-af75-402a-906d-756c3b1954c8 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Task: {'id': task-1340785, 'name': Destroy_Task} progress is 33%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.591232] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0188668d-25e1-420a-be60-fb5771349282 tempest-ServerAddressesTestJSON-1053293248 tempest-ServerAddressesTestJSON-1053293248-project-member] Releasing lock "refresh_cache-49bc1350-0095-406b-bc68-005eb6b681a8" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 812.591580] env[62066]: DEBUG nova.compute.manager [None req-0188668d-25e1-420a-be60-fb5771349282 tempest-ServerAddressesTestJSON-1053293248 tempest-ServerAddressesTestJSON-1053293248-project-member] [instance: 49bc1350-0095-406b-bc68-005eb6b681a8] Instance network_info: |[{"id": "f2c406e7-6447-47bf-9ee1-dba97dc7e5bc", "address": "fa:16:3e:3d:e5:24", "network": {"id": "9b50c52c-3d6f-48fc-a350-c669a9dbe974", "bridge": "br-int", "label": "tempest-ServerAddressesTestJSON-1706648399-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "39a464e941734bfaad54e66d920290fa", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b1f3e6c3-5584-4852-9017-476ab8ac4946", "external-id": "nsx-vlan-transportzone-304", "segmentation_id": 304, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf2c406e7-64", "ovs_interfaceid": "f2c406e7-6447-47bf-9ee1-dba97dc7e5bc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 812.592060] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-0188668d-25e1-420a-be60-fb5771349282 tempest-ServerAddressesTestJSON-1053293248 tempest-ServerAddressesTestJSON-1053293248-project-member] [instance: 49bc1350-0095-406b-bc68-005eb6b681a8] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:3d:e5:24', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'b1f3e6c3-5584-4852-9017-476ab8ac4946', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f2c406e7-6447-47bf-9ee1-dba97dc7e5bc', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 812.599669] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-0188668d-25e1-420a-be60-fb5771349282 tempest-ServerAddressesTestJSON-1053293248 tempest-ServerAddressesTestJSON-1053293248-project-member] Creating folder: Project (39a464e941734bfaad54e66d920290fa). Parent ref: group-v285980. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 812.600137] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f06e4c62-f92d-42de-80a7-691baf3ac66a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.610436] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-0188668d-25e1-420a-be60-fb5771349282 tempest-ServerAddressesTestJSON-1053293248 tempest-ServerAddressesTestJSON-1053293248-project-member] Created folder: Project (39a464e941734bfaad54e66d920290fa) in parent group-v285980. [ 812.610666] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-0188668d-25e1-420a-be60-fb5771349282 tempest-ServerAddressesTestJSON-1053293248 tempest-ServerAddressesTestJSON-1053293248-project-member] Creating folder: Instances. Parent ref: group-v286034. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 812.614535] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-488bbe5f-d140-4fba-9649-5e221b112c66 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.616115] env[62066]: DEBUG oslo_vmware.api [None req-8034bd7f-2294-4f4c-b924-f35bec7ebaac tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] Task: {'id': task-1340787, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.079119} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.616379] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-8034bd7f-2294-4f4c-b924-f35bec7ebaac tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 812.617474] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6c3ec356-a24a-4f50-982b-0441cc8bbf07 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.622402] env[62066]: DEBUG oslo_vmware.api [None req-8034bd7f-2294-4f4c-b924-f35bec7ebaac tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] Waiting for the task: (returnval){ [ 812.622402] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]52c177e0-f85b-946c-c2e2-5c6ea4752395" [ 812.622402] env[62066]: _type = "Task" [ 812.622402] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.628653] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-0188668d-25e1-420a-be60-fb5771349282 tempest-ServerAddressesTestJSON-1053293248 tempest-ServerAddressesTestJSON-1053293248-project-member] Created folder: Instances in parent group-v286034. [ 812.628956] env[62066]: DEBUG oslo.service.loopingcall [None req-0188668d-25e1-420a-be60-fb5771349282 tempest-ServerAddressesTestJSON-1053293248 tempest-ServerAddressesTestJSON-1053293248-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 812.629515] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 49bc1350-0095-406b-bc68-005eb6b681a8] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 812.629741] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-005d1b85-999c-4454-b327-07da70e8f570 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.646496] env[62066]: DEBUG oslo_vmware.api [None req-8034bd7f-2294-4f4c-b924-f35bec7ebaac tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52c177e0-f85b-946c-c2e2-5c6ea4752395, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.652201] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 812.652201] env[62066]: value = "task-1340790" [ 812.652201] env[62066]: _type = "Task" [ 812.652201] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.662369] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1340790, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.705989] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-744d31a0-5625-4b2d-b7e1-dc3d02cf7ec2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.717945] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51b64a07-4346-4cbf-b546-ae07f42aad5a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.750016] env[62066]: DEBUG nova.compute.manager [req-8ec05625-d8f6-406c-86a3-26a23ea2f260 req-b5e9ac18-ce2a-416f-8697-6cb363f24dde service nova] [instance: 31d47299-83a3-4f27-aeb7-95c4cd36c5ac] Detach interface failed, port_id=4d836af3-8ac2-4b67-842b-c868c679bd5e, reason: Instance 31d47299-83a3-4f27-aeb7-95c4cd36c5ac could not be found. {{(pid=62066) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 812.750294] env[62066]: DEBUG nova.compute.manager [req-8ec05625-d8f6-406c-86a3-26a23ea2f260 req-b5e9ac18-ce2a-416f-8697-6cb363f24dde service nova] [instance: 49bc1350-0095-406b-bc68-005eb6b681a8] Received event network-changed-f2c406e7-6447-47bf-9ee1-dba97dc7e5bc {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 812.750486] env[62066]: DEBUG nova.compute.manager [req-8ec05625-d8f6-406c-86a3-26a23ea2f260 req-b5e9ac18-ce2a-416f-8697-6cb363f24dde service nova] [instance: 49bc1350-0095-406b-bc68-005eb6b681a8] Refreshing instance network info cache due to event network-changed-f2c406e7-6447-47bf-9ee1-dba97dc7e5bc. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 812.750729] env[62066]: DEBUG oslo_concurrency.lockutils [req-8ec05625-d8f6-406c-86a3-26a23ea2f260 req-b5e9ac18-ce2a-416f-8697-6cb363f24dde service nova] Acquiring lock "refresh_cache-49bc1350-0095-406b-bc68-005eb6b681a8" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 812.750878] env[62066]: DEBUG oslo_concurrency.lockutils [req-8ec05625-d8f6-406c-86a3-26a23ea2f260 req-b5e9ac18-ce2a-416f-8697-6cb363f24dde service nova] Acquired lock "refresh_cache-49bc1350-0095-406b-bc68-005eb6b681a8" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 812.751056] env[62066]: DEBUG nova.network.neutron [req-8ec05625-d8f6-406c-86a3-26a23ea2f260 req-b5e9ac18-ce2a-416f-8697-6cb363f24dde service nova] [instance: 49bc1350-0095-406b-bc68-005eb6b681a8] Refreshing network info cache for port f2c406e7-6447-47bf-9ee1-dba97dc7e5bc {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 812.753265] env[62066]: DEBUG nova.network.neutron [None req-90f5cbab-7c0d-4b69-92aa-09bf24d9f7da tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] [instance: 050040cb-6ee3-4ad2-960a-fcebb53ac394] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 812.956499] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f68407a-05ee-4af0-a634-3b5e704c2cec {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.966574] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57517cdc-886d-43b7-a0d7-fdf601051ce6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.007663] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-027362dd-0f09-4c70-9b26-ca015ada0a16 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.017025] env[62066]: DEBUG oslo_vmware.api [None req-dcd99ffb-af75-402a-906d-756c3b1954c8 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Task: {'id': task-1340785, 'name': Destroy_Task, 'duration_secs': 1.054298} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.017025] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-dcd99ffb-af75-402a-906d-756c3b1954c8 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] [instance: 1a4b0637-1a56-41ef-b89b-6b56d24ed206] Destroyed the VM [ 813.017025] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-dcd99ffb-af75-402a-906d-756c3b1954c8 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] [instance: 1a4b0637-1a56-41ef-b89b-6b56d24ed206] Deleting Snapshot of the VM instance {{(pid=62066) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 813.017025] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-b79bcb70-38ee-4bb5-99af-1af0712b8036 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.021770] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7456bb16-3778-4179-b183-ee7a07a82ac5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.026679] env[62066]: DEBUG oslo_vmware.api [None req-dcd99ffb-af75-402a-906d-756c3b1954c8 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Waiting for the task: (returnval){ [ 813.026679] env[62066]: value = "task-1340791" [ 813.026679] env[62066]: _type = "Task" [ 813.026679] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.037507] env[62066]: DEBUG nova.compute.provider_tree [None req-9e052131-3359-4888-9d78-89440b639dcc tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Updating inventory in ProviderTree for provider 8e4f7194-1498-4f08-8723-ab7260524bcb with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 813.043430] env[62066]: DEBUG oslo_vmware.api [None req-dcd99ffb-af75-402a-906d-756c3b1954c8 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Task: {'id': task-1340791, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.086439] env[62066]: DEBUG nova.network.neutron [None req-90f5cbab-7c0d-4b69-92aa-09bf24d9f7da tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] [instance: 050040cb-6ee3-4ad2-960a-fcebb53ac394] Updating instance_info_cache with network_info: [{"id": "72521b90-587a-4f59-b744-919ef3087539", "address": "fa:16:3e:be:38:d9", "network": {"id": "a94fb60b-330a-459e-95f0-9088f43282d6", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1044524819", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.139", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b69df8557ba0480893da58e1e8bf591d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9643129c-1d95-4422-9df1-2c21289bd5d6", "external-id": "nsx-vlan-transportzone-917", "segmentation_id": 917, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap72521b90-58", "ovs_interfaceid": "72521b90-587a-4f59-b744-919ef3087539", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "d329136d-bd14-4bee-b09a-39ef97845251", "address": "fa:16:3e:c2:f4:e3", "network": {"id": "8f173085-0641-4db1-b78f-55605d436189", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1329311803", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.121", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "b69df8557ba0480893da58e1e8bf591d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "76cff786-08f3-4a4a-8e68-6ae052a5dff3", "external-id": "nsx-vlan-transportzone-664", "segmentation_id": 664, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd329136d-bd", "ovs_interfaceid": "d329136d-bd14-4bee-b09a-39ef97845251", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 813.133102] env[62066]: DEBUG oslo_vmware.api [None req-8034bd7f-2294-4f4c-b924-f35bec7ebaac tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52c177e0-f85b-946c-c2e2-5c6ea4752395, 'name': SearchDatastore_Task, 'duration_secs': 0.010484} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.133386] env[62066]: DEBUG oslo_concurrency.lockutils [None req-8034bd7f-2294-4f4c-b924-f35bec7ebaac tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] Releasing lock "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 813.133673] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-8034bd7f-2294-4f4c-b924-f35bec7ebaac tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk to [datastore1] 738ebb2f-7cc6-4d0f-871c-1428c9c21384/738ebb2f-7cc6-4d0f-871c-1428c9c21384.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 813.133961] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f6cc809d-278e-4fb4-906b-f468fa41739d tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Acquired lock "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 813.134943] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-f6cc809d-278e-4fb4-906b-f468fa41739d tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 813.134943] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-56a4788c-465a-4d25-9483-4161c886bc0a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.136740] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f9692400-6f24-4749-8cb9-425ce6b61386 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.143324] env[62066]: DEBUG oslo_vmware.api [None req-8034bd7f-2294-4f4c-b924-f35bec7ebaac tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] Waiting for the task: (returnval){ [ 813.143324] env[62066]: value = "task-1340792" [ 813.143324] env[62066]: _type = "Task" [ 813.143324] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.144433] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-f6cc809d-278e-4fb4-906b-f468fa41739d tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 813.144610] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-f6cc809d-278e-4fb4-906b-f468fa41739d tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 813.147842] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6d30c503-5288-4953-afa7-de88b8d443c9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.158464] env[62066]: DEBUG oslo_vmware.api [None req-8034bd7f-2294-4f4c-b924-f35bec7ebaac tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] Task: {'id': task-1340792, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.159729] env[62066]: DEBUG oslo_vmware.api [None req-f6cc809d-278e-4fb4-906b-f468fa41739d tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Waiting for the task: (returnval){ [ 813.159729] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]52eb0fa9-374e-6ed4-fc0c-e264a55e541a" [ 813.159729] env[62066]: _type = "Task" [ 813.159729] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.165831] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1340790, 'name': CreateVM_Task, 'duration_secs': 0.386596} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.166287] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 49bc1350-0095-406b-bc68-005eb6b681a8] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 813.166944] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0188668d-25e1-420a-be60-fb5771349282 tempest-ServerAddressesTestJSON-1053293248 tempest-ServerAddressesTestJSON-1053293248-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 813.167138] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0188668d-25e1-420a-be60-fb5771349282 tempest-ServerAddressesTestJSON-1053293248 tempest-ServerAddressesTestJSON-1053293248-project-member] Acquired lock "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 813.167500] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0188668d-25e1-420a-be60-fb5771349282 tempest-ServerAddressesTestJSON-1053293248 tempest-ServerAddressesTestJSON-1053293248-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 813.170762] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9922a872-b2c4-45a9-83fb-881a49b44ef8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.172565] env[62066]: DEBUG oslo_vmware.api [None req-f6cc809d-278e-4fb4-906b-f468fa41739d tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52eb0fa9-374e-6ed4-fc0c-e264a55e541a, 'name': SearchDatastore_Task, 'duration_secs': 0.008736} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.173708] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e16fc886-2450-4c59-a63e-c254ae50f976 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.176817] env[62066]: DEBUG oslo_vmware.api [None req-0188668d-25e1-420a-be60-fb5771349282 tempest-ServerAddressesTestJSON-1053293248 tempest-ServerAddressesTestJSON-1053293248-project-member] Waiting for the task: (returnval){ [ 813.176817] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]5200fbf7-4990-c239-a496-a52535b7d645" [ 813.176817] env[62066]: _type = "Task" [ 813.176817] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.183291] env[62066]: DEBUG oslo_vmware.api [None req-f6cc809d-278e-4fb4-906b-f468fa41739d tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Waiting for the task: (returnval){ [ 813.183291] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]526c45d5-a8ef-7481-1a96-3c544e871eeb" [ 813.183291] env[62066]: _type = "Task" [ 813.183291] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.190582] env[62066]: DEBUG oslo_vmware.api [None req-0188668d-25e1-420a-be60-fb5771349282 tempest-ServerAddressesTestJSON-1053293248 tempest-ServerAddressesTestJSON-1053293248-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]5200fbf7-4990-c239-a496-a52535b7d645, 'name': SearchDatastore_Task, 'duration_secs': 0.00753} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.190582] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0188668d-25e1-420a-be60-fb5771349282 tempest-ServerAddressesTestJSON-1053293248 tempest-ServerAddressesTestJSON-1053293248-project-member] Releasing lock "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 813.190582] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-0188668d-25e1-420a-be60-fb5771349282 tempest-ServerAddressesTestJSON-1053293248 tempest-ServerAddressesTestJSON-1053293248-project-member] [instance: 49bc1350-0095-406b-bc68-005eb6b681a8] Processing image 50ff584c-3b50-4395-af07-3e66769bc9f7 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 813.190582] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0188668d-25e1-420a-be60-fb5771349282 tempest-ServerAddressesTestJSON-1053293248 tempest-ServerAddressesTestJSON-1053293248-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 813.193758] env[62066]: DEBUG oslo_vmware.api [None req-f6cc809d-278e-4fb4-906b-f468fa41739d tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]526c45d5-a8ef-7481-1a96-3c544e871eeb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.208524] env[62066]: DEBUG nova.compute.manager [req-175538ae-aa08-4d76-ade0-903122a3704d req-312cbd66-aefa-40b5-8212-14923eb477e2 service nova] [instance: 050040cb-6ee3-4ad2-960a-fcebb53ac394] Received event network-vif-plugged-d329136d-bd14-4bee-b09a-39ef97845251 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 813.208744] env[62066]: DEBUG oslo_concurrency.lockutils [req-175538ae-aa08-4d76-ade0-903122a3704d req-312cbd66-aefa-40b5-8212-14923eb477e2 service nova] Acquiring lock "050040cb-6ee3-4ad2-960a-fcebb53ac394-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 813.208951] env[62066]: DEBUG oslo_concurrency.lockutils [req-175538ae-aa08-4d76-ade0-903122a3704d req-312cbd66-aefa-40b5-8212-14923eb477e2 service nova] Lock "050040cb-6ee3-4ad2-960a-fcebb53ac394-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 813.209139] env[62066]: DEBUG oslo_concurrency.lockutils [req-175538ae-aa08-4d76-ade0-903122a3704d req-312cbd66-aefa-40b5-8212-14923eb477e2 service nova] Lock "050040cb-6ee3-4ad2-960a-fcebb53ac394-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 813.209345] env[62066]: DEBUG nova.compute.manager [req-175538ae-aa08-4d76-ade0-903122a3704d req-312cbd66-aefa-40b5-8212-14923eb477e2 service nova] [instance: 050040cb-6ee3-4ad2-960a-fcebb53ac394] No waiting events found dispatching network-vif-plugged-d329136d-bd14-4bee-b09a-39ef97845251 {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 813.209517] env[62066]: WARNING nova.compute.manager [req-175538ae-aa08-4d76-ade0-903122a3704d req-312cbd66-aefa-40b5-8212-14923eb477e2 service nova] [instance: 050040cb-6ee3-4ad2-960a-fcebb53ac394] Received unexpected event network-vif-plugged-d329136d-bd14-4bee-b09a-39ef97845251 for instance with vm_state building and task_state spawning. [ 813.209683] env[62066]: DEBUG nova.compute.manager [req-175538ae-aa08-4d76-ade0-903122a3704d req-312cbd66-aefa-40b5-8212-14923eb477e2 service nova] [instance: 050040cb-6ee3-4ad2-960a-fcebb53ac394] Received event network-changed-d329136d-bd14-4bee-b09a-39ef97845251 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 813.209840] env[62066]: DEBUG nova.compute.manager [req-175538ae-aa08-4d76-ade0-903122a3704d req-312cbd66-aefa-40b5-8212-14923eb477e2 service nova] [instance: 050040cb-6ee3-4ad2-960a-fcebb53ac394] Refreshing instance network info cache due to event network-changed-d329136d-bd14-4bee-b09a-39ef97845251. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 813.210014] env[62066]: DEBUG oslo_concurrency.lockutils [req-175538ae-aa08-4d76-ade0-903122a3704d req-312cbd66-aefa-40b5-8212-14923eb477e2 service nova] Acquiring lock "refresh_cache-050040cb-6ee3-4ad2-960a-fcebb53ac394" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 813.445283] env[62066]: DEBUG nova.compute.manager [None req-8d3cb2ee-be58-4e07-b64f-9d5c34ab9462 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 49f46244-34fa-48a1-95a2-8e95850f345d] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 813.446046] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04ca5b6c-c105-4861-b62f-cecc5412dca8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.469308] env[62066]: DEBUG nova.network.neutron [req-8ec05625-d8f6-406c-86a3-26a23ea2f260 req-b5e9ac18-ce2a-416f-8697-6cb363f24dde service nova] [instance: 49bc1350-0095-406b-bc68-005eb6b681a8] Updated VIF entry in instance network info cache for port f2c406e7-6447-47bf-9ee1-dba97dc7e5bc. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 813.469726] env[62066]: DEBUG nova.network.neutron [req-8ec05625-d8f6-406c-86a3-26a23ea2f260 req-b5e9ac18-ce2a-416f-8697-6cb363f24dde service nova] [instance: 49bc1350-0095-406b-bc68-005eb6b681a8] Updating instance_info_cache with network_info: [{"id": "f2c406e7-6447-47bf-9ee1-dba97dc7e5bc", "address": "fa:16:3e:3d:e5:24", "network": {"id": "9b50c52c-3d6f-48fc-a350-c669a9dbe974", "bridge": "br-int", "label": "tempest-ServerAddressesTestJSON-1706648399-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "39a464e941734bfaad54e66d920290fa", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b1f3e6c3-5584-4852-9017-476ab8ac4946", "external-id": "nsx-vlan-transportzone-304", "segmentation_id": 304, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf2c406e7-64", "ovs_interfaceid": "f2c406e7-6447-47bf-9ee1-dba97dc7e5bc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 813.543960] env[62066]: DEBUG oslo_vmware.api [None req-dcd99ffb-af75-402a-906d-756c3b1954c8 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Task: {'id': task-1340791, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.576457] env[62066]: ERROR nova.scheduler.client.report [None req-9e052131-3359-4888-9d78-89440b639dcc tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] [req-3e5fa2b9-43bf-471a-8695-77fd932578f3] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 8e4f7194-1498-4f08-8723-ab7260524bcb. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-3e5fa2b9-43bf-471a-8695-77fd932578f3"}]} [ 813.589465] env[62066]: DEBUG oslo_concurrency.lockutils [None req-90f5cbab-7c0d-4b69-92aa-09bf24d9f7da tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] Releasing lock "refresh_cache-050040cb-6ee3-4ad2-960a-fcebb53ac394" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 813.591115] env[62066]: DEBUG nova.compute.manager [None req-90f5cbab-7c0d-4b69-92aa-09bf24d9f7da tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] [instance: 050040cb-6ee3-4ad2-960a-fcebb53ac394] Instance network_info: |[{"id": "72521b90-587a-4f59-b744-919ef3087539", "address": "fa:16:3e:be:38:d9", "network": {"id": "a94fb60b-330a-459e-95f0-9088f43282d6", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1044524819", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.139", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b69df8557ba0480893da58e1e8bf591d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9643129c-1d95-4422-9df1-2c21289bd5d6", "external-id": "nsx-vlan-transportzone-917", "segmentation_id": 917, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap72521b90-58", "ovs_interfaceid": "72521b90-587a-4f59-b744-919ef3087539", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "d329136d-bd14-4bee-b09a-39ef97845251", "address": "fa:16:3e:c2:f4:e3", "network": {"id": "8f173085-0641-4db1-b78f-55605d436189", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1329311803", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.121", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "b69df8557ba0480893da58e1e8bf591d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "76cff786-08f3-4a4a-8e68-6ae052a5dff3", "external-id": "nsx-vlan-transportzone-664", "segmentation_id": 664, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd329136d-bd", "ovs_interfaceid": "d329136d-bd14-4bee-b09a-39ef97845251", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 813.591115] env[62066]: DEBUG oslo_concurrency.lockutils [req-175538ae-aa08-4d76-ade0-903122a3704d req-312cbd66-aefa-40b5-8212-14923eb477e2 service nova] Acquired lock "refresh_cache-050040cb-6ee3-4ad2-960a-fcebb53ac394" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 813.591115] env[62066]: DEBUG nova.network.neutron [req-175538ae-aa08-4d76-ade0-903122a3704d req-312cbd66-aefa-40b5-8212-14923eb477e2 service nova] [instance: 050040cb-6ee3-4ad2-960a-fcebb53ac394] Refreshing network info cache for port d329136d-bd14-4bee-b09a-39ef97845251 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 813.591944] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-90f5cbab-7c0d-4b69-92aa-09bf24d9f7da tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] [instance: 050040cb-6ee3-4ad2-960a-fcebb53ac394] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:be:38:d9', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9643129c-1d95-4422-9df1-2c21289bd5d6', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '72521b90-587a-4f59-b744-919ef3087539', 'vif_model': 'vmxnet3'}, {'network_name': 'br-int', 'mac_address': 'fa:16:3e:c2:f4:e3', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '76cff786-08f3-4a4a-8e68-6ae052a5dff3', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd329136d-bd14-4bee-b09a-39ef97845251', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 813.603415] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-90f5cbab-7c0d-4b69-92aa-09bf24d9f7da tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] Creating folder: Project (b69df8557ba0480893da58e1e8bf591d). Parent ref: group-v285980. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 813.605612] env[62066]: DEBUG nova.scheduler.client.report [None req-9e052131-3359-4888-9d78-89440b639dcc tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Refreshing inventories for resource provider 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 813.607914] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b97d0af8-f48b-40b4-9f78-ff70e31dba62 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.620789] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-90f5cbab-7c0d-4b69-92aa-09bf24d9f7da tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] Created folder: Project (b69df8557ba0480893da58e1e8bf591d) in parent group-v285980. [ 813.621049] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-90f5cbab-7c0d-4b69-92aa-09bf24d9f7da tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] Creating folder: Instances. Parent ref: group-v286037. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 813.621326] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-62d5d3af-7e63-486f-90c6-3ae8b5f1e89f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.625325] env[62066]: DEBUG nova.scheduler.client.report [None req-9e052131-3359-4888-9d78-89440b639dcc tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Updating ProviderTree inventory for provider 8e4f7194-1498-4f08-8723-ab7260524bcb from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 813.625561] env[62066]: DEBUG nova.compute.provider_tree [None req-9e052131-3359-4888-9d78-89440b639dcc tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Updating inventory in ProviderTree for provider 8e4f7194-1498-4f08-8723-ab7260524bcb with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 813.630406] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-90f5cbab-7c0d-4b69-92aa-09bf24d9f7da tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] Created folder: Instances in parent group-v286037. [ 813.630668] env[62066]: DEBUG oslo.service.loopingcall [None req-90f5cbab-7c0d-4b69-92aa-09bf24d9f7da tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 813.630877] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 050040cb-6ee3-4ad2-960a-fcebb53ac394] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 813.631115] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-bde1f40b-a696-44ff-a28b-7f07f75b2dcd {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.649098] env[62066]: DEBUG nova.scheduler.client.report [None req-9e052131-3359-4888-9d78-89440b639dcc tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Refreshing aggregate associations for resource provider 8e4f7194-1498-4f08-8723-ab7260524bcb, aggregates: None {{(pid=62066) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 813.660395] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 813.660395] env[62066]: value = "task-1340795" [ 813.660395] env[62066]: _type = "Task" [ 813.660395] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.663929] env[62066]: DEBUG oslo_vmware.api [None req-8034bd7f-2294-4f4c-b924-f35bec7ebaac tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] Task: {'id': task-1340792, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.673221] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1340795, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.674360] env[62066]: DEBUG nova.scheduler.client.report [None req-9e052131-3359-4888-9d78-89440b639dcc tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Refreshing trait associations for resource provider 8e4f7194-1498-4f08-8723-ab7260524bcb, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,HW_ARCH_X86_64,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=62066) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 813.696167] env[62066]: DEBUG oslo_vmware.api [None req-f6cc809d-278e-4fb4-906b-f468fa41739d tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]526c45d5-a8ef-7481-1a96-3c544e871eeb, 'name': SearchDatastore_Task, 'duration_secs': 0.016298} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.699107] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f6cc809d-278e-4fb4-906b-f468fa41739d tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Releasing lock "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 813.699545] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-f6cc809d-278e-4fb4-906b-f468fa41739d tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk to [datastore1] 5f645eaa-6edc-4362-82ad-38c4c57b2be2/5f645eaa-6edc-4362-82ad-38c4c57b2be2.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 813.700240] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0188668d-25e1-420a-be60-fb5771349282 tempest-ServerAddressesTestJSON-1053293248 tempest-ServerAddressesTestJSON-1053293248-project-member] Acquired lock "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 813.700481] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-0188668d-25e1-420a-be60-fb5771349282 tempest-ServerAddressesTestJSON-1053293248 tempest-ServerAddressesTestJSON-1053293248-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 813.700730] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a12621ca-5b14-4f3c-bb00-72fe27f9ed03 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.702754] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-29054880-c325-4b51-84b9-b4a1272e4b30 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.716028] env[62066]: DEBUG oslo_vmware.api [None req-f6cc809d-278e-4fb4-906b-f468fa41739d tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Waiting for the task: (returnval){ [ 813.716028] env[62066]: value = "task-1340796" [ 813.716028] env[62066]: _type = "Task" [ 813.716028] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.716028] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-0188668d-25e1-420a-be60-fb5771349282 tempest-ServerAddressesTestJSON-1053293248 tempest-ServerAddressesTestJSON-1053293248-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 813.716028] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-0188668d-25e1-420a-be60-fb5771349282 tempest-ServerAddressesTestJSON-1053293248 tempest-ServerAddressesTestJSON-1053293248-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 813.718811] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8bcfc29e-700f-4c97-a247-750eacc0bbf5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.729470] env[62066]: DEBUG oslo_vmware.api [None req-f6cc809d-278e-4fb4-906b-f468fa41739d tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Task: {'id': task-1340796, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.729793] env[62066]: DEBUG oslo_vmware.api [None req-0188668d-25e1-420a-be60-fb5771349282 tempest-ServerAddressesTestJSON-1053293248 tempest-ServerAddressesTestJSON-1053293248-project-member] Waiting for the task: (returnval){ [ 813.729793] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]52f12a33-52a1-b6b4-d53d-5626a84f0d3c" [ 813.729793] env[62066]: _type = "Task" [ 813.729793] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.739119] env[62066]: DEBUG oslo_vmware.api [None req-0188668d-25e1-420a-be60-fb5771349282 tempest-ServerAddressesTestJSON-1053293248 tempest-ServerAddressesTestJSON-1053293248-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52f12a33-52a1-b6b4-d53d-5626a84f0d3c, 'name': SearchDatastore_Task, 'duration_secs': 0.009025} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.740065] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bd399b77-1e92-4ab4-914c-08cd0c361e56 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.749502] env[62066]: DEBUG oslo_vmware.api [None req-0188668d-25e1-420a-be60-fb5771349282 tempest-ServerAddressesTestJSON-1053293248 tempest-ServerAddressesTestJSON-1053293248-project-member] Waiting for the task: (returnval){ [ 813.749502] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]527a31bc-4ead-a13b-8aee-9b659e481f9c" [ 813.749502] env[62066]: _type = "Task" [ 813.749502] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.764373] env[62066]: DEBUG oslo_vmware.api [None req-0188668d-25e1-420a-be60-fb5771349282 tempest-ServerAddressesTestJSON-1053293248 tempest-ServerAddressesTestJSON-1053293248-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]527a31bc-4ead-a13b-8aee-9b659e481f9c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.956671] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a981300d-4046-4ef5-98d3-42ff4561b213 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.961126] env[62066]: INFO nova.compute.manager [None req-8d3cb2ee-be58-4e07-b64f-9d5c34ab9462 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 49f46244-34fa-48a1-95a2-8e95850f345d] instance snapshotting [ 813.961863] env[62066]: WARNING nova.compute.manager [None req-8d3cb2ee-be58-4e07-b64f-9d5c34ab9462 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 49f46244-34fa-48a1-95a2-8e95850f345d] trying to snapshot a non-running instance: (state: 7 expected: 1) [ 813.966021] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00e89fe8-e0f2-46f5-a139-c8341bcd437b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.971048] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26d15004-a8a1-4037-a24f-48f36e048579 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.990016] env[62066]: DEBUG oslo_concurrency.lockutils [req-8ec05625-d8f6-406c-86a3-26a23ea2f260 req-b5e9ac18-ce2a-416f-8697-6cb363f24dde service nova] Releasing lock "refresh_cache-49bc1350-0095-406b-bc68-005eb6b681a8" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 813.991680] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53dd57d6-562b-4283-8a94-bc1ac06a08e7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.022544] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f26c387b-f011-4910-a5e9-fe013ef17c34 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.040479] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94503dcf-7c80-43ea-90f6-a0004eae6d27 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.044714] env[62066]: DEBUG oslo_vmware.api [None req-dcd99ffb-af75-402a-906d-756c3b1954c8 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Task: {'id': task-1340791, 'name': RemoveSnapshot_Task} progress is 97%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.055198] env[62066]: DEBUG nova.compute.provider_tree [None req-9e052131-3359-4888-9d78-89440b639dcc tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Updating inventory in ProviderTree for provider 8e4f7194-1498-4f08-8723-ab7260524bcb with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 814.168648] env[62066]: DEBUG oslo_vmware.api [None req-8034bd7f-2294-4f4c-b924-f35bec7ebaac tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] Task: {'id': task-1340792, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.566731} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.173486] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-8034bd7f-2294-4f4c-b924-f35bec7ebaac tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk to [datastore1] 738ebb2f-7cc6-4d0f-871c-1428c9c21384/738ebb2f-7cc6-4d0f-871c-1428c9c21384.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 814.176380] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-8034bd7f-2294-4f4c-b924-f35bec7ebaac tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] [instance: 738ebb2f-7cc6-4d0f-871c-1428c9c21384] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 814.176380] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-bc3e2336-fca7-44b2-9efb-305b5f6e6e22 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.181885] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1340795, 'name': CreateVM_Task} progress is 99%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.184149] env[62066]: DEBUG oslo_vmware.api [None req-8034bd7f-2294-4f4c-b924-f35bec7ebaac tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] Waiting for the task: (returnval){ [ 814.184149] env[62066]: value = "task-1340797" [ 814.184149] env[62066]: _type = "Task" [ 814.184149] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 814.196469] env[62066]: DEBUG oslo_vmware.api [None req-8034bd7f-2294-4f4c-b924-f35bec7ebaac tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] Task: {'id': task-1340797, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.228598] env[62066]: DEBUG oslo_vmware.api [None req-f6cc809d-278e-4fb4-906b-f468fa41739d tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Task: {'id': task-1340796, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.264536] env[62066]: DEBUG oslo_vmware.api [None req-0188668d-25e1-420a-be60-fb5771349282 tempest-ServerAddressesTestJSON-1053293248 tempest-ServerAddressesTestJSON-1053293248-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]527a31bc-4ead-a13b-8aee-9b659e481f9c, 'name': SearchDatastore_Task, 'duration_secs': 0.011768} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.264536] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0188668d-25e1-420a-be60-fb5771349282 tempest-ServerAddressesTestJSON-1053293248 tempest-ServerAddressesTestJSON-1053293248-project-member] Releasing lock "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 814.267017] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-0188668d-25e1-420a-be60-fb5771349282 tempest-ServerAddressesTestJSON-1053293248 tempest-ServerAddressesTestJSON-1053293248-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk to [datastore1] 49bc1350-0095-406b-bc68-005eb6b681a8/49bc1350-0095-406b-bc68-005eb6b681a8.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 814.267017] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-952338fb-d43a-4de7-a959-d8dc72937384 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.274930] env[62066]: DEBUG oslo_vmware.api [None req-0188668d-25e1-420a-be60-fb5771349282 tempest-ServerAddressesTestJSON-1053293248 tempest-ServerAddressesTestJSON-1053293248-project-member] Waiting for the task: (returnval){ [ 814.274930] env[62066]: value = "task-1340798" [ 814.274930] env[62066]: _type = "Task" [ 814.274930] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 814.282904] env[62066]: DEBUG oslo_vmware.api [None req-0188668d-25e1-420a-be60-fb5771349282 tempest-ServerAddressesTestJSON-1053293248 tempest-ServerAddressesTestJSON-1053293248-project-member] Task: {'id': task-1340798, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.357515] env[62066]: DEBUG nova.network.neutron [req-175538ae-aa08-4d76-ade0-903122a3704d req-312cbd66-aefa-40b5-8212-14923eb477e2 service nova] [instance: 050040cb-6ee3-4ad2-960a-fcebb53ac394] Updated VIF entry in instance network info cache for port d329136d-bd14-4bee-b09a-39ef97845251. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 814.357963] env[62066]: DEBUG nova.network.neutron [req-175538ae-aa08-4d76-ade0-903122a3704d req-312cbd66-aefa-40b5-8212-14923eb477e2 service nova] [instance: 050040cb-6ee3-4ad2-960a-fcebb53ac394] Updating instance_info_cache with network_info: [{"id": "72521b90-587a-4f59-b744-919ef3087539", "address": "fa:16:3e:be:38:d9", "network": {"id": "a94fb60b-330a-459e-95f0-9088f43282d6", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1044524819", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.139", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b69df8557ba0480893da58e1e8bf591d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9643129c-1d95-4422-9df1-2c21289bd5d6", "external-id": "nsx-vlan-transportzone-917", "segmentation_id": 917, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap72521b90-58", "ovs_interfaceid": "72521b90-587a-4f59-b744-919ef3087539", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "d329136d-bd14-4bee-b09a-39ef97845251", "address": "fa:16:3e:c2:f4:e3", "network": {"id": "8f173085-0641-4db1-b78f-55605d436189", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1329311803", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.121", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "b69df8557ba0480893da58e1e8bf591d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "76cff786-08f3-4a4a-8e68-6ae052a5dff3", "external-id": "nsx-vlan-transportzone-664", "segmentation_id": 664, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd329136d-bd", "ovs_interfaceid": "d329136d-bd14-4bee-b09a-39ef97845251", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 814.535673] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-8d3cb2ee-be58-4e07-b64f-9d5c34ab9462 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 49f46244-34fa-48a1-95a2-8e95850f345d] Creating Snapshot of the VM instance {{(pid=62066) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 814.536022] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-904d970d-a6e7-4ae8-a9c0-f1283775d8c6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.546768] env[62066]: DEBUG oslo_vmware.api [None req-dcd99ffb-af75-402a-906d-756c3b1954c8 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Task: {'id': task-1340791, 'name': RemoveSnapshot_Task, 'duration_secs': 1.320364} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.548511] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-dcd99ffb-af75-402a-906d-756c3b1954c8 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] [instance: 1a4b0637-1a56-41ef-b89b-6b56d24ed206] Deleted Snapshot of the VM instance {{(pid=62066) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 814.548803] env[62066]: INFO nova.compute.manager [None req-dcd99ffb-af75-402a-906d-756c3b1954c8 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] [instance: 1a4b0637-1a56-41ef-b89b-6b56d24ed206] Took 17.10 seconds to snapshot the instance on the hypervisor. [ 814.551814] env[62066]: DEBUG oslo_vmware.api [None req-8d3cb2ee-be58-4e07-b64f-9d5c34ab9462 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Waiting for the task: (returnval){ [ 814.551814] env[62066]: value = "task-1340799" [ 814.551814] env[62066]: _type = "Task" [ 814.551814] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 814.563719] env[62066]: DEBUG oslo_vmware.api [None req-8d3cb2ee-be58-4e07-b64f-9d5c34ab9462 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': task-1340799, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.598330] env[62066]: DEBUG nova.scheduler.client.report [None req-9e052131-3359-4888-9d78-89440b639dcc tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Updated inventory for provider 8e4f7194-1498-4f08-8723-ab7260524bcb with generation 81 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 814.598330] env[62066]: DEBUG nova.compute.provider_tree [None req-9e052131-3359-4888-9d78-89440b639dcc tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Updating resource provider 8e4f7194-1498-4f08-8723-ab7260524bcb generation from 81 to 82 during operation: update_inventory {{(pid=62066) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 814.598575] env[62066]: DEBUG nova.compute.provider_tree [None req-9e052131-3359-4888-9d78-89440b639dcc tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Updating inventory in ProviderTree for provider 8e4f7194-1498-4f08-8723-ab7260524bcb with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 814.677967] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1340795, 'name': CreateVM_Task, 'duration_secs': 0.572574} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.678170] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 050040cb-6ee3-4ad2-960a-fcebb53ac394] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 814.679009] env[62066]: DEBUG oslo_concurrency.lockutils [None req-90f5cbab-7c0d-4b69-92aa-09bf24d9f7da tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 814.679186] env[62066]: DEBUG oslo_concurrency.lockutils [None req-90f5cbab-7c0d-4b69-92aa-09bf24d9f7da tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] Acquired lock "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 814.680177] env[62066]: DEBUG oslo_concurrency.lockutils [None req-90f5cbab-7c0d-4b69-92aa-09bf24d9f7da tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 814.680177] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7048cf07-f92c-4fc0-b4a4-049464c26f09 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.685038] env[62066]: DEBUG oslo_vmware.api [None req-90f5cbab-7c0d-4b69-92aa-09bf24d9f7da tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] Waiting for the task: (returnval){ [ 814.685038] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]52d3f074-9dd0-5557-7e3d-827c720fa213" [ 814.685038] env[62066]: _type = "Task" [ 814.685038] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 814.696521] env[62066]: DEBUG oslo_vmware.api [None req-8034bd7f-2294-4f4c-b924-f35bec7ebaac tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] Task: {'id': task-1340797, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.078896} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.699942] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-8034bd7f-2294-4f4c-b924-f35bec7ebaac tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] [instance: 738ebb2f-7cc6-4d0f-871c-1428c9c21384] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 814.700489] env[62066]: DEBUG oslo_vmware.api [None req-90f5cbab-7c0d-4b69-92aa-09bf24d9f7da tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52d3f074-9dd0-5557-7e3d-827c720fa213, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.701707] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1613fb7c-60c8-488c-a093-c2b5a83c89f5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.727561] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-8034bd7f-2294-4f4c-b924-f35bec7ebaac tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] [instance: 738ebb2f-7cc6-4d0f-871c-1428c9c21384] Reconfiguring VM instance instance-00000041 to attach disk [datastore1] 738ebb2f-7cc6-4d0f-871c-1428c9c21384/738ebb2f-7cc6-4d0f-871c-1428c9c21384.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 814.731166] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e055b5fe-2dbe-4ca1-9a6d-7946a51f6312 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.751997] env[62066]: DEBUG oslo_vmware.api [None req-f6cc809d-278e-4fb4-906b-f468fa41739d tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Task: {'id': task-1340796, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.53935} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.754046] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-f6cc809d-278e-4fb4-906b-f468fa41739d tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk to [datastore1] 5f645eaa-6edc-4362-82ad-38c4c57b2be2/5f645eaa-6edc-4362-82ad-38c4c57b2be2.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 814.754368] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-f6cc809d-278e-4fb4-906b-f468fa41739d tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] [instance: 5f645eaa-6edc-4362-82ad-38c4c57b2be2] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 814.754734] env[62066]: DEBUG oslo_vmware.api [None req-8034bd7f-2294-4f4c-b924-f35bec7ebaac tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] Waiting for the task: (returnval){ [ 814.754734] env[62066]: value = "task-1340800" [ 814.754734] env[62066]: _type = "Task" [ 814.754734] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 814.755016] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4e807365-dea0-426f-8f68-dc4d0baab010 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.765821] env[62066]: DEBUG oslo_vmware.api [None req-f6cc809d-278e-4fb4-906b-f468fa41739d tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Waiting for the task: (returnval){ [ 814.765821] env[62066]: value = "task-1340801" [ 814.765821] env[62066]: _type = "Task" [ 814.765821] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 814.769494] env[62066]: DEBUG oslo_vmware.api [None req-8034bd7f-2294-4f4c-b924-f35bec7ebaac tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] Task: {'id': task-1340800, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.784568] env[62066]: DEBUG oslo_vmware.api [None req-f6cc809d-278e-4fb4-906b-f468fa41739d tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Task: {'id': task-1340801, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.787424] env[62066]: DEBUG oslo_vmware.api [None req-0188668d-25e1-420a-be60-fb5771349282 tempest-ServerAddressesTestJSON-1053293248 tempest-ServerAddressesTestJSON-1053293248-project-member] Task: {'id': task-1340798, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.862508] env[62066]: DEBUG oslo_concurrency.lockutils [req-175538ae-aa08-4d76-ade0-903122a3704d req-312cbd66-aefa-40b5-8212-14923eb477e2 service nova] Releasing lock "refresh_cache-050040cb-6ee3-4ad2-960a-fcebb53ac394" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 815.064622] env[62066]: DEBUG oslo_vmware.api [None req-8d3cb2ee-be58-4e07-b64f-9d5c34ab9462 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': task-1340799, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.105245] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9e052131-3359-4888-9d78-89440b639dcc tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.455s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 815.105245] env[62066]: DEBUG nova.compute.manager [None req-9e052131-3359-4888-9d78-89440b639dcc tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] [instance: 8ff6fbb9-c90f-498d-9a85-d220a8c2f794] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 815.108286] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3d67ea1c-5f75-4620-b677-6f1e0faf1bd6 tempest-ServersV294TestFqdnHostnames-843300434 tempest-ServersV294TestFqdnHostnames-843300434-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.149s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 815.110042] env[62066]: INFO nova.compute.claims [None req-3d67ea1c-5f75-4620-b677-6f1e0faf1bd6 tempest-ServersV294TestFqdnHostnames-843300434 tempest-ServersV294TestFqdnHostnames-843300434-project-member] [instance: 2709d0ad-5dd8-4e3c-b1e6-3d1e2e242380] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 815.198368] env[62066]: DEBUG oslo_vmware.api [None req-90f5cbab-7c0d-4b69-92aa-09bf24d9f7da tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52d3f074-9dd0-5557-7e3d-827c720fa213, 'name': SearchDatastore_Task, 'duration_secs': 0.050237} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.198691] env[62066]: DEBUG oslo_concurrency.lockutils [None req-90f5cbab-7c0d-4b69-92aa-09bf24d9f7da tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] Releasing lock "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 815.199219] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-90f5cbab-7c0d-4b69-92aa-09bf24d9f7da tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] [instance: 050040cb-6ee3-4ad2-960a-fcebb53ac394] Processing image 50ff584c-3b50-4395-af07-3e66769bc9f7 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 815.199219] env[62066]: DEBUG oslo_concurrency.lockutils [None req-90f5cbab-7c0d-4b69-92aa-09bf24d9f7da tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 815.199383] env[62066]: DEBUG oslo_concurrency.lockutils [None req-90f5cbab-7c0d-4b69-92aa-09bf24d9f7da tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] Acquired lock "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 815.199589] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-90f5cbab-7c0d-4b69-92aa-09bf24d9f7da tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 815.199867] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ecab4a45-30d5-4989-a0dd-ce5c0b9dca37 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.207893] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-90f5cbab-7c0d-4b69-92aa-09bf24d9f7da tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 815.208105] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-90f5cbab-7c0d-4b69-92aa-09bf24d9f7da tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 815.208790] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e5db33df-50fe-4f83-bef9-3711050c334d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.214094] env[62066]: DEBUG oslo_vmware.api [None req-90f5cbab-7c0d-4b69-92aa-09bf24d9f7da tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] Waiting for the task: (returnval){ [ 815.214094] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]5271b6df-e41e-8ecf-2c95-09f833f70051" [ 815.214094] env[62066]: _type = "Task" [ 815.214094] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.222412] env[62066]: DEBUG oslo_vmware.api [None req-90f5cbab-7c0d-4b69-92aa-09bf24d9f7da tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]5271b6df-e41e-8ecf-2c95-09f833f70051, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.266598] env[62066]: DEBUG oslo_vmware.api [None req-8034bd7f-2294-4f4c-b924-f35bec7ebaac tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] Task: {'id': task-1340800, 'name': ReconfigVM_Task, 'duration_secs': 0.293605} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.266926] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-8034bd7f-2294-4f4c-b924-f35bec7ebaac tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] [instance: 738ebb2f-7cc6-4d0f-871c-1428c9c21384] Reconfigured VM instance instance-00000041 to attach disk [datastore1] 738ebb2f-7cc6-4d0f-871c-1428c9c21384/738ebb2f-7cc6-4d0f-871c-1428c9c21384.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 815.267604] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-926f9a19-038f-4473-ae36-75859f3cc3af {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.273993] env[62066]: DEBUG oslo_vmware.api [None req-8034bd7f-2294-4f4c-b924-f35bec7ebaac tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] Waiting for the task: (returnval){ [ 815.273993] env[62066]: value = "task-1340802" [ 815.273993] env[62066]: _type = "Task" [ 815.273993] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.282173] env[62066]: DEBUG oslo_vmware.api [None req-f6cc809d-278e-4fb4-906b-f468fa41739d tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Task: {'id': task-1340801, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072226} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.282731] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-f6cc809d-278e-4fb4-906b-f468fa41739d tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] [instance: 5f645eaa-6edc-4362-82ad-38c4c57b2be2] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 815.283762] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfe5d4dc-769b-497b-8af0-4b09a641921f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.291549] env[62066]: DEBUG oslo_vmware.api [None req-8034bd7f-2294-4f4c-b924-f35bec7ebaac tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] Task: {'id': task-1340802, 'name': Rename_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.291784] env[62066]: DEBUG oslo_vmware.api [None req-0188668d-25e1-420a-be60-fb5771349282 tempest-ServerAddressesTestJSON-1053293248 tempest-ServerAddressesTestJSON-1053293248-project-member] Task: {'id': task-1340798, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.525143} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.292317] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-0188668d-25e1-420a-be60-fb5771349282 tempest-ServerAddressesTestJSON-1053293248 tempest-ServerAddressesTestJSON-1053293248-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk to [datastore1] 49bc1350-0095-406b-bc68-005eb6b681a8/49bc1350-0095-406b-bc68-005eb6b681a8.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 815.292532] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-0188668d-25e1-420a-be60-fb5771349282 tempest-ServerAddressesTestJSON-1053293248 tempest-ServerAddressesTestJSON-1053293248-project-member] [instance: 49bc1350-0095-406b-bc68-005eb6b681a8] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 815.292764] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e3ce0ce5-5167-44cc-abd8-218c52ebbc8d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.312455] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-f6cc809d-278e-4fb4-906b-f468fa41739d tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] [instance: 5f645eaa-6edc-4362-82ad-38c4c57b2be2] Reconfiguring VM instance instance-00000042 to attach disk [datastore1] 5f645eaa-6edc-4362-82ad-38c4c57b2be2/5f645eaa-6edc-4362-82ad-38c4c57b2be2.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 815.313177] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-12c29510-426a-47e9-a78a-36bf6b9c04eb {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.328490] env[62066]: DEBUG oslo_vmware.api [None req-0188668d-25e1-420a-be60-fb5771349282 tempest-ServerAddressesTestJSON-1053293248 tempest-ServerAddressesTestJSON-1053293248-project-member] Waiting for the task: (returnval){ [ 815.328490] env[62066]: value = "task-1340803" [ 815.328490] env[62066]: _type = "Task" [ 815.328490] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.334029] env[62066]: DEBUG oslo_vmware.api [None req-f6cc809d-278e-4fb4-906b-f468fa41739d tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Waiting for the task: (returnval){ [ 815.334029] env[62066]: value = "task-1340804" [ 815.334029] env[62066]: _type = "Task" [ 815.334029] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.340213] env[62066]: DEBUG oslo_vmware.api [None req-0188668d-25e1-420a-be60-fb5771349282 tempest-ServerAddressesTestJSON-1053293248 tempest-ServerAddressesTestJSON-1053293248-project-member] Task: {'id': task-1340803, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.345029] env[62066]: DEBUG oslo_vmware.api [None req-f6cc809d-278e-4fb4-906b-f468fa41739d tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Task: {'id': task-1340804, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.567131] env[62066]: DEBUG oslo_vmware.api [None req-8d3cb2ee-be58-4e07-b64f-9d5c34ab9462 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': task-1340799, 'name': CreateSnapshot_Task, 'duration_secs': 0.570246} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.567442] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-8d3cb2ee-be58-4e07-b64f-9d5c34ab9462 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 49f46244-34fa-48a1-95a2-8e95850f345d] Created Snapshot of the VM instance {{(pid=62066) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 815.568211] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f50252ad-d2db-4283-90b9-a2e7b10c3d2f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.615716] env[62066]: DEBUG nova.compute.utils [None req-9e052131-3359-4888-9d78-89440b639dcc tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 815.616708] env[62066]: DEBUG nova.compute.manager [None req-9e052131-3359-4888-9d78-89440b639dcc tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] [instance: 8ff6fbb9-c90f-498d-9a85-d220a8c2f794] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 815.616899] env[62066]: DEBUG nova.network.neutron [None req-9e052131-3359-4888-9d78-89440b639dcc tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] [instance: 8ff6fbb9-c90f-498d-9a85-d220a8c2f794] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 815.659824] env[62066]: DEBUG nova.policy [None req-9e052131-3359-4888-9d78-89440b639dcc tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8a9b164f679f4f4a9fa28f1362ceb50d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1f05d8a66d0c4a479e49a947e4b4bbff', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 815.724434] env[62066]: DEBUG oslo_vmware.api [None req-90f5cbab-7c0d-4b69-92aa-09bf24d9f7da tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]5271b6df-e41e-8ecf-2c95-09f833f70051, 'name': SearchDatastore_Task, 'duration_secs': 0.008242} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.725241] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-80cc011a-303a-42f3-9f90-13c588bf3769 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.730366] env[62066]: DEBUG oslo_vmware.api [None req-90f5cbab-7c0d-4b69-92aa-09bf24d9f7da tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] Waiting for the task: (returnval){ [ 815.730366] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]5228246b-77e2-2897-54dd-e2654d88c66f" [ 815.730366] env[62066]: _type = "Task" [ 815.730366] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.737990] env[62066]: DEBUG oslo_vmware.api [None req-90f5cbab-7c0d-4b69-92aa-09bf24d9f7da tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]5228246b-77e2-2897-54dd-e2654d88c66f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.783550] env[62066]: DEBUG oslo_vmware.api [None req-8034bd7f-2294-4f4c-b924-f35bec7ebaac tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] Task: {'id': task-1340802, 'name': Rename_Task, 'duration_secs': 0.144907} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.783832] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-8034bd7f-2294-4f4c-b924-f35bec7ebaac tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] [instance: 738ebb2f-7cc6-4d0f-871c-1428c9c21384] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 815.784107] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8cee7ed5-5541-4786-b329-9adef7cf49a1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.790477] env[62066]: DEBUG oslo_vmware.api [None req-8034bd7f-2294-4f4c-b924-f35bec7ebaac tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] Waiting for the task: (returnval){ [ 815.790477] env[62066]: value = "task-1340805" [ 815.790477] env[62066]: _type = "Task" [ 815.790477] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.799379] env[62066]: DEBUG oslo_vmware.api [None req-8034bd7f-2294-4f4c-b924-f35bec7ebaac tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] Task: {'id': task-1340805, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.849568] env[62066]: DEBUG oslo_vmware.api [None req-0188668d-25e1-420a-be60-fb5771349282 tempest-ServerAddressesTestJSON-1053293248 tempest-ServerAddressesTestJSON-1053293248-project-member] Task: {'id': task-1340803, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073922} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.849785] env[62066]: DEBUG oslo_vmware.api [None req-f6cc809d-278e-4fb4-906b-f468fa41739d tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Task: {'id': task-1340804, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.850051] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-0188668d-25e1-420a-be60-fb5771349282 tempest-ServerAddressesTestJSON-1053293248 tempest-ServerAddressesTestJSON-1053293248-project-member] [instance: 49bc1350-0095-406b-bc68-005eb6b681a8] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 815.850906] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6ccadaa-bc6a-4bfc-9b21-de2d315775e3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.874714] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-0188668d-25e1-420a-be60-fb5771349282 tempest-ServerAddressesTestJSON-1053293248 tempest-ServerAddressesTestJSON-1053293248-project-member] [instance: 49bc1350-0095-406b-bc68-005eb6b681a8] Reconfiguring VM instance instance-00000044 to attach disk [datastore1] 49bc1350-0095-406b-bc68-005eb6b681a8/49bc1350-0095-406b-bc68-005eb6b681a8.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 815.875102] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-969cb465-9550-4922-9f1f-c00e2163e506 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.897021] env[62066]: DEBUG oslo_vmware.api [None req-0188668d-25e1-420a-be60-fb5771349282 tempest-ServerAddressesTestJSON-1053293248 tempest-ServerAddressesTestJSON-1053293248-project-member] Waiting for the task: (returnval){ [ 815.897021] env[62066]: value = "task-1340806" [ 815.897021] env[62066]: _type = "Task" [ 815.897021] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.905726] env[62066]: DEBUG oslo_vmware.api [None req-0188668d-25e1-420a-be60-fb5771349282 tempest-ServerAddressesTestJSON-1053293248 tempest-ServerAddressesTestJSON-1053293248-project-member] Task: {'id': task-1340806, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.025793] env[62066]: DEBUG nova.network.neutron [None req-9e052131-3359-4888-9d78-89440b639dcc tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] [instance: 8ff6fbb9-c90f-498d-9a85-d220a8c2f794] Successfully created port: 8c324833-7bf9-4007-85aa-5b16871f63c7 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 816.088775] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-8d3cb2ee-be58-4e07-b64f-9d5c34ab9462 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 49f46244-34fa-48a1-95a2-8e95850f345d] Creating linked-clone VM from snapshot {{(pid=62066) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 816.089571] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-61feb474-4606-4681-a01c-563a145ec2bc {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.099462] env[62066]: DEBUG oslo_vmware.api [None req-8d3cb2ee-be58-4e07-b64f-9d5c34ab9462 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Waiting for the task: (returnval){ [ 816.099462] env[62066]: value = "task-1340807" [ 816.099462] env[62066]: _type = "Task" [ 816.099462] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.107355] env[62066]: DEBUG oslo_vmware.api [None req-8d3cb2ee-be58-4e07-b64f-9d5c34ab9462 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': task-1340807, 'name': CloneVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.122083] env[62066]: DEBUG nova.compute.manager [None req-9e052131-3359-4888-9d78-89440b639dcc tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] [instance: 8ff6fbb9-c90f-498d-9a85-d220a8c2f794] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 816.240700] env[62066]: DEBUG oslo_vmware.api [None req-90f5cbab-7c0d-4b69-92aa-09bf24d9f7da tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]5228246b-77e2-2897-54dd-e2654d88c66f, 'name': SearchDatastore_Task, 'duration_secs': 0.008865} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.243576] env[62066]: DEBUG oslo_concurrency.lockutils [None req-90f5cbab-7c0d-4b69-92aa-09bf24d9f7da tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] Releasing lock "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 816.243857] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-90f5cbab-7c0d-4b69-92aa-09bf24d9f7da tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk to [datastore1] 050040cb-6ee3-4ad2-960a-fcebb53ac394/050040cb-6ee3-4ad2-960a-fcebb53ac394.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 816.244512] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5cf6af02-9e41-45a6-90d5-ea398e813599 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.252813] env[62066]: DEBUG oslo_vmware.api [None req-90f5cbab-7c0d-4b69-92aa-09bf24d9f7da tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] Waiting for the task: (returnval){ [ 816.252813] env[62066]: value = "task-1340808" [ 816.252813] env[62066]: _type = "Task" [ 816.252813] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.266592] env[62066]: DEBUG oslo_vmware.api [None req-90f5cbab-7c0d-4b69-92aa-09bf24d9f7da tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] Task: {'id': task-1340808, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.304297] env[62066]: DEBUG oslo_vmware.api [None req-8034bd7f-2294-4f4c-b924-f35bec7ebaac tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] Task: {'id': task-1340805, 'name': PowerOnVM_Task, 'duration_secs': 0.503666} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.304682] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-8034bd7f-2294-4f4c-b924-f35bec7ebaac tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] [instance: 738ebb2f-7cc6-4d0f-871c-1428c9c21384] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 816.305286] env[62066]: INFO nova.compute.manager [None req-8034bd7f-2294-4f4c-b924-f35bec7ebaac tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] [instance: 738ebb2f-7cc6-4d0f-871c-1428c9c21384] Took 12.60 seconds to spawn the instance on the hypervisor. [ 816.305286] env[62066]: DEBUG nova.compute.manager [None req-8034bd7f-2294-4f4c-b924-f35bec7ebaac tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] [instance: 738ebb2f-7cc6-4d0f-871c-1428c9c21384] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 816.306619] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30d71a0c-0ca1-4af4-aa01-f16578ee49e9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.346600] env[62066]: DEBUG oslo_vmware.api [None req-f6cc809d-278e-4fb4-906b-f468fa41739d tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Task: {'id': task-1340804, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.394841] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abdf78d0-e127-4148-aad8-e4492d86f315 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.409017] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5b6665d-c4f6-452e-8fcc-869575f51d60 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.413050] env[62066]: DEBUG oslo_vmware.api [None req-0188668d-25e1-420a-be60-fb5771349282 tempest-ServerAddressesTestJSON-1053293248 tempest-ServerAddressesTestJSON-1053293248-project-member] Task: {'id': task-1340806, 'name': ReconfigVM_Task, 'duration_secs': 0.353797} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.413373] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-0188668d-25e1-420a-be60-fb5771349282 tempest-ServerAddressesTestJSON-1053293248 tempest-ServerAddressesTestJSON-1053293248-project-member] [instance: 49bc1350-0095-406b-bc68-005eb6b681a8] Reconfigured VM instance instance-00000044 to attach disk [datastore1] 49bc1350-0095-406b-bc68-005eb6b681a8/49bc1350-0095-406b-bc68-005eb6b681a8.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 816.414667] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-537e631d-baad-46b7-951d-11e3042be562 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.444768] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3808bdea-0d51-492d-9c57-d4f5d4403389 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.447635] env[62066]: DEBUG oslo_vmware.api [None req-0188668d-25e1-420a-be60-fb5771349282 tempest-ServerAddressesTestJSON-1053293248 tempest-ServerAddressesTestJSON-1053293248-project-member] Waiting for the task: (returnval){ [ 816.447635] env[62066]: value = "task-1340809" [ 816.447635] env[62066]: _type = "Task" [ 816.447635] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.454897] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-215f827d-d27a-48c6-9bb8-e78e061cac17 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.463715] env[62066]: DEBUG oslo_vmware.api [None req-0188668d-25e1-420a-be60-fb5771349282 tempest-ServerAddressesTestJSON-1053293248 tempest-ServerAddressesTestJSON-1053293248-project-member] Task: {'id': task-1340809, 'name': Rename_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.474780] env[62066]: DEBUG nova.compute.provider_tree [None req-3d67ea1c-5f75-4620-b677-6f1e0faf1bd6 tempest-ServersV294TestFqdnHostnames-843300434 tempest-ServersV294TestFqdnHostnames-843300434-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 816.478798] env[62066]: DEBUG nova.compute.manager [None req-41432ebd-8922-4676-9cc7-6eed1922da68 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] [instance: 1e1a2ede-4ad8-4600-851e-6f2046b3f919] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 816.479724] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b809ec49-d71c-467d-8c98-26e047ca6f09 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.610076] env[62066]: DEBUG oslo_vmware.api [None req-8d3cb2ee-be58-4e07-b64f-9d5c34ab9462 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': task-1340807, 'name': CloneVM_Task} progress is 93%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.762693] env[62066]: DEBUG oslo_vmware.api [None req-90f5cbab-7c0d-4b69-92aa-09bf24d9f7da tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] Task: {'id': task-1340808, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.831290] env[62066]: INFO nova.compute.manager [None req-8034bd7f-2294-4f4c-b924-f35bec7ebaac tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] [instance: 738ebb2f-7cc6-4d0f-871c-1428c9c21384] Took 29.99 seconds to build instance. [ 816.847222] env[62066]: DEBUG oslo_vmware.api [None req-f6cc809d-278e-4fb4-906b-f468fa41739d tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Task: {'id': task-1340804, 'name': ReconfigVM_Task, 'duration_secs': 1.397298} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.848204] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-f6cc809d-278e-4fb4-906b-f468fa41739d tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] [instance: 5f645eaa-6edc-4362-82ad-38c4c57b2be2] Reconfigured VM instance instance-00000042 to attach disk [datastore1] 5f645eaa-6edc-4362-82ad-38c4c57b2be2/5f645eaa-6edc-4362-82ad-38c4c57b2be2.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 816.848989] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e69698db-5242-4f17-b6d6-4540cb717b2a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.856041] env[62066]: DEBUG oslo_vmware.api [None req-f6cc809d-278e-4fb4-906b-f468fa41739d tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Waiting for the task: (returnval){ [ 816.856041] env[62066]: value = "task-1340810" [ 816.856041] env[62066]: _type = "Task" [ 816.856041] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.864570] env[62066]: DEBUG oslo_vmware.api [None req-f6cc809d-278e-4fb4-906b-f468fa41739d tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Task: {'id': task-1340810, 'name': Rename_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.958248] env[62066]: DEBUG oslo_vmware.api [None req-0188668d-25e1-420a-be60-fb5771349282 tempest-ServerAddressesTestJSON-1053293248 tempest-ServerAddressesTestJSON-1053293248-project-member] Task: {'id': task-1340809, 'name': Rename_Task, 'duration_secs': 0.332046} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.958529] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-0188668d-25e1-420a-be60-fb5771349282 tempest-ServerAddressesTestJSON-1053293248 tempest-ServerAddressesTestJSON-1053293248-project-member] [instance: 49bc1350-0095-406b-bc68-005eb6b681a8] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 816.958935] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-937d44aa-9dd9-405a-863d-5b50d5cde3fe {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.966662] env[62066]: DEBUG oslo_vmware.api [None req-0188668d-25e1-420a-be60-fb5771349282 tempest-ServerAddressesTestJSON-1053293248 tempest-ServerAddressesTestJSON-1053293248-project-member] Waiting for the task: (returnval){ [ 816.966662] env[62066]: value = "task-1340811" [ 816.966662] env[62066]: _type = "Task" [ 816.966662] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.975014] env[62066]: DEBUG oslo_vmware.api [None req-0188668d-25e1-420a-be60-fb5771349282 tempest-ServerAddressesTestJSON-1053293248 tempest-ServerAddressesTestJSON-1053293248-project-member] Task: {'id': task-1340811, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.977970] env[62066]: DEBUG nova.scheduler.client.report [None req-3d67ea1c-5f75-4620-b677-6f1e0faf1bd6 tempest-ServersV294TestFqdnHostnames-843300434 tempest-ServersV294TestFqdnHostnames-843300434-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 816.990911] env[62066]: INFO nova.compute.manager [None req-41432ebd-8922-4676-9cc7-6eed1922da68 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] [instance: 1e1a2ede-4ad8-4600-851e-6f2046b3f919] instance snapshotting [ 816.994584] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56060ad5-9b3b-47c6-81a3-a5dca3c86028 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.016400] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d2bb646-0cd7-419e-941d-c744046d8c5a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.113373] env[62066]: DEBUG oslo_vmware.api [None req-8d3cb2ee-be58-4e07-b64f-9d5c34ab9462 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': task-1340807, 'name': CloneVM_Task} progress is 94%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.135635] env[62066]: DEBUG nova.compute.manager [None req-9e052131-3359-4888-9d78-89440b639dcc tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] [instance: 8ff6fbb9-c90f-498d-9a85-d220a8c2f794] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 817.157706] env[62066]: DEBUG nova.virt.hardware [None req-9e052131-3359-4888-9d78-89440b639dcc tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-23T13:40:41Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-23T13:40:23Z,direct_url=,disk_format='vmdk',id=50ff584c-3b50-4395-af07-3e66769bc9f7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='eb52f7069a374c61ae946f052007c6d9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-23T13:40:24Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 817.157956] env[62066]: DEBUG nova.virt.hardware [None req-9e052131-3359-4888-9d78-89440b639dcc tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 817.158130] env[62066]: DEBUG nova.virt.hardware [None req-9e052131-3359-4888-9d78-89440b639dcc tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 817.158324] env[62066]: DEBUG nova.virt.hardware [None req-9e052131-3359-4888-9d78-89440b639dcc tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 817.158529] env[62066]: DEBUG nova.virt.hardware [None req-9e052131-3359-4888-9d78-89440b639dcc tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 817.158624] env[62066]: DEBUG nova.virt.hardware [None req-9e052131-3359-4888-9d78-89440b639dcc tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 817.158837] env[62066]: DEBUG nova.virt.hardware [None req-9e052131-3359-4888-9d78-89440b639dcc tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 817.158998] env[62066]: DEBUG nova.virt.hardware [None req-9e052131-3359-4888-9d78-89440b639dcc tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 817.159203] env[62066]: DEBUG nova.virt.hardware [None req-9e052131-3359-4888-9d78-89440b639dcc tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 817.159380] env[62066]: DEBUG nova.virt.hardware [None req-9e052131-3359-4888-9d78-89440b639dcc tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 817.159577] env[62066]: DEBUG nova.virt.hardware [None req-9e052131-3359-4888-9d78-89440b639dcc tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 817.160531] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc9d18b2-a82d-410e-a6ff-859a2d3bc9ef {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.169266] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb165f8d-2dae-4a6d-b4f1-c75f3ec9ac79 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.264741] env[62066]: DEBUG oslo_vmware.api [None req-90f5cbab-7c0d-4b69-92aa-09bf24d9f7da tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] Task: {'id': task-1340808, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.595189} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.265040] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-90f5cbab-7c0d-4b69-92aa-09bf24d9f7da tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk to [datastore1] 050040cb-6ee3-4ad2-960a-fcebb53ac394/050040cb-6ee3-4ad2-960a-fcebb53ac394.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 817.265286] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-90f5cbab-7c0d-4b69-92aa-09bf24d9f7da tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] [instance: 050040cb-6ee3-4ad2-960a-fcebb53ac394] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 817.265609] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9e19a144-e185-4202-9493-2cbbe1d07fbe {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.274643] env[62066]: DEBUG oslo_vmware.api [None req-90f5cbab-7c0d-4b69-92aa-09bf24d9f7da tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] Waiting for the task: (returnval){ [ 817.274643] env[62066]: value = "task-1340812" [ 817.274643] env[62066]: _type = "Task" [ 817.274643] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.283909] env[62066]: DEBUG oslo_vmware.api [None req-90f5cbab-7c0d-4b69-92aa-09bf24d9f7da tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] Task: {'id': task-1340812, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.334034] env[62066]: DEBUG oslo_concurrency.lockutils [None req-8034bd7f-2294-4f4c-b924-f35bec7ebaac tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] Lock "738ebb2f-7cc6-4d0f-871c-1428c9c21384" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 118.605s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 817.368072] env[62066]: DEBUG oslo_vmware.api [None req-f6cc809d-278e-4fb4-906b-f468fa41739d tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Task: {'id': task-1340810, 'name': Rename_Task, 'duration_secs': 0.259278} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.368285] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-f6cc809d-278e-4fb4-906b-f468fa41739d tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] [instance: 5f645eaa-6edc-4362-82ad-38c4c57b2be2] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 817.368519] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-cb86281b-4998-4b35-a4cb-e4ccdc7de22c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.377279] env[62066]: DEBUG oslo_vmware.api [None req-f6cc809d-278e-4fb4-906b-f468fa41739d tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Waiting for the task: (returnval){ [ 817.377279] env[62066]: value = "task-1340813" [ 817.377279] env[62066]: _type = "Task" [ 817.377279] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.387156] env[62066]: DEBUG oslo_vmware.api [None req-f6cc809d-278e-4fb4-906b-f468fa41739d tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Task: {'id': task-1340813, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.478939] env[62066]: DEBUG oslo_vmware.api [None req-0188668d-25e1-420a-be60-fb5771349282 tempest-ServerAddressesTestJSON-1053293248 tempest-ServerAddressesTestJSON-1053293248-project-member] Task: {'id': task-1340811, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.483088] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3d67ea1c-5f75-4620-b677-6f1e0faf1bd6 tempest-ServersV294TestFqdnHostnames-843300434 tempest-ServersV294TestFqdnHostnames-843300434-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.375s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 817.483665] env[62066]: DEBUG nova.compute.manager [None req-3d67ea1c-5f75-4620-b677-6f1e0faf1bd6 tempest-ServersV294TestFqdnHostnames-843300434 tempest-ServersV294TestFqdnHostnames-843300434-project-member] [instance: 2709d0ad-5dd8-4e3c-b1e6-3d1e2e242380] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 817.486559] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b8836db2-293e-442c-80f2-42d8bc4c2918 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.925s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 817.488401] env[62066]: INFO nova.compute.claims [None req-b8836db2-293e-442c-80f2-42d8bc4c2918 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 88daaaa6-f385-4161-bc74-9fffabb1145d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 817.527929] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-41432ebd-8922-4676-9cc7-6eed1922da68 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] [instance: 1e1a2ede-4ad8-4600-851e-6f2046b3f919] Creating Snapshot of the VM instance {{(pid=62066) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 817.528514] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-113263c0-bc55-4ee0-9b71-a3a230b1e505 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.539319] env[62066]: DEBUG oslo_vmware.api [None req-41432ebd-8922-4676-9cc7-6eed1922da68 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Waiting for the task: (returnval){ [ 817.539319] env[62066]: value = "task-1340814" [ 817.539319] env[62066]: _type = "Task" [ 817.539319] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.552145] env[62066]: DEBUG oslo_vmware.api [None req-41432ebd-8922-4676-9cc7-6eed1922da68 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Task: {'id': task-1340814, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.615507] env[62066]: DEBUG oslo_vmware.api [None req-8d3cb2ee-be58-4e07-b64f-9d5c34ab9462 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': task-1340807, 'name': CloneVM_Task} progress is 94%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.719452] env[62066]: DEBUG nova.compute.manager [req-b1341fc1-3c9d-4394-8627-054e110cf982 req-d4bfbc96-2ce0-416a-a1c0-65ac9ddaad27 service nova] [instance: 8ff6fbb9-c90f-498d-9a85-d220a8c2f794] Received event network-vif-plugged-8c324833-7bf9-4007-85aa-5b16871f63c7 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 817.719452] env[62066]: DEBUG oslo_concurrency.lockutils [req-b1341fc1-3c9d-4394-8627-054e110cf982 req-d4bfbc96-2ce0-416a-a1c0-65ac9ddaad27 service nova] Acquiring lock "8ff6fbb9-c90f-498d-9a85-d220a8c2f794-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 817.719452] env[62066]: DEBUG oslo_concurrency.lockutils [req-b1341fc1-3c9d-4394-8627-054e110cf982 req-d4bfbc96-2ce0-416a-a1c0-65ac9ddaad27 service nova] Lock "8ff6fbb9-c90f-498d-9a85-d220a8c2f794-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 817.719452] env[62066]: DEBUG oslo_concurrency.lockutils [req-b1341fc1-3c9d-4394-8627-054e110cf982 req-d4bfbc96-2ce0-416a-a1c0-65ac9ddaad27 service nova] Lock "8ff6fbb9-c90f-498d-9a85-d220a8c2f794-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 817.719452] env[62066]: DEBUG nova.compute.manager [req-b1341fc1-3c9d-4394-8627-054e110cf982 req-d4bfbc96-2ce0-416a-a1c0-65ac9ddaad27 service nova] [instance: 8ff6fbb9-c90f-498d-9a85-d220a8c2f794] No waiting events found dispatching network-vif-plugged-8c324833-7bf9-4007-85aa-5b16871f63c7 {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 817.719452] env[62066]: WARNING nova.compute.manager [req-b1341fc1-3c9d-4394-8627-054e110cf982 req-d4bfbc96-2ce0-416a-a1c0-65ac9ddaad27 service nova] [instance: 8ff6fbb9-c90f-498d-9a85-d220a8c2f794] Received unexpected event network-vif-plugged-8c324833-7bf9-4007-85aa-5b16871f63c7 for instance with vm_state building and task_state spawning. [ 817.791952] env[62066]: DEBUG oslo_concurrency.lockutils [None req-423c892e-6435-422f-ab41-7d39bb22efa3 tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] Acquiring lock "738ebb2f-7cc6-4d0f-871c-1428c9c21384" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 817.793039] env[62066]: DEBUG oslo_concurrency.lockutils [None req-423c892e-6435-422f-ab41-7d39bb22efa3 tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] Lock "738ebb2f-7cc6-4d0f-871c-1428c9c21384" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 817.793039] env[62066]: DEBUG oslo_concurrency.lockutils [None req-423c892e-6435-422f-ab41-7d39bb22efa3 tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] Acquiring lock "738ebb2f-7cc6-4d0f-871c-1428c9c21384-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 817.793039] env[62066]: DEBUG oslo_concurrency.lockutils [None req-423c892e-6435-422f-ab41-7d39bb22efa3 tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] Lock "738ebb2f-7cc6-4d0f-871c-1428c9c21384-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 817.793214] env[62066]: DEBUG oslo_concurrency.lockutils [None req-423c892e-6435-422f-ab41-7d39bb22efa3 tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] Lock "738ebb2f-7cc6-4d0f-871c-1428c9c21384-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 817.795384] env[62066]: INFO nova.compute.manager [None req-423c892e-6435-422f-ab41-7d39bb22efa3 tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] [instance: 738ebb2f-7cc6-4d0f-871c-1428c9c21384] Terminating instance [ 817.797419] env[62066]: DEBUG oslo_vmware.api [None req-90f5cbab-7c0d-4b69-92aa-09bf24d9f7da tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] Task: {'id': task-1340812, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.257845} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.797986] env[62066]: DEBUG nova.compute.manager [None req-423c892e-6435-422f-ab41-7d39bb22efa3 tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] [instance: 738ebb2f-7cc6-4d0f-871c-1428c9c21384] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 817.798305] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-423c892e-6435-422f-ab41-7d39bb22efa3 tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] [instance: 738ebb2f-7cc6-4d0f-871c-1428c9c21384] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 817.798493] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-90f5cbab-7c0d-4b69-92aa-09bf24d9f7da tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] [instance: 050040cb-6ee3-4ad2-960a-fcebb53ac394] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 817.799329] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d182e51c-2ac9-4070-8b26-ee93e3de2db4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.802808] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-265e7fbd-e1ed-4699-a63f-d49a52cefee9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.834604] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-90f5cbab-7c0d-4b69-92aa-09bf24d9f7da tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] [instance: 050040cb-6ee3-4ad2-960a-fcebb53ac394] Reconfiguring VM instance instance-00000043 to attach disk [datastore1] 050040cb-6ee3-4ad2-960a-fcebb53ac394/050040cb-6ee3-4ad2-960a-fcebb53ac394.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 817.834982] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-423c892e-6435-422f-ab41-7d39bb22efa3 tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] [instance: 738ebb2f-7cc6-4d0f-871c-1428c9c21384] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 817.835235] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d5844596-dd7a-427c-8085-b20a5554dbcd {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.849438] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-bbf435b4-9ebf-40f2-b519-f4aca7cbb38b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.858636] env[62066]: DEBUG oslo_vmware.api [None req-90f5cbab-7c0d-4b69-92aa-09bf24d9f7da tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] Waiting for the task: (returnval){ [ 817.858636] env[62066]: value = "task-1340816" [ 817.858636] env[62066]: _type = "Task" [ 817.858636] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.860602] env[62066]: DEBUG oslo_vmware.api [None req-423c892e-6435-422f-ab41-7d39bb22efa3 tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] Waiting for the task: (returnval){ [ 817.860602] env[62066]: value = "task-1340815" [ 817.860602] env[62066]: _type = "Task" [ 817.860602] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.874093] env[62066]: DEBUG oslo_vmware.api [None req-423c892e-6435-422f-ab41-7d39bb22efa3 tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] Task: {'id': task-1340815, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.878724] env[62066]: DEBUG oslo_vmware.api [None req-90f5cbab-7c0d-4b69-92aa-09bf24d9f7da tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] Task: {'id': task-1340816, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.888633] env[62066]: DEBUG oslo_vmware.api [None req-f6cc809d-278e-4fb4-906b-f468fa41739d tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Task: {'id': task-1340813, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.979533] env[62066]: DEBUG oslo_vmware.api [None req-0188668d-25e1-420a-be60-fb5771349282 tempest-ServerAddressesTestJSON-1053293248 tempest-ServerAddressesTestJSON-1053293248-project-member] Task: {'id': task-1340811, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.996709] env[62066]: DEBUG nova.compute.utils [None req-3d67ea1c-5f75-4620-b677-6f1e0faf1bd6 tempest-ServersV294TestFqdnHostnames-843300434 tempest-ServersV294TestFqdnHostnames-843300434-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 817.998696] env[62066]: DEBUG nova.compute.manager [None req-3d67ea1c-5f75-4620-b677-6f1e0faf1bd6 tempest-ServersV294TestFqdnHostnames-843300434 tempest-ServersV294TestFqdnHostnames-843300434-project-member] [instance: 2709d0ad-5dd8-4e3c-b1e6-3d1e2e242380] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 817.998847] env[62066]: DEBUG nova.network.neutron [None req-3d67ea1c-5f75-4620-b677-6f1e0faf1bd6 tempest-ServersV294TestFqdnHostnames-843300434 tempest-ServersV294TestFqdnHostnames-843300434-project-member] [instance: 2709d0ad-5dd8-4e3c-b1e6-3d1e2e242380] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 818.050547] env[62066]: DEBUG oslo_vmware.api [None req-41432ebd-8922-4676-9cc7-6eed1922da68 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Task: {'id': task-1340814, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.052286] env[62066]: DEBUG nova.policy [None req-3d67ea1c-5f75-4620-b677-6f1e0faf1bd6 tempest-ServersV294TestFqdnHostnames-843300434 tempest-ServersV294TestFqdnHostnames-843300434-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e7056cfde5b6404e8f95759a20be7804', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd1785af300d545e6913c0e91148f4563', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 818.114765] env[62066]: DEBUG oslo_vmware.api [None req-8d3cb2ee-be58-4e07-b64f-9d5c34ab9462 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': task-1340807, 'name': CloneVM_Task} progress is 94%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.193156] env[62066]: DEBUG nova.network.neutron [None req-9e052131-3359-4888-9d78-89440b639dcc tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] [instance: 8ff6fbb9-c90f-498d-9a85-d220a8c2f794] Successfully updated port: 8c324833-7bf9-4007-85aa-5b16871f63c7 {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 818.233037] env[62066]: DEBUG nova.compute.manager [req-45e414ba-4bc2-4636-918c-7198361230d4 req-200539b7-3e7c-45ae-88b4-115684d46ffe service nova] [instance: 8ff6fbb9-c90f-498d-9a85-d220a8c2f794] Received event network-changed-8c324833-7bf9-4007-85aa-5b16871f63c7 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 818.233037] env[62066]: DEBUG nova.compute.manager [req-45e414ba-4bc2-4636-918c-7198361230d4 req-200539b7-3e7c-45ae-88b4-115684d46ffe service nova] [instance: 8ff6fbb9-c90f-498d-9a85-d220a8c2f794] Refreshing instance network info cache due to event network-changed-8c324833-7bf9-4007-85aa-5b16871f63c7. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 818.233037] env[62066]: DEBUG oslo_concurrency.lockutils [req-45e414ba-4bc2-4636-918c-7198361230d4 req-200539b7-3e7c-45ae-88b4-115684d46ffe service nova] Acquiring lock "refresh_cache-8ff6fbb9-c90f-498d-9a85-d220a8c2f794" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 818.233037] env[62066]: DEBUG oslo_concurrency.lockutils [req-45e414ba-4bc2-4636-918c-7198361230d4 req-200539b7-3e7c-45ae-88b4-115684d46ffe service nova] Acquired lock "refresh_cache-8ff6fbb9-c90f-498d-9a85-d220a8c2f794" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 818.233423] env[62066]: DEBUG nova.network.neutron [req-45e414ba-4bc2-4636-918c-7198361230d4 req-200539b7-3e7c-45ae-88b4-115684d46ffe service nova] [instance: 8ff6fbb9-c90f-498d-9a85-d220a8c2f794] Refreshing network info cache for port 8c324833-7bf9-4007-85aa-5b16871f63c7 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 818.378875] env[62066]: DEBUG oslo_vmware.api [None req-90f5cbab-7c0d-4b69-92aa-09bf24d9f7da tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] Task: {'id': task-1340816, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.382676] env[62066]: DEBUG oslo_vmware.api [None req-423c892e-6435-422f-ab41-7d39bb22efa3 tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] Task: {'id': task-1340815, 'name': PowerOffVM_Task, 'duration_secs': 0.260477} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.386441] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-423c892e-6435-422f-ab41-7d39bb22efa3 tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] [instance: 738ebb2f-7cc6-4d0f-871c-1428c9c21384] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 818.386706] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-423c892e-6435-422f-ab41-7d39bb22efa3 tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] [instance: 738ebb2f-7cc6-4d0f-871c-1428c9c21384] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 818.387033] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-84313fa1-d608-470d-8784-940094296a4b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.393929] env[62066]: DEBUG nova.network.neutron [None req-3d67ea1c-5f75-4620-b677-6f1e0faf1bd6 tempest-ServersV294TestFqdnHostnames-843300434 tempest-ServersV294TestFqdnHostnames-843300434-project-member] [instance: 2709d0ad-5dd8-4e3c-b1e6-3d1e2e242380] Successfully created port: c5cdc663-1e55-4238-bad4-7fb406b34a36 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 818.400584] env[62066]: DEBUG oslo_vmware.api [None req-f6cc809d-278e-4fb4-906b-f468fa41739d tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Task: {'id': task-1340813, 'name': PowerOnVM_Task, 'duration_secs': 0.906924} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.400729] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-f6cc809d-278e-4fb4-906b-f468fa41739d tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] [instance: 5f645eaa-6edc-4362-82ad-38c4c57b2be2] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 818.400905] env[62066]: INFO nova.compute.manager [None req-f6cc809d-278e-4fb4-906b-f468fa41739d tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] [instance: 5f645eaa-6edc-4362-82ad-38c4c57b2be2] Took 13.33 seconds to spawn the instance on the hypervisor. [ 818.401136] env[62066]: DEBUG nova.compute.manager [None req-f6cc809d-278e-4fb4-906b-f468fa41739d tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] [instance: 5f645eaa-6edc-4362-82ad-38c4c57b2be2] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 818.402062] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-082a50c7-ff21-4988-895d-0bfab61f6946 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.479984] env[62066]: DEBUG oslo_vmware.api [None req-0188668d-25e1-420a-be60-fb5771349282 tempest-ServerAddressesTestJSON-1053293248 tempest-ServerAddressesTestJSON-1053293248-project-member] Task: {'id': task-1340811, 'name': PowerOnVM_Task, 'duration_secs': 1.035288} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.480310] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-0188668d-25e1-420a-be60-fb5771349282 tempest-ServerAddressesTestJSON-1053293248 tempest-ServerAddressesTestJSON-1053293248-project-member] [instance: 49bc1350-0095-406b-bc68-005eb6b681a8] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 818.480563] env[62066]: INFO nova.compute.manager [None req-0188668d-25e1-420a-be60-fb5771349282 tempest-ServerAddressesTestJSON-1053293248 tempest-ServerAddressesTestJSON-1053293248-project-member] [instance: 49bc1350-0095-406b-bc68-005eb6b681a8] Took 8.56 seconds to spawn the instance on the hypervisor. [ 818.480748] env[62066]: DEBUG nova.compute.manager [None req-0188668d-25e1-420a-be60-fb5771349282 tempest-ServerAddressesTestJSON-1053293248 tempest-ServerAddressesTestJSON-1053293248-project-member] [instance: 49bc1350-0095-406b-bc68-005eb6b681a8] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 818.481590] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd6829d3-72c4-4de4-a2a3-67c60d6310ba {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.492879] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-423c892e-6435-422f-ab41-7d39bb22efa3 tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] [instance: 738ebb2f-7cc6-4d0f-871c-1428c9c21384] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 818.493134] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-423c892e-6435-422f-ab41-7d39bb22efa3 tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] [instance: 738ebb2f-7cc6-4d0f-871c-1428c9c21384] Deleting contents of the VM from datastore datastore1 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 818.493338] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-423c892e-6435-422f-ab41-7d39bb22efa3 tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] Deleting the datastore file [datastore1] 738ebb2f-7cc6-4d0f-871c-1428c9c21384 {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 818.493929] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b95d0055-bc90-4bfb-a261-b21335ae520f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.502637] env[62066]: DEBUG nova.compute.manager [None req-3d67ea1c-5f75-4620-b677-6f1e0faf1bd6 tempest-ServersV294TestFqdnHostnames-843300434 tempest-ServersV294TestFqdnHostnames-843300434-project-member] [instance: 2709d0ad-5dd8-4e3c-b1e6-3d1e2e242380] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 818.511378] env[62066]: DEBUG oslo_vmware.api [None req-423c892e-6435-422f-ab41-7d39bb22efa3 tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] Waiting for the task: (returnval){ [ 818.511378] env[62066]: value = "task-1340818" [ 818.511378] env[62066]: _type = "Task" [ 818.511378] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.523497] env[62066]: DEBUG oslo_vmware.api [None req-423c892e-6435-422f-ab41-7d39bb22efa3 tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] Task: {'id': task-1340818, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.557435] env[62066]: DEBUG oslo_vmware.api [None req-41432ebd-8922-4676-9cc7-6eed1922da68 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Task: {'id': task-1340814, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.613870] env[62066]: DEBUG oslo_vmware.api [None req-8d3cb2ee-be58-4e07-b64f-9d5c34ab9462 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': task-1340807, 'name': CloneVM_Task, 'duration_secs': 2.363305} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.614260] env[62066]: INFO nova.virt.vmwareapi.vmops [None req-8d3cb2ee-be58-4e07-b64f-9d5c34ab9462 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 49f46244-34fa-48a1-95a2-8e95850f345d] Created linked-clone VM from snapshot [ 818.618650] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef0212a2-e22f-4172-b371-c447faf0072e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.630053] env[62066]: DEBUG nova.virt.vmwareapi.images [None req-8d3cb2ee-be58-4e07-b64f-9d5c34ab9462 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 49f46244-34fa-48a1-95a2-8e95850f345d] Uploading image 843e9ee8-31ee-40e2-9f6f-50ffd87c6796 {{(pid=62066) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 818.656286] env[62066]: DEBUG oslo_vmware.rw_handles [None req-8d3cb2ee-be58-4e07-b64f-9d5c34ab9462 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 818.656286] env[62066]: value = "vm-286041" [ 818.656286] env[62066]: _type = "VirtualMachine" [ 818.656286] env[62066]: }. {{(pid=62066) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 818.656577] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-585ef20d-9db2-46c0-b727-24bff321aab1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.665239] env[62066]: DEBUG oslo_vmware.rw_handles [None req-8d3cb2ee-be58-4e07-b64f-9d5c34ab9462 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Lease: (returnval){ [ 818.665239] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]528e1e19-5129-182c-f74c-6edb2a3595ff" [ 818.665239] env[62066]: _type = "HttpNfcLease" [ 818.665239] env[62066]: } obtained for exporting VM: (result){ [ 818.665239] env[62066]: value = "vm-286041" [ 818.665239] env[62066]: _type = "VirtualMachine" [ 818.665239] env[62066]: }. {{(pid=62066) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 818.665515] env[62066]: DEBUG oslo_vmware.api [None req-8d3cb2ee-be58-4e07-b64f-9d5c34ab9462 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Waiting for the lease: (returnval){ [ 818.665515] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]528e1e19-5129-182c-f74c-6edb2a3595ff" [ 818.665515] env[62066]: _type = "HttpNfcLease" [ 818.665515] env[62066]: } to be ready. {{(pid=62066) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 818.673197] env[62066]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 818.673197] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]528e1e19-5129-182c-f74c-6edb2a3595ff" [ 818.673197] env[62066]: _type = "HttpNfcLease" [ 818.673197] env[62066]: } is initializing. {{(pid=62066) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 818.700187] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9e052131-3359-4888-9d78-89440b639dcc tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Acquiring lock "refresh_cache-8ff6fbb9-c90f-498d-9a85-d220a8c2f794" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 818.772543] env[62066]: DEBUG nova.network.neutron [req-45e414ba-4bc2-4636-918c-7198361230d4 req-200539b7-3e7c-45ae-88b4-115684d46ffe service nova] [instance: 8ff6fbb9-c90f-498d-9a85-d220a8c2f794] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 818.779885] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc1b2d4b-ff7b-4bd7-b66e-315cd9212456 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.789984] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45a94730-3312-45d7-81e7-a9cd240d4766 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.837040] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2be36934-3033-4fcd-8a6f-cd525ef38c9f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.840941] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba966e46-8935-42af-806d-f6f9a6460095 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.856737] env[62066]: DEBUG nova.compute.provider_tree [None req-b8836db2-293e-442c-80f2-42d8bc4c2918 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 818.874188] env[62066]: DEBUG oslo_vmware.api [None req-90f5cbab-7c0d-4b69-92aa-09bf24d9f7da tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] Task: {'id': task-1340816, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.883057] env[62066]: DEBUG nova.network.neutron [req-45e414ba-4bc2-4636-918c-7198361230d4 req-200539b7-3e7c-45ae-88b4-115684d46ffe service nova] [instance: 8ff6fbb9-c90f-498d-9a85-d220a8c2f794] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 818.920540] env[62066]: INFO nova.compute.manager [None req-f6cc809d-278e-4fb4-906b-f468fa41739d tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] [instance: 5f645eaa-6edc-4362-82ad-38c4c57b2be2] Took 28.03 seconds to build instance. [ 819.002887] env[62066]: INFO nova.compute.manager [None req-0188668d-25e1-420a-be60-fb5771349282 tempest-ServerAddressesTestJSON-1053293248 tempest-ServerAddressesTestJSON-1053293248-project-member] [instance: 49bc1350-0095-406b-bc68-005eb6b681a8] Took 26.31 seconds to build instance. [ 819.026197] env[62066]: DEBUG oslo_vmware.api [None req-423c892e-6435-422f-ab41-7d39bb22efa3 tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] Task: {'id': task-1340818, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.056510] env[62066]: DEBUG oslo_vmware.api [None req-41432ebd-8922-4676-9cc7-6eed1922da68 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Task: {'id': task-1340814, 'name': CreateSnapshot_Task, 'duration_secs': 1.349731} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.056627] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-41432ebd-8922-4676-9cc7-6eed1922da68 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] [instance: 1e1a2ede-4ad8-4600-851e-6f2046b3f919] Created Snapshot of the VM instance {{(pid=62066) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 819.057475] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e17c1fc5-44ed-4b00-9724-0b87bbf5a1f1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.115504] env[62066]: DEBUG nova.compute.manager [req-8f063caa-ffe9-4ab4-9670-314c92d8b44f req-2d4a098a-d9cf-485a-b105-6fdb4af0d670 service nova] [instance: 5f645eaa-6edc-4362-82ad-38c4c57b2be2] Received event network-changed-65fcb910-75be-4a68-8cec-476f7cc7cd00 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 819.115504] env[62066]: DEBUG nova.compute.manager [req-8f063caa-ffe9-4ab4-9670-314c92d8b44f req-2d4a098a-d9cf-485a-b105-6fdb4af0d670 service nova] [instance: 5f645eaa-6edc-4362-82ad-38c4c57b2be2] Refreshing instance network info cache due to event network-changed-65fcb910-75be-4a68-8cec-476f7cc7cd00. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 819.115727] env[62066]: DEBUG oslo_concurrency.lockutils [req-8f063caa-ffe9-4ab4-9670-314c92d8b44f req-2d4a098a-d9cf-485a-b105-6fdb4af0d670 service nova] Acquiring lock "refresh_cache-5f645eaa-6edc-4362-82ad-38c4c57b2be2" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 819.115877] env[62066]: DEBUG oslo_concurrency.lockutils [req-8f063caa-ffe9-4ab4-9670-314c92d8b44f req-2d4a098a-d9cf-485a-b105-6fdb4af0d670 service nova] Acquired lock "refresh_cache-5f645eaa-6edc-4362-82ad-38c4c57b2be2" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 819.116273] env[62066]: DEBUG nova.network.neutron [req-8f063caa-ffe9-4ab4-9670-314c92d8b44f req-2d4a098a-d9cf-485a-b105-6fdb4af0d670 service nova] [instance: 5f645eaa-6edc-4362-82ad-38c4c57b2be2] Refreshing network info cache for port 65fcb910-75be-4a68-8cec-476f7cc7cd00 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 819.174635] env[62066]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 819.174635] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]528e1e19-5129-182c-f74c-6edb2a3595ff" [ 819.174635] env[62066]: _type = "HttpNfcLease" [ 819.174635] env[62066]: } is ready. {{(pid=62066) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 819.174981] env[62066]: DEBUG oslo_vmware.rw_handles [None req-8d3cb2ee-be58-4e07-b64f-9d5c34ab9462 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 819.174981] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]528e1e19-5129-182c-f74c-6edb2a3595ff" [ 819.174981] env[62066]: _type = "HttpNfcLease" [ 819.174981] env[62066]: }. {{(pid=62066) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 819.175734] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0a2ea64-2c24-4640-bcc6-eb477afdf3f5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.183908] env[62066]: DEBUG oslo_vmware.rw_handles [None req-8d3cb2ee-be58-4e07-b64f-9d5c34ab9462 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5227be83-0a35-4646-6e67-879db120d0e5/disk-0.vmdk from lease info. {{(pid=62066) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 819.184104] env[62066]: DEBUG oslo_vmware.rw_handles [None req-8d3cb2ee-be58-4e07-b64f-9d5c34ab9462 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5227be83-0a35-4646-6e67-879db120d0e5/disk-0.vmdk for reading. {{(pid=62066) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 819.359745] env[62066]: DEBUG nova.scheduler.client.report [None req-b8836db2-293e-442c-80f2-42d8bc4c2918 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 819.373955] env[62066]: DEBUG oslo_vmware.api [None req-90f5cbab-7c0d-4b69-92aa-09bf24d9f7da tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] Task: {'id': task-1340816, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.385593] env[62066]: DEBUG oslo_concurrency.lockutils [req-45e414ba-4bc2-4636-918c-7198361230d4 req-200539b7-3e7c-45ae-88b4-115684d46ffe service nova] Releasing lock "refresh_cache-8ff6fbb9-c90f-498d-9a85-d220a8c2f794" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 819.385593] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9e052131-3359-4888-9d78-89440b639dcc tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Acquired lock "refresh_cache-8ff6fbb9-c90f-498d-9a85-d220a8c2f794" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 819.385726] env[62066]: DEBUG nova.network.neutron [None req-9e052131-3359-4888-9d78-89440b639dcc tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] [instance: 8ff6fbb9-c90f-498d-9a85-d220a8c2f794] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 819.389427] env[62066]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-eca59a6b-bad3-4804-ad99-8a90e8f4d846 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.426487] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f6cc809d-278e-4fb4-906b-f468fa41739d tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Lock "5f645eaa-6edc-4362-82ad-38c4c57b2be2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 111.130s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 819.505361] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0188668d-25e1-420a-be60-fb5771349282 tempest-ServerAddressesTestJSON-1053293248 tempest-ServerAddressesTestJSON-1053293248-project-member] Lock "49bc1350-0095-406b-bc68-005eb6b681a8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 99.711s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 819.521435] env[62066]: DEBUG nova.compute.manager [None req-3d67ea1c-5f75-4620-b677-6f1e0faf1bd6 tempest-ServersV294TestFqdnHostnames-843300434 tempest-ServersV294TestFqdnHostnames-843300434-project-member] [instance: 2709d0ad-5dd8-4e3c-b1e6-3d1e2e242380] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 819.529238] env[62066]: DEBUG oslo_vmware.api [None req-423c892e-6435-422f-ab41-7d39bb22efa3 tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] Task: {'id': task-1340818, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.558099] env[62066]: DEBUG nova.virt.hardware [None req-3d67ea1c-5f75-4620-b677-6f1e0faf1bd6 tempest-ServersV294TestFqdnHostnames-843300434 tempest-ServersV294TestFqdnHostnames-843300434-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-23T13:40:41Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-23T13:40:23Z,direct_url=,disk_format='vmdk',id=50ff584c-3b50-4395-af07-3e66769bc9f7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='eb52f7069a374c61ae946f052007c6d9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-23T13:40:24Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 819.558363] env[62066]: DEBUG nova.virt.hardware [None req-3d67ea1c-5f75-4620-b677-6f1e0faf1bd6 tempest-ServersV294TestFqdnHostnames-843300434 tempest-ServersV294TestFqdnHostnames-843300434-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 819.558525] env[62066]: DEBUG nova.virt.hardware [None req-3d67ea1c-5f75-4620-b677-6f1e0faf1bd6 tempest-ServersV294TestFqdnHostnames-843300434 tempest-ServersV294TestFqdnHostnames-843300434-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 819.558714] env[62066]: DEBUG nova.virt.hardware [None req-3d67ea1c-5f75-4620-b677-6f1e0faf1bd6 tempest-ServersV294TestFqdnHostnames-843300434 tempest-ServersV294TestFqdnHostnames-843300434-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 819.558867] env[62066]: DEBUG nova.virt.hardware [None req-3d67ea1c-5f75-4620-b677-6f1e0faf1bd6 tempest-ServersV294TestFqdnHostnames-843300434 tempest-ServersV294TestFqdnHostnames-843300434-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 819.559026] env[62066]: DEBUG nova.virt.hardware [None req-3d67ea1c-5f75-4620-b677-6f1e0faf1bd6 tempest-ServersV294TestFqdnHostnames-843300434 tempest-ServersV294TestFqdnHostnames-843300434-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 819.559246] env[62066]: DEBUG nova.virt.hardware [None req-3d67ea1c-5f75-4620-b677-6f1e0faf1bd6 tempest-ServersV294TestFqdnHostnames-843300434 tempest-ServersV294TestFqdnHostnames-843300434-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 819.563023] env[62066]: DEBUG nova.virt.hardware [None req-3d67ea1c-5f75-4620-b677-6f1e0faf1bd6 tempest-ServersV294TestFqdnHostnames-843300434 tempest-ServersV294TestFqdnHostnames-843300434-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 819.563023] env[62066]: DEBUG nova.virt.hardware [None req-3d67ea1c-5f75-4620-b677-6f1e0faf1bd6 tempest-ServersV294TestFqdnHostnames-843300434 tempest-ServersV294TestFqdnHostnames-843300434-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 819.563023] env[62066]: DEBUG nova.virt.hardware [None req-3d67ea1c-5f75-4620-b677-6f1e0faf1bd6 tempest-ServersV294TestFqdnHostnames-843300434 tempest-ServersV294TestFqdnHostnames-843300434-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 819.563023] env[62066]: DEBUG nova.virt.hardware [None req-3d67ea1c-5f75-4620-b677-6f1e0faf1bd6 tempest-ServersV294TestFqdnHostnames-843300434 tempest-ServersV294TestFqdnHostnames-843300434-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 819.563023] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42df3ee3-8e00-457a-aded-8bf57c613935 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.569281] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d69bf17-6df4-4e82-8837-09296f497c36 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.581609] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-41432ebd-8922-4676-9cc7-6eed1922da68 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] [instance: 1e1a2ede-4ad8-4600-851e-6f2046b3f919] Creating linked-clone VM from snapshot {{(pid=62066) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 819.582801] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-ada521e6-cb17-47a9-b035-1e3afe6c146d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.601129] env[62066]: DEBUG oslo_vmware.api [None req-41432ebd-8922-4676-9cc7-6eed1922da68 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Waiting for the task: (returnval){ [ 819.601129] env[62066]: value = "task-1340820" [ 819.601129] env[62066]: _type = "Task" [ 819.601129] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.610348] env[62066]: DEBUG oslo_vmware.api [None req-41432ebd-8922-4676-9cc7-6eed1922da68 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Task: {'id': task-1340820, 'name': CloneVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.765282] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3a2ae741-acf0-4694-aaae-0cdc986fd6f3 tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Acquiring lock "5f645eaa-6edc-4362-82ad-38c4c57b2be2" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 819.765664] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3a2ae741-acf0-4694-aaae-0cdc986fd6f3 tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Lock "5f645eaa-6edc-4362-82ad-38c4c57b2be2" acquired by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 819.766042] env[62066]: INFO nova.compute.manager [None req-3a2ae741-acf0-4694-aaae-0cdc986fd6f3 tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] [instance: 5f645eaa-6edc-4362-82ad-38c4c57b2be2] Rebooting instance [ 819.818635] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1e760fa3-f701-40c8-8666-21a2eeae0751 tempest-ServerAddressesTestJSON-1053293248 tempest-ServerAddressesTestJSON-1053293248-project-member] Acquiring lock "49bc1350-0095-406b-bc68-005eb6b681a8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 819.818996] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1e760fa3-f701-40c8-8666-21a2eeae0751 tempest-ServerAddressesTestJSON-1053293248 tempest-ServerAddressesTestJSON-1053293248-project-member] Lock "49bc1350-0095-406b-bc68-005eb6b681a8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 819.819530] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1e760fa3-f701-40c8-8666-21a2eeae0751 tempest-ServerAddressesTestJSON-1053293248 tempest-ServerAddressesTestJSON-1053293248-project-member] Acquiring lock "49bc1350-0095-406b-bc68-005eb6b681a8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 819.820426] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1e760fa3-f701-40c8-8666-21a2eeae0751 tempest-ServerAddressesTestJSON-1053293248 tempest-ServerAddressesTestJSON-1053293248-project-member] Lock "49bc1350-0095-406b-bc68-005eb6b681a8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 819.820656] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1e760fa3-f701-40c8-8666-21a2eeae0751 tempest-ServerAddressesTestJSON-1053293248 tempest-ServerAddressesTestJSON-1053293248-project-member] Lock "49bc1350-0095-406b-bc68-005eb6b681a8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 819.824457] env[62066]: INFO nova.compute.manager [None req-1e760fa3-f701-40c8-8666-21a2eeae0751 tempest-ServerAddressesTestJSON-1053293248 tempest-ServerAddressesTestJSON-1053293248-project-member] [instance: 49bc1350-0095-406b-bc68-005eb6b681a8] Terminating instance [ 819.830562] env[62066]: DEBUG nova.compute.manager [None req-1e760fa3-f701-40c8-8666-21a2eeae0751 tempest-ServerAddressesTestJSON-1053293248 tempest-ServerAddressesTestJSON-1053293248-project-member] [instance: 49bc1350-0095-406b-bc68-005eb6b681a8] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 819.831147] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-1e760fa3-f701-40c8-8666-21a2eeae0751 tempest-ServerAddressesTestJSON-1053293248 tempest-ServerAddressesTestJSON-1053293248-project-member] [instance: 49bc1350-0095-406b-bc68-005eb6b681a8] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 819.832299] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fa54ca8-cb91-4441-b586-11b412e49473 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.854047] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-1e760fa3-f701-40c8-8666-21a2eeae0751 tempest-ServerAddressesTestJSON-1053293248 tempest-ServerAddressesTestJSON-1053293248-project-member] [instance: 49bc1350-0095-406b-bc68-005eb6b681a8] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 819.856245] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e94949ec-dd98-4843-a0cd-9cbeba3e2a30 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.866850] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b8836db2-293e-442c-80f2-42d8bc4c2918 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.380s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 819.867490] env[62066]: DEBUG nova.compute.manager [None req-b8836db2-293e-442c-80f2-42d8bc4c2918 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 88daaaa6-f385-4161-bc74-9fffabb1145d] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 819.871628] env[62066]: DEBUG oslo_vmware.api [None req-1e760fa3-f701-40c8-8666-21a2eeae0751 tempest-ServerAddressesTestJSON-1053293248 tempest-ServerAddressesTestJSON-1053293248-project-member] Waiting for the task: (returnval){ [ 819.871628] env[62066]: value = "task-1340821" [ 819.871628] env[62066]: _type = "Task" [ 819.871628] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.880176] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d2edd17d-ac3d-4b81-90f1-278c62ed416c tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.975s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 819.881654] env[62066]: INFO nova.compute.claims [None req-d2edd17d-ac3d-4b81-90f1-278c62ed416c tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] [instance: de2e2be7-efdb-45a8-842a-640ab9deb1d9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 819.901629] env[62066]: DEBUG oslo_vmware.api [None req-90f5cbab-7c0d-4b69-92aa-09bf24d9f7da tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] Task: {'id': task-1340816, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.906611] env[62066]: DEBUG oslo_vmware.api [None req-1e760fa3-f701-40c8-8666-21a2eeae0751 tempest-ServerAddressesTestJSON-1053293248 tempest-ServerAddressesTestJSON-1053293248-project-member] Task: {'id': task-1340821, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.954742] env[62066]: DEBUG nova.network.neutron [None req-9e052131-3359-4888-9d78-89440b639dcc tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] [instance: 8ff6fbb9-c90f-498d-9a85-d220a8c2f794] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 820.031975] env[62066]: DEBUG oslo_vmware.api [None req-423c892e-6435-422f-ab41-7d39bb22efa3 tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] Task: {'id': task-1340818, 'name': DeleteDatastoreFile_Task, 'duration_secs': 1.301828} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.031975] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-423c892e-6435-422f-ab41-7d39bb22efa3 tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 820.031975] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-423c892e-6435-422f-ab41-7d39bb22efa3 tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] [instance: 738ebb2f-7cc6-4d0f-871c-1428c9c21384] Deleted contents of the VM from datastore datastore1 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 820.031975] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-423c892e-6435-422f-ab41-7d39bb22efa3 tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] [instance: 738ebb2f-7cc6-4d0f-871c-1428c9c21384] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 820.031975] env[62066]: INFO nova.compute.manager [None req-423c892e-6435-422f-ab41-7d39bb22efa3 tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] [instance: 738ebb2f-7cc6-4d0f-871c-1428c9c21384] Took 2.23 seconds to destroy the instance on the hypervisor. [ 820.031975] env[62066]: DEBUG oslo.service.loopingcall [None req-423c892e-6435-422f-ab41-7d39bb22efa3 tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 820.031975] env[62066]: DEBUG nova.compute.manager [-] [instance: 738ebb2f-7cc6-4d0f-871c-1428c9c21384] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 820.031975] env[62066]: DEBUG nova.network.neutron [-] [instance: 738ebb2f-7cc6-4d0f-871c-1428c9c21384] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 820.057315] env[62066]: DEBUG nova.network.neutron [req-8f063caa-ffe9-4ab4-9670-314c92d8b44f req-2d4a098a-d9cf-485a-b105-6fdb4af0d670 service nova] [instance: 5f645eaa-6edc-4362-82ad-38c4c57b2be2] Updated VIF entry in instance network info cache for port 65fcb910-75be-4a68-8cec-476f7cc7cd00. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 820.057717] env[62066]: DEBUG nova.network.neutron [req-8f063caa-ffe9-4ab4-9670-314c92d8b44f req-2d4a098a-d9cf-485a-b105-6fdb4af0d670 service nova] [instance: 5f645eaa-6edc-4362-82ad-38c4c57b2be2] Updating instance_info_cache with network_info: [{"id": "65fcb910-75be-4a68-8cec-476f7cc7cd00", "address": "fa:16:3e:20:29:72", "network": {"id": "b0f4607a-9d59-457e-aabc-5851c90b3366", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-1154549021-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a9b239f4cb3242d0bd1dcd98a2a3bfc0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c02dd284-ab80-451c-93eb-48c8360acb9c", "external-id": "nsx-vlan-transportzone-818", "segmentation_id": 818, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap65fcb910-75", "ovs_interfaceid": "65fcb910-75be-4a68-8cec-476f7cc7cd00", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 820.113623] env[62066]: DEBUG oslo_vmware.api [None req-41432ebd-8922-4676-9cc7-6eed1922da68 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Task: {'id': task-1340820, 'name': CloneVM_Task} progress is 94%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.300483] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3a2ae741-acf0-4694-aaae-0cdc986fd6f3 tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Acquiring lock "refresh_cache-5f645eaa-6edc-4362-82ad-38c4c57b2be2" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 820.383023] env[62066]: DEBUG nova.compute.utils [None req-b8836db2-293e-442c-80f2-42d8bc4c2918 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 820.385595] env[62066]: DEBUG nova.compute.manager [None req-b8836db2-293e-442c-80f2-42d8bc4c2918 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 88daaaa6-f385-4161-bc74-9fffabb1145d] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 820.386142] env[62066]: DEBUG nova.network.neutron [None req-b8836db2-293e-442c-80f2-42d8bc4c2918 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 88daaaa6-f385-4161-bc74-9fffabb1145d] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 820.390081] env[62066]: DEBUG oslo_vmware.api [None req-90f5cbab-7c0d-4b69-92aa-09bf24d9f7da tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] Task: {'id': task-1340816, 'name': ReconfigVM_Task, 'duration_secs': 2.166338} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.399623] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-90f5cbab-7c0d-4b69-92aa-09bf24d9f7da tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] [instance: 050040cb-6ee3-4ad2-960a-fcebb53ac394] Reconfigured VM instance instance-00000043 to attach disk [datastore1] 050040cb-6ee3-4ad2-960a-fcebb53ac394/050040cb-6ee3-4ad2-960a-fcebb53ac394.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 820.404545] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-113c036e-f6f7-4420-bffd-742cb8c1bc43 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.407667] env[62066]: DEBUG nova.compute.manager [req-449b997e-5a81-423c-9865-623ce79fe420 req-4dcbb844-c902-451c-a858-b0c8b436e0f5 service nova] [instance: 2709d0ad-5dd8-4e3c-b1e6-3d1e2e242380] Received event network-vif-plugged-c5cdc663-1e55-4238-bad4-7fb406b34a36 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 820.407924] env[62066]: DEBUG oslo_concurrency.lockutils [req-449b997e-5a81-423c-9865-623ce79fe420 req-4dcbb844-c902-451c-a858-b0c8b436e0f5 service nova] Acquiring lock "2709d0ad-5dd8-4e3c-b1e6-3d1e2e242380-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 820.408287] env[62066]: DEBUG oslo_concurrency.lockutils [req-449b997e-5a81-423c-9865-623ce79fe420 req-4dcbb844-c902-451c-a858-b0c8b436e0f5 service nova] Lock "2709d0ad-5dd8-4e3c-b1e6-3d1e2e242380-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 820.408595] env[62066]: DEBUG oslo_concurrency.lockutils [req-449b997e-5a81-423c-9865-623ce79fe420 req-4dcbb844-c902-451c-a858-b0c8b436e0f5 service nova] Lock "2709d0ad-5dd8-4e3c-b1e6-3d1e2e242380-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 820.408849] env[62066]: DEBUG nova.compute.manager [req-449b997e-5a81-423c-9865-623ce79fe420 req-4dcbb844-c902-451c-a858-b0c8b436e0f5 service nova] [instance: 2709d0ad-5dd8-4e3c-b1e6-3d1e2e242380] No waiting events found dispatching network-vif-plugged-c5cdc663-1e55-4238-bad4-7fb406b34a36 {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 820.409053] env[62066]: WARNING nova.compute.manager [req-449b997e-5a81-423c-9865-623ce79fe420 req-4dcbb844-c902-451c-a858-b0c8b436e0f5 service nova] [instance: 2709d0ad-5dd8-4e3c-b1e6-3d1e2e242380] Received unexpected event network-vif-plugged-c5cdc663-1e55-4238-bad4-7fb406b34a36 for instance with vm_state building and task_state spawning. [ 820.417484] env[62066]: DEBUG oslo_vmware.api [None req-1e760fa3-f701-40c8-8666-21a2eeae0751 tempest-ServerAddressesTestJSON-1053293248 tempest-ServerAddressesTestJSON-1053293248-project-member] Task: {'id': task-1340821, 'name': PowerOffVM_Task, 'duration_secs': 0.216576} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.420793] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-1e760fa3-f701-40c8-8666-21a2eeae0751 tempest-ServerAddressesTestJSON-1053293248 tempest-ServerAddressesTestJSON-1053293248-project-member] [instance: 49bc1350-0095-406b-bc68-005eb6b681a8] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 820.420793] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-1e760fa3-f701-40c8-8666-21a2eeae0751 tempest-ServerAddressesTestJSON-1053293248 tempest-ServerAddressesTestJSON-1053293248-project-member] [instance: 49bc1350-0095-406b-bc68-005eb6b681a8] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 820.421264] env[62066]: DEBUG oslo_vmware.api [None req-90f5cbab-7c0d-4b69-92aa-09bf24d9f7da tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] Waiting for the task: (returnval){ [ 820.421264] env[62066]: value = "task-1340822" [ 820.421264] env[62066]: _type = "Task" [ 820.421264] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.421493] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c1b046ea-f84d-4ced-992c-28d436888a82 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.436167] env[62066]: DEBUG oslo_vmware.api [None req-90f5cbab-7c0d-4b69-92aa-09bf24d9f7da tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] Task: {'id': task-1340822, 'name': Rename_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.465219] env[62066]: DEBUG nova.network.neutron [None req-9e052131-3359-4888-9d78-89440b639dcc tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] [instance: 8ff6fbb9-c90f-498d-9a85-d220a8c2f794] Updating instance_info_cache with network_info: [{"id": "8c324833-7bf9-4007-85aa-5b16871f63c7", "address": "fa:16:3e:00:85:34", "network": {"id": "0e17b5e2-526e-4d4c-a463-5e34408044cf", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1588316661-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1f05d8a66d0c4a479e49a947e4b4bbff", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ffcecdaa-a7b8-49fc-9371-dbdb7744688e", "external-id": "nsx-vlan-transportzone-994", "segmentation_id": 994, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8c324833-7b", "ovs_interfaceid": "8c324833-7bf9-4007-85aa-5b16871f63c7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 820.484173] env[62066]: DEBUG nova.network.neutron [None req-3d67ea1c-5f75-4620-b677-6f1e0faf1bd6 tempest-ServersV294TestFqdnHostnames-843300434 tempest-ServersV294TestFqdnHostnames-843300434-project-member] [instance: 2709d0ad-5dd8-4e3c-b1e6-3d1e2e242380] Successfully updated port: c5cdc663-1e55-4238-bad4-7fb406b34a36 {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 820.500572] env[62066]: DEBUG nova.policy [None req-b8836db2-293e-442c-80f2-42d8bc4c2918 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2ab20cbd0bab4ae7ba46d9135605a509', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd609babc987a42e2a8ddb4bfb9c3b103', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 820.526029] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-1e760fa3-f701-40c8-8666-21a2eeae0751 tempest-ServerAddressesTestJSON-1053293248 tempest-ServerAddressesTestJSON-1053293248-project-member] [instance: 49bc1350-0095-406b-bc68-005eb6b681a8] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 820.526029] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-1e760fa3-f701-40c8-8666-21a2eeae0751 tempest-ServerAddressesTestJSON-1053293248 tempest-ServerAddressesTestJSON-1053293248-project-member] [instance: 49bc1350-0095-406b-bc68-005eb6b681a8] Deleting contents of the VM from datastore datastore1 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 820.526029] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-1e760fa3-f701-40c8-8666-21a2eeae0751 tempest-ServerAddressesTestJSON-1053293248 tempest-ServerAddressesTestJSON-1053293248-project-member] Deleting the datastore file [datastore1] 49bc1350-0095-406b-bc68-005eb6b681a8 {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 820.526029] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-533e5a6a-8a12-4c00-9456-8d37a4be739f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.532392] env[62066]: DEBUG oslo_vmware.api [None req-1e760fa3-f701-40c8-8666-21a2eeae0751 tempest-ServerAddressesTestJSON-1053293248 tempest-ServerAddressesTestJSON-1053293248-project-member] Waiting for the task: (returnval){ [ 820.532392] env[62066]: value = "task-1340824" [ 820.532392] env[62066]: _type = "Task" [ 820.532392] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.543018] env[62066]: DEBUG oslo_vmware.api [None req-1e760fa3-f701-40c8-8666-21a2eeae0751 tempest-ServerAddressesTestJSON-1053293248 tempest-ServerAddressesTestJSON-1053293248-project-member] Task: {'id': task-1340824, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.564847] env[62066]: DEBUG oslo_concurrency.lockutils [req-8f063caa-ffe9-4ab4-9670-314c92d8b44f req-2d4a098a-d9cf-485a-b105-6fdb4af0d670 service nova] Releasing lock "refresh_cache-5f645eaa-6edc-4362-82ad-38c4c57b2be2" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 820.564847] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3a2ae741-acf0-4694-aaae-0cdc986fd6f3 tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Acquired lock "refresh_cache-5f645eaa-6edc-4362-82ad-38c4c57b2be2" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 820.564847] env[62066]: DEBUG nova.network.neutron [None req-3a2ae741-acf0-4694-aaae-0cdc986fd6f3 tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] [instance: 5f645eaa-6edc-4362-82ad-38c4c57b2be2] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 820.614479] env[62066]: DEBUG oslo_vmware.api [None req-41432ebd-8922-4676-9cc7-6eed1922da68 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Task: {'id': task-1340820, 'name': CloneVM_Task} progress is 94%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.801586] env[62066]: DEBUG nova.network.neutron [None req-b8836db2-293e-442c-80f2-42d8bc4c2918 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 88daaaa6-f385-4161-bc74-9fffabb1145d] Successfully created port: ba002f90-5218-46f2-b7e1-cd96ae3018c7 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 820.891571] env[62066]: DEBUG nova.compute.manager [None req-b8836db2-293e-442c-80f2-42d8bc4c2918 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 88daaaa6-f385-4161-bc74-9fffabb1145d] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 820.910833] env[62066]: DEBUG nova.network.neutron [-] [instance: 738ebb2f-7cc6-4d0f-871c-1428c9c21384] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 820.942193] env[62066]: DEBUG oslo_vmware.api [None req-90f5cbab-7c0d-4b69-92aa-09bf24d9f7da tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] Task: {'id': task-1340822, 'name': Rename_Task, 'duration_secs': 0.179572} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.942193] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-90f5cbab-7c0d-4b69-92aa-09bf24d9f7da tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] [instance: 050040cb-6ee3-4ad2-960a-fcebb53ac394] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 820.942193] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6a698f93-292f-41b0-9cbf-28e9fc5a889f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.951983] env[62066]: DEBUG oslo_vmware.api [None req-90f5cbab-7c0d-4b69-92aa-09bf24d9f7da tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] Waiting for the task: (returnval){ [ 820.951983] env[62066]: value = "task-1340825" [ 820.951983] env[62066]: _type = "Task" [ 820.951983] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.965695] env[62066]: DEBUG oslo_vmware.api [None req-90f5cbab-7c0d-4b69-92aa-09bf24d9f7da tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] Task: {'id': task-1340825, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.972029] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9e052131-3359-4888-9d78-89440b639dcc tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Releasing lock "refresh_cache-8ff6fbb9-c90f-498d-9a85-d220a8c2f794" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 820.972029] env[62066]: DEBUG nova.compute.manager [None req-9e052131-3359-4888-9d78-89440b639dcc tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] [instance: 8ff6fbb9-c90f-498d-9a85-d220a8c2f794] Instance network_info: |[{"id": "8c324833-7bf9-4007-85aa-5b16871f63c7", "address": "fa:16:3e:00:85:34", "network": {"id": "0e17b5e2-526e-4d4c-a463-5e34408044cf", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1588316661-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1f05d8a66d0c4a479e49a947e4b4bbff", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ffcecdaa-a7b8-49fc-9371-dbdb7744688e", "external-id": "nsx-vlan-transportzone-994", "segmentation_id": 994, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8c324833-7b", "ovs_interfaceid": "8c324833-7bf9-4007-85aa-5b16871f63c7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 820.972961] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-9e052131-3359-4888-9d78-89440b639dcc tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] [instance: 8ff6fbb9-c90f-498d-9a85-d220a8c2f794] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:00:85:34', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ffcecdaa-a7b8-49fc-9371-dbdb7744688e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8c324833-7bf9-4007-85aa-5b16871f63c7', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 820.985165] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-9e052131-3359-4888-9d78-89440b639dcc tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Creating folder: Project (1f05d8a66d0c4a479e49a947e4b4bbff). Parent ref: group-v285980. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 820.985165] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e9f13ae3-3d7c-4444-bac4-e5c859fb2da5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.987284] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3d67ea1c-5f75-4620-b677-6f1e0faf1bd6 tempest-ServersV294TestFqdnHostnames-843300434 tempest-ServersV294TestFqdnHostnames-843300434-project-member] Acquiring lock "refresh_cache-2709d0ad-5dd8-4e3c-b1e6-3d1e2e242380" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 820.987284] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3d67ea1c-5f75-4620-b677-6f1e0faf1bd6 tempest-ServersV294TestFqdnHostnames-843300434 tempest-ServersV294TestFqdnHostnames-843300434-project-member] Acquired lock "refresh_cache-2709d0ad-5dd8-4e3c-b1e6-3d1e2e242380" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 820.987284] env[62066]: DEBUG nova.network.neutron [None req-3d67ea1c-5f75-4620-b677-6f1e0faf1bd6 tempest-ServersV294TestFqdnHostnames-843300434 tempest-ServersV294TestFqdnHostnames-843300434-project-member] [instance: 2709d0ad-5dd8-4e3c-b1e6-3d1e2e242380] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 820.996696] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-9e052131-3359-4888-9d78-89440b639dcc tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Created folder: Project (1f05d8a66d0c4a479e49a947e4b4bbff) in parent group-v285980. [ 820.999479] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-9e052131-3359-4888-9d78-89440b639dcc tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Creating folder: Instances. Parent ref: group-v286044. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 820.999479] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e33905cf-1d7d-4836-a64e-f414d8a4142c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.012559] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-9e052131-3359-4888-9d78-89440b639dcc tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Created folder: Instances in parent group-v286044. [ 821.012874] env[62066]: DEBUG oslo.service.loopingcall [None req-9e052131-3359-4888-9d78-89440b639dcc tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 821.013282] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8ff6fbb9-c90f-498d-9a85-d220a8c2f794] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 821.013838] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-400da5dd-1b68-445e-a691-c38284d825bc {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.045775] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 821.045775] env[62066]: value = "task-1340828" [ 821.045775] env[62066]: _type = "Task" [ 821.045775] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.051772] env[62066]: DEBUG oslo_vmware.api [None req-1e760fa3-f701-40c8-8666-21a2eeae0751 tempest-ServerAddressesTestJSON-1053293248 tempest-ServerAddressesTestJSON-1053293248-project-member] Task: {'id': task-1340824, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.226671} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.052322] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-1e760fa3-f701-40c8-8666-21a2eeae0751 tempest-ServerAddressesTestJSON-1053293248 tempest-ServerAddressesTestJSON-1053293248-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 821.052539] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-1e760fa3-f701-40c8-8666-21a2eeae0751 tempest-ServerAddressesTestJSON-1053293248 tempest-ServerAddressesTestJSON-1053293248-project-member] [instance: 49bc1350-0095-406b-bc68-005eb6b681a8] Deleted contents of the VM from datastore datastore1 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 821.052726] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-1e760fa3-f701-40c8-8666-21a2eeae0751 tempest-ServerAddressesTestJSON-1053293248 tempest-ServerAddressesTestJSON-1053293248-project-member] [instance: 49bc1350-0095-406b-bc68-005eb6b681a8] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 821.053426] env[62066]: INFO nova.compute.manager [None req-1e760fa3-f701-40c8-8666-21a2eeae0751 tempest-ServerAddressesTestJSON-1053293248 tempest-ServerAddressesTestJSON-1053293248-project-member] [instance: 49bc1350-0095-406b-bc68-005eb6b681a8] Took 1.22 seconds to destroy the instance on the hypervisor. [ 821.053426] env[62066]: DEBUG oslo.service.loopingcall [None req-1e760fa3-f701-40c8-8666-21a2eeae0751 tempest-ServerAddressesTestJSON-1053293248 tempest-ServerAddressesTestJSON-1053293248-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 821.053745] env[62066]: DEBUG nova.compute.manager [-] [instance: 49bc1350-0095-406b-bc68-005eb6b681a8] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 821.054585] env[62066]: DEBUG nova.network.neutron [-] [instance: 49bc1350-0095-406b-bc68-005eb6b681a8] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 821.061553] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1340828, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.114776] env[62066]: DEBUG oslo_vmware.api [None req-41432ebd-8922-4676-9cc7-6eed1922da68 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Task: {'id': task-1340820, 'name': CloneVM_Task, 'duration_secs': 1.382158} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.115161] env[62066]: INFO nova.virt.vmwareapi.vmops [None req-41432ebd-8922-4676-9cc7-6eed1922da68 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] [instance: 1e1a2ede-4ad8-4600-851e-6f2046b3f919] Created linked-clone VM from snapshot [ 821.116187] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b6e43f8-7e86-4358-9c39-437d15558058 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.131325] env[62066]: DEBUG nova.virt.vmwareapi.images [None req-41432ebd-8922-4676-9cc7-6eed1922da68 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] [instance: 1e1a2ede-4ad8-4600-851e-6f2046b3f919] Uploading image b6a3f5de-6b26-4b8a-96cd-dc69c51db39d {{(pid=62066) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 821.157539] env[62066]: DEBUG nova.compute.manager [req-54fa84f4-619a-49b2-a61d-8f33265b27b4 req-cb2f8ee3-aa84-4875-a443-dd197a7f3404 service nova] [instance: 738ebb2f-7cc6-4d0f-871c-1428c9c21384] Received event network-vif-deleted-fa65affa-3b23-4d64-a714-83d41caa3293 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 821.165589] env[62066]: DEBUG oslo_vmware.rw_handles [None req-41432ebd-8922-4676-9cc7-6eed1922da68 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 821.165589] env[62066]: value = "vm-286043" [ 821.165589] env[62066]: _type = "VirtualMachine" [ 821.165589] env[62066]: }. {{(pid=62066) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 821.166193] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-e0f41b51-9f9a-490e-834a-9b538411ab22 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.177578] env[62066]: DEBUG oslo_vmware.rw_handles [None req-41432ebd-8922-4676-9cc7-6eed1922da68 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Lease: (returnval){ [ 821.177578] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]5218d213-e794-e7be-ce3d-b834fa9d8ecd" [ 821.177578] env[62066]: _type = "HttpNfcLease" [ 821.177578] env[62066]: } obtained for exporting VM: (result){ [ 821.177578] env[62066]: value = "vm-286043" [ 821.177578] env[62066]: _type = "VirtualMachine" [ 821.177578] env[62066]: }. {{(pid=62066) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 821.177578] env[62066]: DEBUG oslo_vmware.api [None req-41432ebd-8922-4676-9cc7-6eed1922da68 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Waiting for the lease: (returnval){ [ 821.177578] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]5218d213-e794-e7be-ce3d-b834fa9d8ecd" [ 821.177578] env[62066]: _type = "HttpNfcLease" [ 821.177578] env[62066]: } to be ready. {{(pid=62066) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 821.187124] env[62066]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 821.187124] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]5218d213-e794-e7be-ce3d-b834fa9d8ecd" [ 821.187124] env[62066]: _type = "HttpNfcLease" [ 821.187124] env[62066]: } is initializing. {{(pid=62066) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 821.275557] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b91f7af-2a0c-4ccc-8bf2-f547b5d87a31 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.284639] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa4e3c38-630d-4d7e-879a-c4a0ea809821 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.321118] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d471481-d940-4308-92cd-f435139b728d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.329241] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73473f7c-1ba3-405d-92be-266c8e32f9d7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.344565] env[62066]: DEBUG nova.compute.provider_tree [None req-d2edd17d-ac3d-4b81-90f1-278c62ed416c tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 821.416025] env[62066]: INFO nova.compute.manager [-] [instance: 738ebb2f-7cc6-4d0f-871c-1428c9c21384] Took 1.38 seconds to deallocate network for instance. [ 821.465640] env[62066]: DEBUG oslo_vmware.api [None req-90f5cbab-7c0d-4b69-92aa-09bf24d9f7da tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] Task: {'id': task-1340825, 'name': PowerOnVM_Task} progress is 1%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.489167] env[62066]: DEBUG nova.network.neutron [None req-3a2ae741-acf0-4694-aaae-0cdc986fd6f3 tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] [instance: 5f645eaa-6edc-4362-82ad-38c4c57b2be2] Updating instance_info_cache with network_info: [{"id": "65fcb910-75be-4a68-8cec-476f7cc7cd00", "address": "fa:16:3e:20:29:72", "network": {"id": "b0f4607a-9d59-457e-aabc-5851c90b3366", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-1154549021-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a9b239f4cb3242d0bd1dcd98a2a3bfc0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c02dd284-ab80-451c-93eb-48c8360acb9c", "external-id": "nsx-vlan-transportzone-818", "segmentation_id": 818, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap65fcb910-75", "ovs_interfaceid": "65fcb910-75be-4a68-8cec-476f7cc7cd00", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 821.532391] env[62066]: DEBUG nova.network.neutron [None req-3d67ea1c-5f75-4620-b677-6f1e0faf1bd6 tempest-ServersV294TestFqdnHostnames-843300434 tempest-ServersV294TestFqdnHostnames-843300434-project-member] [instance: 2709d0ad-5dd8-4e3c-b1e6-3d1e2e242380] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 821.554211] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1340828, 'name': CreateVM_Task, 'duration_secs': 0.406594} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.554395] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8ff6fbb9-c90f-498d-9a85-d220a8c2f794] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 821.555098] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9e052131-3359-4888-9d78-89440b639dcc tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 821.555486] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9e052131-3359-4888-9d78-89440b639dcc tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Acquired lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 821.555746] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9e052131-3359-4888-9d78-89440b639dcc tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 821.556063] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-099e7c3c-433c-409d-96be-35ea983e5699 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.560959] env[62066]: DEBUG oslo_vmware.api [None req-9e052131-3359-4888-9d78-89440b639dcc tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Waiting for the task: (returnval){ [ 821.560959] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]52325530-1268-8f30-9d39-5f637aad0785" [ 821.560959] env[62066]: _type = "Task" [ 821.560959] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.572386] env[62066]: DEBUG oslo_vmware.api [None req-9e052131-3359-4888-9d78-89440b639dcc tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52325530-1268-8f30-9d39-5f637aad0785, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.687031] env[62066]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 821.687031] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]5218d213-e794-e7be-ce3d-b834fa9d8ecd" [ 821.687031] env[62066]: _type = "HttpNfcLease" [ 821.687031] env[62066]: } is ready. {{(pid=62066) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 821.687271] env[62066]: DEBUG nova.network.neutron [None req-3d67ea1c-5f75-4620-b677-6f1e0faf1bd6 tempest-ServersV294TestFqdnHostnames-843300434 tempest-ServersV294TestFqdnHostnames-843300434-project-member] [instance: 2709d0ad-5dd8-4e3c-b1e6-3d1e2e242380] Updating instance_info_cache with network_info: [{"id": "c5cdc663-1e55-4238-bad4-7fb406b34a36", "address": "fa:16:3e:42:1a:81", "network": {"id": "2df566a5-1f6f-4665-a06d-1441530a4f80", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-1008410317-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d1785af300d545e6913c0e91148f4563", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b4d548e7-d762-406a-bb2d-dc7168a8ca67", "external-id": "nsx-vlan-transportzone-796", "segmentation_id": 796, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc5cdc663-1e", "ovs_interfaceid": "c5cdc663-1e55-4238-bad4-7fb406b34a36", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 821.688498] env[62066]: DEBUG oslo_vmware.rw_handles [None req-41432ebd-8922-4676-9cc7-6eed1922da68 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 821.688498] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]5218d213-e794-e7be-ce3d-b834fa9d8ecd" [ 821.688498] env[62066]: _type = "HttpNfcLease" [ 821.688498] env[62066]: }. {{(pid=62066) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 821.689308] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4123e40b-2737-4993-9749-380f102628bf {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.698074] env[62066]: DEBUG oslo_vmware.rw_handles [None req-41432ebd-8922-4676-9cc7-6eed1922da68 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/520e24d9-6862-b3ad-ba9f-81ea21948905/disk-0.vmdk from lease info. {{(pid=62066) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 821.698074] env[62066]: DEBUG oslo_vmware.rw_handles [None req-41432ebd-8922-4676-9cc7-6eed1922da68 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/520e24d9-6862-b3ad-ba9f-81ea21948905/disk-0.vmdk for reading. {{(pid=62066) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 821.807363] env[62066]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-8fdd82ab-86e0-4fe0-916f-3dc2de7a7afe {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.848920] env[62066]: DEBUG nova.scheduler.client.report [None req-d2edd17d-ac3d-4b81-90f1-278c62ed416c tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 821.908782] env[62066]: DEBUG nova.compute.manager [None req-b8836db2-293e-442c-80f2-42d8bc4c2918 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 88daaaa6-f385-4161-bc74-9fffabb1145d] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 821.927355] env[62066]: DEBUG oslo_concurrency.lockutils [None req-423c892e-6435-422f-ab41-7d39bb22efa3 tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 821.964465] env[62066]: DEBUG oslo_vmware.api [None req-90f5cbab-7c0d-4b69-92aa-09bf24d9f7da tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] Task: {'id': task-1340825, 'name': PowerOnVM_Task} progress is 64%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.993842] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3a2ae741-acf0-4694-aaae-0cdc986fd6f3 tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Releasing lock "refresh_cache-5f645eaa-6edc-4362-82ad-38c4c57b2be2" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 821.996675] env[62066]: DEBUG nova.compute.manager [None req-3a2ae741-acf0-4694-aaae-0cdc986fd6f3 tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] [instance: 5f645eaa-6edc-4362-82ad-38c4c57b2be2] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 821.997598] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df1a0f1a-4d40-4d21-bc32-d05c68bfa4ae {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.047892] env[62066]: DEBUG nova.virt.hardware [None req-b8836db2-293e-442c-80f2-42d8bc4c2918 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-23T13:40:41Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-23T13:40:23Z,direct_url=,disk_format='vmdk',id=50ff584c-3b50-4395-af07-3e66769bc9f7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='eb52f7069a374c61ae946f052007c6d9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-23T13:40:24Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 822.048545] env[62066]: DEBUG nova.virt.hardware [None req-b8836db2-293e-442c-80f2-42d8bc4c2918 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 822.048545] env[62066]: DEBUG nova.virt.hardware [None req-b8836db2-293e-442c-80f2-42d8bc4c2918 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 822.048545] env[62066]: DEBUG nova.virt.hardware [None req-b8836db2-293e-442c-80f2-42d8bc4c2918 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 822.049930] env[62066]: DEBUG nova.virt.hardware [None req-b8836db2-293e-442c-80f2-42d8bc4c2918 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 822.049930] env[62066]: DEBUG nova.virt.hardware [None req-b8836db2-293e-442c-80f2-42d8bc4c2918 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 822.049930] env[62066]: DEBUG nova.virt.hardware [None req-b8836db2-293e-442c-80f2-42d8bc4c2918 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 822.050544] env[62066]: DEBUG nova.virt.hardware [None req-b8836db2-293e-442c-80f2-42d8bc4c2918 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 822.050544] env[62066]: DEBUG nova.virt.hardware [None req-b8836db2-293e-442c-80f2-42d8bc4c2918 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 822.050621] env[62066]: DEBUG nova.virt.hardware [None req-b8836db2-293e-442c-80f2-42d8bc4c2918 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 822.051048] env[62066]: DEBUG nova.virt.hardware [None req-b8836db2-293e-442c-80f2-42d8bc4c2918 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 822.052625] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f86d6298-6020-430e-ac41-c6c6eaf6e89c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.073453] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7a35f34-5574-4c7f-9f64-e0d4aa92ea73 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.083913] env[62066]: DEBUG nova.network.neutron [-] [instance: 49bc1350-0095-406b-bc68-005eb6b681a8] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 822.086490] env[62066]: DEBUG oslo_vmware.api [None req-9e052131-3359-4888-9d78-89440b639dcc tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52325530-1268-8f30-9d39-5f637aad0785, 'name': SearchDatastore_Task, 'duration_secs': 0.010868} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.086490] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9e052131-3359-4888-9d78-89440b639dcc tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Releasing lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 822.086721] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-9e052131-3359-4888-9d78-89440b639dcc tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] [instance: 8ff6fbb9-c90f-498d-9a85-d220a8c2f794] Processing image 50ff584c-3b50-4395-af07-3e66769bc9f7 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 822.086970] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9e052131-3359-4888-9d78-89440b639dcc tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 822.087160] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9e052131-3359-4888-9d78-89440b639dcc tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Acquired lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 822.087365] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-9e052131-3359-4888-9d78-89440b639dcc tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 822.087643] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-336667b9-d0a1-4096-a7ab-65bc24fcf32b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.106153] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-9e052131-3359-4888-9d78-89440b639dcc tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 822.106406] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-9e052131-3359-4888-9d78-89440b639dcc tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 822.107576] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c877350d-4bb2-403d-9808-1c54b10e6989 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.113495] env[62066]: DEBUG oslo_vmware.api [None req-9e052131-3359-4888-9d78-89440b639dcc tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Waiting for the task: (returnval){ [ 822.113495] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]52ad971c-8615-42e6-6628-31cb441fce7c" [ 822.113495] env[62066]: _type = "Task" [ 822.113495] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.122723] env[62066]: DEBUG oslo_vmware.api [None req-9e052131-3359-4888-9d78-89440b639dcc tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52ad971c-8615-42e6-6628-31cb441fce7c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.193032] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3d67ea1c-5f75-4620-b677-6f1e0faf1bd6 tempest-ServersV294TestFqdnHostnames-843300434 tempest-ServersV294TestFqdnHostnames-843300434-project-member] Releasing lock "refresh_cache-2709d0ad-5dd8-4e3c-b1e6-3d1e2e242380" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 822.193519] env[62066]: DEBUG nova.compute.manager [None req-3d67ea1c-5f75-4620-b677-6f1e0faf1bd6 tempest-ServersV294TestFqdnHostnames-843300434 tempest-ServersV294TestFqdnHostnames-843300434-project-member] [instance: 2709d0ad-5dd8-4e3c-b1e6-3d1e2e242380] Instance network_info: |[{"id": "c5cdc663-1e55-4238-bad4-7fb406b34a36", "address": "fa:16:3e:42:1a:81", "network": {"id": "2df566a5-1f6f-4665-a06d-1441530a4f80", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-1008410317-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d1785af300d545e6913c0e91148f4563", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b4d548e7-d762-406a-bb2d-dc7168a8ca67", "external-id": "nsx-vlan-transportzone-796", "segmentation_id": 796, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc5cdc663-1e", "ovs_interfaceid": "c5cdc663-1e55-4238-bad4-7fb406b34a36", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 822.194145] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-3d67ea1c-5f75-4620-b677-6f1e0faf1bd6 tempest-ServersV294TestFqdnHostnames-843300434 tempest-ServersV294TestFqdnHostnames-843300434-project-member] [instance: 2709d0ad-5dd8-4e3c-b1e6-3d1e2e242380] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:42:1a:81', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'b4d548e7-d762-406a-bb2d-dc7168a8ca67', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c5cdc663-1e55-4238-bad4-7fb406b34a36', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 822.203500] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d67ea1c-5f75-4620-b677-6f1e0faf1bd6 tempest-ServersV294TestFqdnHostnames-843300434 tempest-ServersV294TestFqdnHostnames-843300434-project-member] Creating folder: Project (d1785af300d545e6913c0e91148f4563). Parent ref: group-v285980. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 822.203889] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ec3003b5-d295-4ab7-9be0-16fab8982033 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.215212] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-3d67ea1c-5f75-4620-b677-6f1e0faf1bd6 tempest-ServersV294TestFqdnHostnames-843300434 tempest-ServersV294TestFqdnHostnames-843300434-project-member] Created folder: Project (d1785af300d545e6913c0e91148f4563) in parent group-v285980. [ 822.215341] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d67ea1c-5f75-4620-b677-6f1e0faf1bd6 tempest-ServersV294TestFqdnHostnames-843300434 tempest-ServersV294TestFqdnHostnames-843300434-project-member] Creating folder: Instances. Parent ref: group-v286047. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 822.215676] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-df9bbffd-0a80-45dd-9419-666dfc3e331a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.224717] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-3d67ea1c-5f75-4620-b677-6f1e0faf1bd6 tempest-ServersV294TestFqdnHostnames-843300434 tempest-ServersV294TestFqdnHostnames-843300434-project-member] Created folder: Instances in parent group-v286047. [ 822.225089] env[62066]: DEBUG oslo.service.loopingcall [None req-3d67ea1c-5f75-4620-b677-6f1e0faf1bd6 tempest-ServersV294TestFqdnHostnames-843300434 tempest-ServersV294TestFqdnHostnames-843300434-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 822.225398] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2709d0ad-5dd8-4e3c-b1e6-3d1e2e242380] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 822.225722] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a07017ce-5aa3-4289-bf17-fc940acad738 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.246245] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 822.246245] env[62066]: value = "task-1340832" [ 822.246245] env[62066]: _type = "Task" [ 822.246245] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.254270] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1340832, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.352999] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d2edd17d-ac3d-4b81-90f1-278c62ed416c tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.474s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 822.353655] env[62066]: DEBUG nova.compute.manager [None req-d2edd17d-ac3d-4b81-90f1-278c62ed416c tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] [instance: de2e2be7-efdb-45a8-842a-640ab9deb1d9] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 822.357132] env[62066]: DEBUG oslo_concurrency.lockutils [None req-acef35f4-b906-4efb-ae97-e8cdc80e8e5a tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.692s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 822.358969] env[62066]: INFO nova.compute.claims [None req-acef35f4-b906-4efb-ae97-e8cdc80e8e5a tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] [instance: 5e96def2-0cbd-4bd9-93f4-6a365a0142b0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 822.443657] env[62066]: DEBUG nova.compute.manager [req-4f80553e-512c-4475-a128-7dd1852a4579 req-3bcec331-8e70-485a-bb56-4fb8b3721b5d service nova] [instance: 2709d0ad-5dd8-4e3c-b1e6-3d1e2e242380] Received event network-changed-c5cdc663-1e55-4238-bad4-7fb406b34a36 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 822.444057] env[62066]: DEBUG nova.compute.manager [req-4f80553e-512c-4475-a128-7dd1852a4579 req-3bcec331-8e70-485a-bb56-4fb8b3721b5d service nova] [instance: 2709d0ad-5dd8-4e3c-b1e6-3d1e2e242380] Refreshing instance network info cache due to event network-changed-c5cdc663-1e55-4238-bad4-7fb406b34a36. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 822.444449] env[62066]: DEBUG oslo_concurrency.lockutils [req-4f80553e-512c-4475-a128-7dd1852a4579 req-3bcec331-8e70-485a-bb56-4fb8b3721b5d service nova] Acquiring lock "refresh_cache-2709d0ad-5dd8-4e3c-b1e6-3d1e2e242380" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 822.444739] env[62066]: DEBUG oslo_concurrency.lockutils [req-4f80553e-512c-4475-a128-7dd1852a4579 req-3bcec331-8e70-485a-bb56-4fb8b3721b5d service nova] Acquired lock "refresh_cache-2709d0ad-5dd8-4e3c-b1e6-3d1e2e242380" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 822.445023] env[62066]: DEBUG nova.network.neutron [req-4f80553e-512c-4475-a128-7dd1852a4579 req-3bcec331-8e70-485a-bb56-4fb8b3721b5d service nova] [instance: 2709d0ad-5dd8-4e3c-b1e6-3d1e2e242380] Refreshing network info cache for port c5cdc663-1e55-4238-bad4-7fb406b34a36 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 822.466921] env[62066]: DEBUG oslo_vmware.api [None req-90f5cbab-7c0d-4b69-92aa-09bf24d9f7da tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] Task: {'id': task-1340825, 'name': PowerOnVM_Task, 'duration_secs': 1.083108} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.467533] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-90f5cbab-7c0d-4b69-92aa-09bf24d9f7da tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] [instance: 050040cb-6ee3-4ad2-960a-fcebb53ac394] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 822.467888] env[62066]: INFO nova.compute.manager [None req-90f5cbab-7c0d-4b69-92aa-09bf24d9f7da tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] [instance: 050040cb-6ee3-4ad2-960a-fcebb53ac394] Took 15.00 seconds to spawn the instance on the hypervisor. [ 822.468240] env[62066]: DEBUG nova.compute.manager [None req-90f5cbab-7c0d-4b69-92aa-09bf24d9f7da tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] [instance: 050040cb-6ee3-4ad2-960a-fcebb53ac394] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 822.469696] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75aea4de-9e51-4da8-8a4d-ffff0639d576 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.555557] env[62066]: DEBUG nova.network.neutron [None req-b8836db2-293e-442c-80f2-42d8bc4c2918 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 88daaaa6-f385-4161-bc74-9fffabb1145d] Successfully updated port: ba002f90-5218-46f2-b7e1-cd96ae3018c7 {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 822.586473] env[62066]: INFO nova.compute.manager [-] [instance: 49bc1350-0095-406b-bc68-005eb6b681a8] Took 1.53 seconds to deallocate network for instance. [ 822.632578] env[62066]: DEBUG oslo_vmware.api [None req-9e052131-3359-4888-9d78-89440b639dcc tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52ad971c-8615-42e6-6628-31cb441fce7c, 'name': SearchDatastore_Task, 'duration_secs': 0.013546} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.634184] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c80cf18e-7327-4c55-8c56-1875a56f7af7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.641277] env[62066]: DEBUG oslo_vmware.api [None req-9e052131-3359-4888-9d78-89440b639dcc tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Waiting for the task: (returnval){ [ 822.641277] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]523a6a53-a67c-f661-bb70-f09acdc146d2" [ 822.641277] env[62066]: _type = "Task" [ 822.641277] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.649246] env[62066]: DEBUG oslo_vmware.api [None req-9e052131-3359-4888-9d78-89440b639dcc tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]523a6a53-a67c-f661-bb70-f09acdc146d2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.757834] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1340832, 'name': CreateVM_Task, 'duration_secs': 0.466524} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.759050] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2709d0ad-5dd8-4e3c-b1e6-3d1e2e242380] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 822.759050] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3d67ea1c-5f75-4620-b677-6f1e0faf1bd6 tempest-ServersV294TestFqdnHostnames-843300434 tempest-ServersV294TestFqdnHostnames-843300434-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 822.759494] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3d67ea1c-5f75-4620-b677-6f1e0faf1bd6 tempest-ServersV294TestFqdnHostnames-843300434 tempest-ServersV294TestFqdnHostnames-843300434-project-member] Acquired lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 822.760061] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3d67ea1c-5f75-4620-b677-6f1e0faf1bd6 tempest-ServersV294TestFqdnHostnames-843300434 tempest-ServersV294TestFqdnHostnames-843300434-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 822.760700] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-91552c76-7ef5-4192-80b4-d4533ded04d6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.767034] env[62066]: DEBUG oslo_vmware.api [None req-3d67ea1c-5f75-4620-b677-6f1e0faf1bd6 tempest-ServersV294TestFqdnHostnames-843300434 tempest-ServersV294TestFqdnHostnames-843300434-project-member] Waiting for the task: (returnval){ [ 822.767034] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]524d64e4-5438-87e5-7390-bcec16c79938" [ 822.767034] env[62066]: _type = "Task" [ 822.767034] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.777153] env[62066]: DEBUG oslo_vmware.api [None req-3d67ea1c-5f75-4620-b677-6f1e0faf1bd6 tempest-ServersV294TestFqdnHostnames-843300434 tempest-ServersV294TestFqdnHostnames-843300434-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]524d64e4-5438-87e5-7390-bcec16c79938, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.864132] env[62066]: DEBUG nova.compute.utils [None req-d2edd17d-ac3d-4b81-90f1-278c62ed416c tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 822.869431] env[62066]: DEBUG nova.compute.manager [None req-d2edd17d-ac3d-4b81-90f1-278c62ed416c tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] [instance: de2e2be7-efdb-45a8-842a-640ab9deb1d9] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 822.869431] env[62066]: DEBUG nova.network.neutron [None req-d2edd17d-ac3d-4b81-90f1-278c62ed416c tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] [instance: de2e2be7-efdb-45a8-842a-640ab9deb1d9] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 822.915746] env[62066]: DEBUG nova.policy [None req-d2edd17d-ac3d-4b81-90f1-278c62ed416c tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9d11ab2ea74442369fa1d93daf98392b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'dd30584efc254610b3939d1ba3806693', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 822.988787] env[62066]: INFO nova.compute.manager [None req-90f5cbab-7c0d-4b69-92aa-09bf24d9f7da tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] [instance: 050040cb-6ee3-4ad2-960a-fcebb53ac394] Took 31.75 seconds to build instance. [ 823.015832] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd69318f-f46a-470d-8a41-86dff862e23b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.024519] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-3a2ae741-acf0-4694-aaae-0cdc986fd6f3 tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] [instance: 5f645eaa-6edc-4362-82ad-38c4c57b2be2] Doing hard reboot of VM {{(pid=62066) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1063}} [ 823.025332] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ResetVM_Task with opID=oslo.vmware-9f5381d3-645f-4d6d-a4ca-d6b7d30d08ad {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.033991] env[62066]: DEBUG oslo_vmware.api [None req-3a2ae741-acf0-4694-aaae-0cdc986fd6f3 tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Waiting for the task: (returnval){ [ 823.033991] env[62066]: value = "task-1340833" [ 823.033991] env[62066]: _type = "Task" [ 823.033991] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.052084] env[62066]: DEBUG oslo_vmware.api [None req-3a2ae741-acf0-4694-aaae-0cdc986fd6f3 tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Task: {'id': task-1340833, 'name': ResetVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.058791] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b8836db2-293e-442c-80f2-42d8bc4c2918 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Acquiring lock "refresh_cache-88daaaa6-f385-4161-bc74-9fffabb1145d" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 823.059034] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b8836db2-293e-442c-80f2-42d8bc4c2918 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Acquired lock "refresh_cache-88daaaa6-f385-4161-bc74-9fffabb1145d" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 823.059252] env[62066]: DEBUG nova.network.neutron [None req-b8836db2-293e-442c-80f2-42d8bc4c2918 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 88daaaa6-f385-4161-bc74-9fffabb1145d] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 823.093301] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1e760fa3-f701-40c8-8666-21a2eeae0751 tempest-ServerAddressesTestJSON-1053293248 tempest-ServerAddressesTestJSON-1053293248-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 823.154372] env[62066]: DEBUG oslo_vmware.api [None req-9e052131-3359-4888-9d78-89440b639dcc tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]523a6a53-a67c-f661-bb70-f09acdc146d2, 'name': SearchDatastore_Task, 'duration_secs': 0.01581} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.154728] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9e052131-3359-4888-9d78-89440b639dcc tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Releasing lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 823.154940] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-9e052131-3359-4888-9d78-89440b639dcc tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk to [datastore2] 8ff6fbb9-c90f-498d-9a85-d220a8c2f794/8ff6fbb9-c90f-498d-9a85-d220a8c2f794.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 823.155282] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-84cb6355-4134-4e2c-bb16-578ad7e39a28 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.161812] env[62066]: DEBUG oslo_vmware.api [None req-9e052131-3359-4888-9d78-89440b639dcc tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Waiting for the task: (returnval){ [ 823.161812] env[62066]: value = "task-1340834" [ 823.161812] env[62066]: _type = "Task" [ 823.161812] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.170761] env[62066]: DEBUG oslo_vmware.api [None req-9e052131-3359-4888-9d78-89440b639dcc tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Task: {'id': task-1340834, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.278250] env[62066]: DEBUG oslo_vmware.api [None req-3d67ea1c-5f75-4620-b677-6f1e0faf1bd6 tempest-ServersV294TestFqdnHostnames-843300434 tempest-ServersV294TestFqdnHostnames-843300434-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]524d64e4-5438-87e5-7390-bcec16c79938, 'name': SearchDatastore_Task, 'duration_secs': 0.013753} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.278776] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3d67ea1c-5f75-4620-b677-6f1e0faf1bd6 tempest-ServersV294TestFqdnHostnames-843300434 tempest-ServersV294TestFqdnHostnames-843300434-project-member] Releasing lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 823.279091] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-3d67ea1c-5f75-4620-b677-6f1e0faf1bd6 tempest-ServersV294TestFqdnHostnames-843300434 tempest-ServersV294TestFqdnHostnames-843300434-project-member] [instance: 2709d0ad-5dd8-4e3c-b1e6-3d1e2e242380] Processing image 50ff584c-3b50-4395-af07-3e66769bc9f7 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 823.279372] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3d67ea1c-5f75-4620-b677-6f1e0faf1bd6 tempest-ServersV294TestFqdnHostnames-843300434 tempest-ServersV294TestFqdnHostnames-843300434-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 823.279559] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3d67ea1c-5f75-4620-b677-6f1e0faf1bd6 tempest-ServersV294TestFqdnHostnames-843300434 tempest-ServersV294TestFqdnHostnames-843300434-project-member] Acquired lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 823.279868] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-3d67ea1c-5f75-4620-b677-6f1e0faf1bd6 tempest-ServersV294TestFqdnHostnames-843300434 tempest-ServersV294TestFqdnHostnames-843300434-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 823.280414] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b94c2e30-1252-47c9-b14d-9014853a5cb7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.290487] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-3d67ea1c-5f75-4620-b677-6f1e0faf1bd6 tempest-ServersV294TestFqdnHostnames-843300434 tempest-ServersV294TestFqdnHostnames-843300434-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 823.290917] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-3d67ea1c-5f75-4620-b677-6f1e0faf1bd6 tempest-ServersV294TestFqdnHostnames-843300434 tempest-ServersV294TestFqdnHostnames-843300434-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 823.291915] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6e01face-8aa1-4e62-af19-63f6d1fc6856 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.298733] env[62066]: DEBUG oslo_vmware.api [None req-3d67ea1c-5f75-4620-b677-6f1e0faf1bd6 tempest-ServersV294TestFqdnHostnames-843300434 tempest-ServersV294TestFqdnHostnames-843300434-project-member] Waiting for the task: (returnval){ [ 823.298733] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]52930082-e915-e577-f63c-bc8d83990cfc" [ 823.298733] env[62066]: _type = "Task" [ 823.298733] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.309178] env[62066]: DEBUG oslo_vmware.api [None req-3d67ea1c-5f75-4620-b677-6f1e0faf1bd6 tempest-ServersV294TestFqdnHostnames-843300434 tempest-ServersV294TestFqdnHostnames-843300434-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52930082-e915-e577-f63c-bc8d83990cfc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.331110] env[62066]: DEBUG nova.network.neutron [req-4f80553e-512c-4475-a128-7dd1852a4579 req-3bcec331-8e70-485a-bb56-4fb8b3721b5d service nova] [instance: 2709d0ad-5dd8-4e3c-b1e6-3d1e2e242380] Updated VIF entry in instance network info cache for port c5cdc663-1e55-4238-bad4-7fb406b34a36. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 823.331604] env[62066]: DEBUG nova.network.neutron [req-4f80553e-512c-4475-a128-7dd1852a4579 req-3bcec331-8e70-485a-bb56-4fb8b3721b5d service nova] [instance: 2709d0ad-5dd8-4e3c-b1e6-3d1e2e242380] Updating instance_info_cache with network_info: [{"id": "c5cdc663-1e55-4238-bad4-7fb406b34a36", "address": "fa:16:3e:42:1a:81", "network": {"id": "2df566a5-1f6f-4665-a06d-1441530a4f80", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-1008410317-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d1785af300d545e6913c0e91148f4563", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b4d548e7-d762-406a-bb2d-dc7168a8ca67", "external-id": "nsx-vlan-transportzone-796", "segmentation_id": 796, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc5cdc663-1e", "ovs_interfaceid": "c5cdc663-1e55-4238-bad4-7fb406b34a36", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 823.377440] env[62066]: DEBUG nova.compute.manager [None req-d2edd17d-ac3d-4b81-90f1-278c62ed416c tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] [instance: de2e2be7-efdb-45a8-842a-640ab9deb1d9] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 823.405253] env[62066]: DEBUG nova.network.neutron [None req-d2edd17d-ac3d-4b81-90f1-278c62ed416c tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] [instance: de2e2be7-efdb-45a8-842a-640ab9deb1d9] Successfully created port: c833a993-4af0-4bdb-90d2-cc82a796a560 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 823.491136] env[62066]: DEBUG oslo_concurrency.lockutils [None req-90f5cbab-7c0d-4b69-92aa-09bf24d9f7da tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] Lock "050040cb-6ee3-4ad2-960a-fcebb53ac394" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 104.990s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 823.550378] env[62066]: DEBUG oslo_vmware.api [None req-3a2ae741-acf0-4694-aaae-0cdc986fd6f3 tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Task: {'id': task-1340833, 'name': ResetVM_Task, 'duration_secs': 0.117192} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.552796] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-3a2ae741-acf0-4694-aaae-0cdc986fd6f3 tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] [instance: 5f645eaa-6edc-4362-82ad-38c4c57b2be2] Did hard reboot of VM {{(pid=62066) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1067}} [ 823.552796] env[62066]: DEBUG nova.compute.manager [None req-3a2ae741-acf0-4694-aaae-0cdc986fd6f3 tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] [instance: 5f645eaa-6edc-4362-82ad-38c4c57b2be2] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 823.552796] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbd1c436-7356-4e32-b2f8-02de1c03fa16 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.587443] env[62066]: DEBUG oslo_concurrency.lockutils [None req-32db9ed7-c319-413e-978b-0abcfac2cab3 tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] Acquiring lock "050040cb-6ee3-4ad2-960a-fcebb53ac394" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 823.587825] env[62066]: DEBUG oslo_concurrency.lockutils [None req-32db9ed7-c319-413e-978b-0abcfac2cab3 tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] Lock "050040cb-6ee3-4ad2-960a-fcebb53ac394" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 823.588065] env[62066]: DEBUG oslo_concurrency.lockutils [None req-32db9ed7-c319-413e-978b-0abcfac2cab3 tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] Acquiring lock "050040cb-6ee3-4ad2-960a-fcebb53ac394-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 823.588325] env[62066]: DEBUG oslo_concurrency.lockutils [None req-32db9ed7-c319-413e-978b-0abcfac2cab3 tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] Lock "050040cb-6ee3-4ad2-960a-fcebb53ac394-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 823.588435] env[62066]: DEBUG oslo_concurrency.lockutils [None req-32db9ed7-c319-413e-978b-0abcfac2cab3 tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] Lock "050040cb-6ee3-4ad2-960a-fcebb53ac394-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 823.590828] env[62066]: INFO nova.compute.manager [None req-32db9ed7-c319-413e-978b-0abcfac2cab3 tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] [instance: 050040cb-6ee3-4ad2-960a-fcebb53ac394] Terminating instance [ 823.593660] env[62066]: DEBUG nova.compute.manager [None req-32db9ed7-c319-413e-978b-0abcfac2cab3 tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] [instance: 050040cb-6ee3-4ad2-960a-fcebb53ac394] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 823.593960] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-32db9ed7-c319-413e-978b-0abcfac2cab3 tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] [instance: 050040cb-6ee3-4ad2-960a-fcebb53ac394] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 823.595031] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e9ff4ab-d8fd-4c12-968c-72a862b9aab6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.602911] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-32db9ed7-c319-413e-978b-0abcfac2cab3 tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] [instance: 050040cb-6ee3-4ad2-960a-fcebb53ac394] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 823.606480] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-186c9f20-494e-40ea-b126-7e413aaeafab {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.611196] env[62066]: DEBUG nova.network.neutron [None req-b8836db2-293e-442c-80f2-42d8bc4c2918 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 88daaaa6-f385-4161-bc74-9fffabb1145d] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 823.615660] env[62066]: DEBUG oslo_vmware.api [None req-32db9ed7-c319-413e-978b-0abcfac2cab3 tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] Waiting for the task: (returnval){ [ 823.615660] env[62066]: value = "task-1340835" [ 823.615660] env[62066]: _type = "Task" [ 823.615660] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.628092] env[62066]: DEBUG oslo_vmware.api [None req-32db9ed7-c319-413e-978b-0abcfac2cab3 tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] Task: {'id': task-1340835, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.673703] env[62066]: DEBUG oslo_service.periodic_task [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 823.674187] env[62066]: DEBUG oslo_vmware.api [None req-9e052131-3359-4888-9d78-89440b639dcc tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Task: {'id': task-1340834, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.674724] env[62066]: DEBUG oslo_service.periodic_task [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 823.678135] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d1e2182-cd1b-4b69-9ff1-6c08cc7e9e90 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.690201] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4275248-001f-4f8a-8685-337bb3cc7fa9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.726141] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d03cbad6-1eca-4f47-99c6-1618c24f7965 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.738079] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9c5513c-b62a-49a7-8492-57f3cb937d99 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.755856] env[62066]: DEBUG nova.compute.provider_tree [None req-acef35f4-b906-4efb-ae97-e8cdc80e8e5a tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 823.809045] env[62066]: DEBUG oslo_vmware.api [None req-3d67ea1c-5f75-4620-b677-6f1e0faf1bd6 tempest-ServersV294TestFqdnHostnames-843300434 tempest-ServersV294TestFqdnHostnames-843300434-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52930082-e915-e577-f63c-bc8d83990cfc, 'name': SearchDatastore_Task, 'duration_secs': 0.016689} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.809884] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1fba8c26-9231-461f-b402-8d3d6627f950 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.815778] env[62066]: DEBUG oslo_vmware.api [None req-3d67ea1c-5f75-4620-b677-6f1e0faf1bd6 tempest-ServersV294TestFqdnHostnames-843300434 tempest-ServersV294TestFqdnHostnames-843300434-project-member] Waiting for the task: (returnval){ [ 823.815778] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]522d3e26-6fe0-2832-e40d-ba5d25ab3909" [ 823.815778] env[62066]: _type = "Task" [ 823.815778] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.824702] env[62066]: DEBUG oslo_vmware.api [None req-3d67ea1c-5f75-4620-b677-6f1e0faf1bd6 tempest-ServersV294TestFqdnHostnames-843300434 tempest-ServersV294TestFqdnHostnames-843300434-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]522d3e26-6fe0-2832-e40d-ba5d25ab3909, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.834340] env[62066]: DEBUG nova.network.neutron [None req-b8836db2-293e-442c-80f2-42d8bc4c2918 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 88daaaa6-f385-4161-bc74-9fffabb1145d] Updating instance_info_cache with network_info: [{"id": "ba002f90-5218-46f2-b7e1-cd96ae3018c7", "address": "fa:16:3e:54:27:b5", "network": {"id": "1c6c51e4-96f8-4175-b7f5-96b8a42fcb13", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-2006089809-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d609babc987a42e2a8ddb4bfb9c3b103", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a965790c-2d2f-4c2a-9ee7-745f4d53039b", "external-id": "nsx-vlan-transportzone-708", "segmentation_id": 708, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapba002f90-52", "ovs_interfaceid": "ba002f90-5218-46f2-b7e1-cd96ae3018c7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 823.836370] env[62066]: DEBUG oslo_concurrency.lockutils [req-4f80553e-512c-4475-a128-7dd1852a4579 req-3bcec331-8e70-485a-bb56-4fb8b3721b5d service nova] Releasing lock "refresh_cache-2709d0ad-5dd8-4e3c-b1e6-3d1e2e242380" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 823.836696] env[62066]: DEBUG nova.compute.manager [req-4f80553e-512c-4475-a128-7dd1852a4579 req-3bcec331-8e70-485a-bb56-4fb8b3721b5d service nova] [instance: 49bc1350-0095-406b-bc68-005eb6b681a8] Received event network-vif-deleted-f2c406e7-6447-47bf-9ee1-dba97dc7e5bc {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 823.836894] env[62066]: DEBUG nova.compute.manager [req-4f80553e-512c-4475-a128-7dd1852a4579 req-3bcec331-8e70-485a-bb56-4fb8b3721b5d service nova] [instance: 88daaaa6-f385-4161-bc74-9fffabb1145d] Received event network-vif-plugged-ba002f90-5218-46f2-b7e1-cd96ae3018c7 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 823.837249] env[62066]: DEBUG oslo_concurrency.lockutils [req-4f80553e-512c-4475-a128-7dd1852a4579 req-3bcec331-8e70-485a-bb56-4fb8b3721b5d service nova] Acquiring lock "88daaaa6-f385-4161-bc74-9fffabb1145d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 823.837656] env[62066]: DEBUG oslo_concurrency.lockutils [req-4f80553e-512c-4475-a128-7dd1852a4579 req-3bcec331-8e70-485a-bb56-4fb8b3721b5d service nova] Lock "88daaaa6-f385-4161-bc74-9fffabb1145d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 823.837883] env[62066]: DEBUG oslo_concurrency.lockutils [req-4f80553e-512c-4475-a128-7dd1852a4579 req-3bcec331-8e70-485a-bb56-4fb8b3721b5d service nova] Lock "88daaaa6-f385-4161-bc74-9fffabb1145d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 823.838082] env[62066]: DEBUG nova.compute.manager [req-4f80553e-512c-4475-a128-7dd1852a4579 req-3bcec331-8e70-485a-bb56-4fb8b3721b5d service nova] [instance: 88daaaa6-f385-4161-bc74-9fffabb1145d] No waiting events found dispatching network-vif-plugged-ba002f90-5218-46f2-b7e1-cd96ae3018c7 {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 823.838263] env[62066]: WARNING nova.compute.manager [req-4f80553e-512c-4475-a128-7dd1852a4579 req-3bcec331-8e70-485a-bb56-4fb8b3721b5d service nova] [instance: 88daaaa6-f385-4161-bc74-9fffabb1145d] Received unexpected event network-vif-plugged-ba002f90-5218-46f2-b7e1-cd96ae3018c7 for instance with vm_state building and task_state spawning. [ 824.068511] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3a2ae741-acf0-4694-aaae-0cdc986fd6f3 tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Lock "5f645eaa-6edc-4362-82ad-38c4c57b2be2" "released" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: held 4.303s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 824.127752] env[62066]: DEBUG oslo_vmware.api [None req-32db9ed7-c319-413e-978b-0abcfac2cab3 tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] Task: {'id': task-1340835, 'name': PowerOffVM_Task, 'duration_secs': 0.210434} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.128172] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-32db9ed7-c319-413e-978b-0abcfac2cab3 tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] [instance: 050040cb-6ee3-4ad2-960a-fcebb53ac394] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 824.128431] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-32db9ed7-c319-413e-978b-0abcfac2cab3 tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] [instance: 050040cb-6ee3-4ad2-960a-fcebb53ac394] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 824.128756] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-43dc1e9f-2408-4629-b306-7e37beeb31bd {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.173546] env[62066]: DEBUG oslo_vmware.api [None req-9e052131-3359-4888-9d78-89440b639dcc tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Task: {'id': task-1340834, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.730866} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.173924] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-9e052131-3359-4888-9d78-89440b639dcc tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk to [datastore2] 8ff6fbb9-c90f-498d-9a85-d220a8c2f794/8ff6fbb9-c90f-498d-9a85-d220a8c2f794.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 824.174272] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-9e052131-3359-4888-9d78-89440b639dcc tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] [instance: 8ff6fbb9-c90f-498d-9a85-d220a8c2f794] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 824.174574] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-cf75b5de-9bc8-4fd3-8571-a11ce276a06e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.182188] env[62066]: DEBUG oslo_vmware.api [None req-9e052131-3359-4888-9d78-89440b639dcc tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Waiting for the task: (returnval){ [ 824.182188] env[62066]: value = "task-1340837" [ 824.182188] env[62066]: _type = "Task" [ 824.182188] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.193718] env[62066]: DEBUG oslo_service.periodic_task [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 824.193718] env[62066]: DEBUG nova.compute.manager [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Starting heal instance info cache {{(pid=62066) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 824.193857] env[62066]: DEBUG nova.compute.manager [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Rebuilding the list of instances to heal {{(pid=62066) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 824.199913] env[62066]: DEBUG oslo_vmware.api [None req-9e052131-3359-4888-9d78-89440b639dcc tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Task: {'id': task-1340837, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.222479] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-32db9ed7-c319-413e-978b-0abcfac2cab3 tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] [instance: 050040cb-6ee3-4ad2-960a-fcebb53ac394] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 824.222779] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-32db9ed7-c319-413e-978b-0abcfac2cab3 tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] [instance: 050040cb-6ee3-4ad2-960a-fcebb53ac394] Deleting contents of the VM from datastore datastore1 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 824.223133] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-32db9ed7-c319-413e-978b-0abcfac2cab3 tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] Deleting the datastore file [datastore1] 050040cb-6ee3-4ad2-960a-fcebb53ac394 {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 824.223506] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b0eb48e6-0a73-4914-8acd-001305f8adae {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.231884] env[62066]: DEBUG oslo_vmware.api [None req-32db9ed7-c319-413e-978b-0abcfac2cab3 tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] Waiting for the task: (returnval){ [ 824.231884] env[62066]: value = "task-1340838" [ 824.231884] env[62066]: _type = "Task" [ 824.231884] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.241342] env[62066]: DEBUG oslo_vmware.api [None req-32db9ed7-c319-413e-978b-0abcfac2cab3 tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] Task: {'id': task-1340838, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.259263] env[62066]: DEBUG nova.scheduler.client.report [None req-acef35f4-b906-4efb-ae97-e8cdc80e8e5a tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 824.327191] env[62066]: DEBUG oslo_vmware.api [None req-3d67ea1c-5f75-4620-b677-6f1e0faf1bd6 tempest-ServersV294TestFqdnHostnames-843300434 tempest-ServersV294TestFqdnHostnames-843300434-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]522d3e26-6fe0-2832-e40d-ba5d25ab3909, 'name': SearchDatastore_Task, 'duration_secs': 0.052097} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.327534] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3d67ea1c-5f75-4620-b677-6f1e0faf1bd6 tempest-ServersV294TestFqdnHostnames-843300434 tempest-ServersV294TestFqdnHostnames-843300434-project-member] Releasing lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 824.327813] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d67ea1c-5f75-4620-b677-6f1e0faf1bd6 tempest-ServersV294TestFqdnHostnames-843300434 tempest-ServersV294TestFqdnHostnames-843300434-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk to [datastore2] 2709d0ad-5dd8-4e3c-b1e6-3d1e2e242380/2709d0ad-5dd8-4e3c-b1e6-3d1e2e242380.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 824.328154] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-382b77df-8cab-4317-a388-f07041fc2aa2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.335025] env[62066]: DEBUG oslo_vmware.api [None req-3d67ea1c-5f75-4620-b677-6f1e0faf1bd6 tempest-ServersV294TestFqdnHostnames-843300434 tempest-ServersV294TestFqdnHostnames-843300434-project-member] Waiting for the task: (returnval){ [ 824.335025] env[62066]: value = "task-1340839" [ 824.335025] env[62066]: _type = "Task" [ 824.335025] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.339953] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b8836db2-293e-442c-80f2-42d8bc4c2918 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Releasing lock "refresh_cache-88daaaa6-f385-4161-bc74-9fffabb1145d" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 824.340303] env[62066]: DEBUG nova.compute.manager [None req-b8836db2-293e-442c-80f2-42d8bc4c2918 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 88daaaa6-f385-4161-bc74-9fffabb1145d] Instance network_info: |[{"id": "ba002f90-5218-46f2-b7e1-cd96ae3018c7", "address": "fa:16:3e:54:27:b5", "network": {"id": "1c6c51e4-96f8-4175-b7f5-96b8a42fcb13", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-2006089809-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d609babc987a42e2a8ddb4bfb9c3b103", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a965790c-2d2f-4c2a-9ee7-745f4d53039b", "external-id": "nsx-vlan-transportzone-708", "segmentation_id": 708, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapba002f90-52", "ovs_interfaceid": "ba002f90-5218-46f2-b7e1-cd96ae3018c7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 824.343503] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-b8836db2-293e-442c-80f2-42d8bc4c2918 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 88daaaa6-f385-4161-bc74-9fffabb1145d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:54:27:b5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a965790c-2d2f-4c2a-9ee7-745f4d53039b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ba002f90-5218-46f2-b7e1-cd96ae3018c7', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 824.351719] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-b8836db2-293e-442c-80f2-42d8bc4c2918 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Creating folder: Project (d609babc987a42e2a8ddb4bfb9c3b103). Parent ref: group-v285980. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 824.352503] env[62066]: DEBUG oslo_vmware.api [None req-3d67ea1c-5f75-4620-b677-6f1e0faf1bd6 tempest-ServersV294TestFqdnHostnames-843300434 tempest-ServersV294TestFqdnHostnames-843300434-project-member] Task: {'id': task-1340839, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.352779] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2c086b06-4ca8-43ee-b5cd-cd659e39faee {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.363441] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-b8836db2-293e-442c-80f2-42d8bc4c2918 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Created folder: Project (d609babc987a42e2a8ddb4bfb9c3b103) in parent group-v285980. [ 824.363705] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-b8836db2-293e-442c-80f2-42d8bc4c2918 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Creating folder: Instances. Parent ref: group-v286050. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 824.363970] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-89049f5b-be9f-4934-bccb-e76ffc456d4c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.373311] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-b8836db2-293e-442c-80f2-42d8bc4c2918 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Created folder: Instances in parent group-v286050. [ 824.373585] env[62066]: DEBUG oslo.service.loopingcall [None req-b8836db2-293e-442c-80f2-42d8bc4c2918 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 824.373785] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 88daaaa6-f385-4161-bc74-9fffabb1145d] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 824.374034] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f5426582-72b8-440c-bb98-59937c5a3623 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.389144] env[62066]: DEBUG nova.compute.manager [None req-d2edd17d-ac3d-4b81-90f1-278c62ed416c tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] [instance: de2e2be7-efdb-45a8-842a-640ab9deb1d9] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 824.397188] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 824.397188] env[62066]: value = "task-1340842" [ 824.397188] env[62066]: _type = "Task" [ 824.397188] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.404766] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1340842, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.692055] env[62066]: DEBUG oslo_vmware.api [None req-9e052131-3359-4888-9d78-89440b639dcc tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Task: {'id': task-1340837, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.118073} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.692260] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-9e052131-3359-4888-9d78-89440b639dcc tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] [instance: 8ff6fbb9-c90f-498d-9a85-d220a8c2f794] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 824.693057] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b2ab129-600b-418c-af6e-7d1556416337 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.710095] env[62066]: DEBUG nova.compute.manager [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] [instance: 050040cb-6ee3-4ad2-960a-fcebb53ac394] Skipping network cache update for instance because it is being deleted. {{(pid=62066) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9945}} [ 824.710314] env[62066]: DEBUG nova.compute.manager [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] [instance: 8ff6fbb9-c90f-498d-9a85-d220a8c2f794] Skipping network cache update for instance because it is Building. {{(pid=62066) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 824.710376] env[62066]: DEBUG nova.compute.manager [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] [instance: 2709d0ad-5dd8-4e3c-b1e6-3d1e2e242380] Skipping network cache update for instance because it is Building. {{(pid=62066) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 824.710538] env[62066]: DEBUG nova.compute.manager [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] [instance: 88daaaa6-f385-4161-bc74-9fffabb1145d] Skipping network cache update for instance because it is Building. {{(pid=62066) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 824.710676] env[62066]: DEBUG nova.compute.manager [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] [instance: de2e2be7-efdb-45a8-842a-640ab9deb1d9] Skipping network cache update for instance because it is Building. {{(pid=62066) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 824.710796] env[62066]: DEBUG nova.compute.manager [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] [instance: 5e96def2-0cbd-4bd9-93f4-6a365a0142b0] Skipping network cache update for instance because it is Building. {{(pid=62066) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 824.721599] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-9e052131-3359-4888-9d78-89440b639dcc tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] [instance: 8ff6fbb9-c90f-498d-9a85-d220a8c2f794] Reconfiguring VM instance instance-00000045 to attach disk [datastore2] 8ff6fbb9-c90f-498d-9a85-d220a8c2f794/8ff6fbb9-c90f-498d-9a85-d220a8c2f794.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 824.722011] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-988bde22-bcc3-456a-8b27-3c897362747f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.745395] env[62066]: DEBUG oslo_vmware.api [None req-32db9ed7-c319-413e-978b-0abcfac2cab3 tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] Task: {'id': task-1340838, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.263463} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.746834] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-32db9ed7-c319-413e-978b-0abcfac2cab3 tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 824.747075] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-32db9ed7-c319-413e-978b-0abcfac2cab3 tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] [instance: 050040cb-6ee3-4ad2-960a-fcebb53ac394] Deleted contents of the VM from datastore datastore1 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 824.747244] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-32db9ed7-c319-413e-978b-0abcfac2cab3 tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] [instance: 050040cb-6ee3-4ad2-960a-fcebb53ac394] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 824.747425] env[62066]: INFO nova.compute.manager [None req-32db9ed7-c319-413e-978b-0abcfac2cab3 tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] [instance: 050040cb-6ee3-4ad2-960a-fcebb53ac394] Took 1.15 seconds to destroy the instance on the hypervisor. [ 824.747676] env[62066]: DEBUG oslo.service.loopingcall [None req-32db9ed7-c319-413e-978b-0abcfac2cab3 tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 824.747944] env[62066]: DEBUG oslo_vmware.api [None req-9e052131-3359-4888-9d78-89440b639dcc tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Waiting for the task: (returnval){ [ 824.747944] env[62066]: value = "task-1340843" [ 824.747944] env[62066]: _type = "Task" [ 824.747944] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.748160] env[62066]: DEBUG nova.compute.manager [-] [instance: 050040cb-6ee3-4ad2-960a-fcebb53ac394] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 824.748254] env[62066]: DEBUG nova.network.neutron [-] [instance: 050040cb-6ee3-4ad2-960a-fcebb53ac394] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 824.758342] env[62066]: DEBUG oslo_vmware.api [None req-9e052131-3359-4888-9d78-89440b639dcc tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Task: {'id': task-1340843, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.764391] env[62066]: DEBUG oslo_concurrency.lockutils [None req-acef35f4-b906-4efb-ae97-e8cdc80e8e5a tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.407s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 824.764993] env[62066]: DEBUG nova.compute.manager [None req-acef35f4-b906-4efb-ae97-e8cdc80e8e5a tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] [instance: 5e96def2-0cbd-4bd9-93f4-6a365a0142b0] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 824.767840] env[62066]: DEBUG oslo_concurrency.lockutils [None req-02667002-03be-4979-8517-c36490492214 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.373s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 824.769442] env[62066]: INFO nova.compute.claims [None req-02667002-03be-4979-8517-c36490492214 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 86a92b61-f3e9-48f9-8ee2-756669d558ef] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 824.826273] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Acquiring lock "refresh_cache-1a4b0637-1a56-41ef-b89b-6b56d24ed206" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 824.826451] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Acquired lock "refresh_cache-1a4b0637-1a56-41ef-b89b-6b56d24ed206" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 824.826577] env[62066]: DEBUG nova.network.neutron [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] [instance: 1a4b0637-1a56-41ef-b89b-6b56d24ed206] Forcefully refreshing network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 824.826738] env[62066]: DEBUG nova.objects.instance [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Lazy-loading 'info_cache' on Instance uuid 1a4b0637-1a56-41ef-b89b-6b56d24ed206 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 824.845887] env[62066]: DEBUG oslo_vmware.api [None req-3d67ea1c-5f75-4620-b677-6f1e0faf1bd6 tempest-ServersV294TestFqdnHostnames-843300434 tempest-ServersV294TestFqdnHostnames-843300434-project-member] Task: {'id': task-1340839, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.909755] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1340842, 'name': CreateVM_Task} progress is 25%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.260553] env[62066]: DEBUG oslo_vmware.api [None req-9e052131-3359-4888-9d78-89440b639dcc tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Task: {'id': task-1340843, 'name': ReconfigVM_Task, 'duration_secs': 0.422692} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.260913] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-9e052131-3359-4888-9d78-89440b639dcc tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] [instance: 8ff6fbb9-c90f-498d-9a85-d220a8c2f794] Reconfigured VM instance instance-00000045 to attach disk [datastore2] 8ff6fbb9-c90f-498d-9a85-d220a8c2f794/8ff6fbb9-c90f-498d-9a85-d220a8c2f794.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 825.261703] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b1f1177f-b7f0-4758-a83a-b68e77350896 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.269634] env[62066]: DEBUG oslo_vmware.api [None req-9e052131-3359-4888-9d78-89440b639dcc tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Waiting for the task: (returnval){ [ 825.269634] env[62066]: value = "task-1340844" [ 825.269634] env[62066]: _type = "Task" [ 825.269634] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.274802] env[62066]: DEBUG nova.compute.utils [None req-acef35f4-b906-4efb-ae97-e8cdc80e8e5a tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 825.284018] env[62066]: DEBUG nova.compute.manager [None req-acef35f4-b906-4efb-ae97-e8cdc80e8e5a tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] [instance: 5e96def2-0cbd-4bd9-93f4-6a365a0142b0] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 825.284018] env[62066]: DEBUG nova.network.neutron [None req-acef35f4-b906-4efb-ae97-e8cdc80e8e5a tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] [instance: 5e96def2-0cbd-4bd9-93f4-6a365a0142b0] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 825.286113] env[62066]: DEBUG nova.network.neutron [None req-d2edd17d-ac3d-4b81-90f1-278c62ed416c tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] [instance: de2e2be7-efdb-45a8-842a-640ab9deb1d9] Successfully updated port: c833a993-4af0-4bdb-90d2-cc82a796a560 {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 825.293887] env[62066]: DEBUG oslo_vmware.api [None req-9e052131-3359-4888-9d78-89440b639dcc tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Task: {'id': task-1340844, 'name': Rename_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.347567] env[62066]: DEBUG oslo_vmware.api [None req-3d67ea1c-5f75-4620-b677-6f1e0faf1bd6 tempest-ServersV294TestFqdnHostnames-843300434 tempest-ServersV294TestFqdnHostnames-843300434-project-member] Task: {'id': task-1340839, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.587833} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.347863] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d67ea1c-5f75-4620-b677-6f1e0faf1bd6 tempest-ServersV294TestFqdnHostnames-843300434 tempest-ServersV294TestFqdnHostnames-843300434-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk to [datastore2] 2709d0ad-5dd8-4e3c-b1e6-3d1e2e242380/2709d0ad-5dd8-4e3c-b1e6-3d1e2e242380.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 825.348086] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-3d67ea1c-5f75-4620-b677-6f1e0faf1bd6 tempest-ServersV294TestFqdnHostnames-843300434 tempest-ServersV294TestFqdnHostnames-843300434-project-member] [instance: 2709d0ad-5dd8-4e3c-b1e6-3d1e2e242380] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 825.348362] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-139cccdb-323c-494b-b540-7d839f7e8d3d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.358684] env[62066]: DEBUG oslo_vmware.api [None req-3d67ea1c-5f75-4620-b677-6f1e0faf1bd6 tempest-ServersV294TestFqdnHostnames-843300434 tempest-ServersV294TestFqdnHostnames-843300434-project-member] Waiting for the task: (returnval){ [ 825.358684] env[62066]: value = "task-1340845" [ 825.358684] env[62066]: _type = "Task" [ 825.358684] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.369027] env[62066]: DEBUG oslo_vmware.api [None req-3d67ea1c-5f75-4620-b677-6f1e0faf1bd6 tempest-ServersV294TestFqdnHostnames-843300434 tempest-ServersV294TestFqdnHostnames-843300434-project-member] Task: {'id': task-1340845, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.386300] env[62066]: DEBUG nova.policy [None req-acef35f4-b906-4efb-ae97-e8cdc80e8e5a tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'dff959d830ce40869f66802c17b2572d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '59855b1cf91d43e3ba0482f414c4e546', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 825.408590] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1340842, 'name': CreateVM_Task, 'duration_secs': 0.618993} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.408833] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 88daaaa6-f385-4161-bc74-9fffabb1145d] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 825.409475] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b8836db2-293e-442c-80f2-42d8bc4c2918 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 825.409654] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b8836db2-293e-442c-80f2-42d8bc4c2918 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Acquired lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 825.409997] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b8836db2-293e-442c-80f2-42d8bc4c2918 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 825.410279] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8e2dd853-6042-4ec2-9995-021baa4a4f4c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.415421] env[62066]: DEBUG oslo_vmware.api [None req-b8836db2-293e-442c-80f2-42d8bc4c2918 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Waiting for the task: (returnval){ [ 825.415421] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]52994a9d-2e0f-5597-8129-be46e1e7d699" [ 825.415421] env[62066]: _type = "Task" [ 825.415421] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.423771] env[62066]: DEBUG oslo_vmware.api [None req-b8836db2-293e-442c-80f2-42d8bc4c2918 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52994a9d-2e0f-5597-8129-be46e1e7d699, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.738100] env[62066]: DEBUG nova.network.neutron [None req-acef35f4-b906-4efb-ae97-e8cdc80e8e5a tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] [instance: 5e96def2-0cbd-4bd9-93f4-6a365a0142b0] Successfully created port: 0932c7ac-630e-4cef-8991-45b438f39483 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 825.781121] env[62066]: DEBUG nova.compute.manager [None req-acef35f4-b906-4efb-ae97-e8cdc80e8e5a tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] [instance: 5e96def2-0cbd-4bd9-93f4-6a365a0142b0] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 825.784683] env[62066]: DEBUG oslo_vmware.api [None req-9e052131-3359-4888-9d78-89440b639dcc tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Task: {'id': task-1340844, 'name': Rename_Task, 'duration_secs': 0.166762} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.789623] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-9e052131-3359-4888-9d78-89440b639dcc tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] [instance: 8ff6fbb9-c90f-498d-9a85-d220a8c2f794] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 825.789623] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5e95c55a-5a14-46f1-a341-e9d6bd69db8c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.796494] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d2edd17d-ac3d-4b81-90f1-278c62ed416c tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Acquiring lock "refresh_cache-de2e2be7-efdb-45a8-842a-640ab9deb1d9" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 825.796601] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d2edd17d-ac3d-4b81-90f1-278c62ed416c tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Acquired lock "refresh_cache-de2e2be7-efdb-45a8-842a-640ab9deb1d9" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 825.796647] env[62066]: DEBUG nova.network.neutron [None req-d2edd17d-ac3d-4b81-90f1-278c62ed416c tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] [instance: de2e2be7-efdb-45a8-842a-640ab9deb1d9] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 825.800114] env[62066]: DEBUG oslo_vmware.api [None req-9e052131-3359-4888-9d78-89440b639dcc tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Waiting for the task: (returnval){ [ 825.800114] env[62066]: value = "task-1340846" [ 825.800114] env[62066]: _type = "Task" [ 825.800114] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.808723] env[62066]: DEBUG oslo_vmware.api [None req-9e052131-3359-4888-9d78-89440b639dcc tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Task: {'id': task-1340846, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.872383] env[62066]: DEBUG oslo_vmware.api [None req-3d67ea1c-5f75-4620-b677-6f1e0faf1bd6 tempest-ServersV294TestFqdnHostnames-843300434 tempest-ServersV294TestFqdnHostnames-843300434-project-member] Task: {'id': task-1340845, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.124935} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.875612] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-3d67ea1c-5f75-4620-b677-6f1e0faf1bd6 tempest-ServersV294TestFqdnHostnames-843300434 tempest-ServersV294TestFqdnHostnames-843300434-project-member] [instance: 2709d0ad-5dd8-4e3c-b1e6-3d1e2e242380] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 825.876802] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59c57e28-cf3b-4708-becd-15d8c1dce7f0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.901886] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-3d67ea1c-5f75-4620-b677-6f1e0faf1bd6 tempest-ServersV294TestFqdnHostnames-843300434 tempest-ServersV294TestFqdnHostnames-843300434-project-member] [instance: 2709d0ad-5dd8-4e3c-b1e6-3d1e2e242380] Reconfiguring VM instance instance-00000046 to attach disk [datastore2] 2709d0ad-5dd8-4e3c-b1e6-3d1e2e242380/2709d0ad-5dd8-4e3c-b1e6-3d1e2e242380.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 825.907024] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-94b7c350-0eaa-4be2-824e-5e5ecb3c6fb8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.932868] env[62066]: DEBUG oslo_vmware.api [None req-b8836db2-293e-442c-80f2-42d8bc4c2918 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52994a9d-2e0f-5597-8129-be46e1e7d699, 'name': SearchDatastore_Task, 'duration_secs': 0.014451} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.936816] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b8836db2-293e-442c-80f2-42d8bc4c2918 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Releasing lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 825.937178] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-b8836db2-293e-442c-80f2-42d8bc4c2918 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 88daaaa6-f385-4161-bc74-9fffabb1145d] Processing image 50ff584c-3b50-4395-af07-3e66769bc9f7 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 825.937353] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b8836db2-293e-442c-80f2-42d8bc4c2918 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 825.937506] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b8836db2-293e-442c-80f2-42d8bc4c2918 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Acquired lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 825.937689] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-b8836db2-293e-442c-80f2-42d8bc4c2918 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 825.938052] env[62066]: DEBUG oslo_vmware.api [None req-3d67ea1c-5f75-4620-b677-6f1e0faf1bd6 tempest-ServersV294TestFqdnHostnames-843300434 tempest-ServersV294TestFqdnHostnames-843300434-project-member] Waiting for the task: (returnval){ [ 825.938052] env[62066]: value = "task-1340847" [ 825.938052] env[62066]: _type = "Task" [ 825.938052] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.938796] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-41a4819b-cc34-4a73-bd69-ee7dcebf1205 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.949862] env[62066]: DEBUG oslo_vmware.api [None req-3d67ea1c-5f75-4620-b677-6f1e0faf1bd6 tempest-ServersV294TestFqdnHostnames-843300434 tempest-ServersV294TestFqdnHostnames-843300434-project-member] Task: {'id': task-1340847, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.954757] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-b8836db2-293e-442c-80f2-42d8bc4c2918 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 825.954757] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-b8836db2-293e-442c-80f2-42d8bc4c2918 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 825.954757] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9e939ffa-49d6-4b74-ba35-3b589399179f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.961718] env[62066]: DEBUG oslo_vmware.api [None req-b8836db2-293e-442c-80f2-42d8bc4c2918 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Waiting for the task: (returnval){ [ 825.961718] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]5299027a-2e08-6ce0-6410-bdca5e7c038b" [ 825.961718] env[62066]: _type = "Task" [ 825.961718] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.968303] env[62066]: DEBUG oslo_vmware.api [None req-b8836db2-293e-442c-80f2-42d8bc4c2918 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]5299027a-2e08-6ce0-6410-bdca5e7c038b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.105943] env[62066]: DEBUG nova.network.neutron [-] [instance: 050040cb-6ee3-4ad2-960a-fcebb53ac394] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 826.127806] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d128204d-71a1-4b3a-ab42-21eaf125bb9c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.137691] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c1f0b4e-4170-47d3-9dca-9088dd2f5cde {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.168563] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a9a9291-8780-447c-bea1-0acbda7105b0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.176443] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c471da1-0e0e-4189-b883-1f0c21b1ed7b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.190089] env[62066]: DEBUG nova.compute.provider_tree [None req-02667002-03be-4979-8517-c36490492214 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 826.310862] env[62066]: DEBUG oslo_vmware.api [None req-9e052131-3359-4888-9d78-89440b639dcc tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Task: {'id': task-1340846, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.334989] env[62066]: DEBUG nova.network.neutron [None req-d2edd17d-ac3d-4b81-90f1-278c62ed416c tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] [instance: de2e2be7-efdb-45a8-842a-640ab9deb1d9] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 826.451083] env[62066]: DEBUG oslo_vmware.api [None req-3d67ea1c-5f75-4620-b677-6f1e0faf1bd6 tempest-ServersV294TestFqdnHostnames-843300434 tempest-ServersV294TestFqdnHostnames-843300434-project-member] Task: {'id': task-1340847, 'name': ReconfigVM_Task, 'duration_secs': 0.407651} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.451379] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-3d67ea1c-5f75-4620-b677-6f1e0faf1bd6 tempest-ServersV294TestFqdnHostnames-843300434 tempest-ServersV294TestFqdnHostnames-843300434-project-member] [instance: 2709d0ad-5dd8-4e3c-b1e6-3d1e2e242380] Reconfigured VM instance instance-00000046 to attach disk [datastore2] 2709d0ad-5dd8-4e3c-b1e6-3d1e2e242380/2709d0ad-5dd8-4e3c-b1e6-3d1e2e242380.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 826.452050] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1ce1a807-dd2d-4835-965e-4aa9d78e066f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.458805] env[62066]: DEBUG oslo_vmware.api [None req-3d67ea1c-5f75-4620-b677-6f1e0faf1bd6 tempest-ServersV294TestFqdnHostnames-843300434 tempest-ServersV294TestFqdnHostnames-843300434-project-member] Waiting for the task: (returnval){ [ 826.458805] env[62066]: value = "task-1340848" [ 826.458805] env[62066]: _type = "Task" [ 826.458805] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.471287] env[62066]: DEBUG oslo_vmware.api [None req-3d67ea1c-5f75-4620-b677-6f1e0faf1bd6 tempest-ServersV294TestFqdnHostnames-843300434 tempest-ServersV294TestFqdnHostnames-843300434-project-member] Task: {'id': task-1340848, 'name': Rename_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.477269] env[62066]: DEBUG oslo_vmware.api [None req-b8836db2-293e-442c-80f2-42d8bc4c2918 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]5299027a-2e08-6ce0-6410-bdca5e7c038b, 'name': SearchDatastore_Task, 'duration_secs': 0.01477} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.478128] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d906b59a-4557-4e8a-a160-a804b6f1f646 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.484192] env[62066]: DEBUG oslo_vmware.api [None req-b8836db2-293e-442c-80f2-42d8bc4c2918 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Waiting for the task: (returnval){ [ 826.484192] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]529ada2a-d27d-38bb-5a0c-75f261c6a642" [ 826.484192] env[62066]: _type = "Task" [ 826.484192] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.494333] env[62066]: DEBUG oslo_vmware.api [None req-b8836db2-293e-442c-80f2-42d8bc4c2918 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]529ada2a-d27d-38bb-5a0c-75f261c6a642, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.532263] env[62066]: DEBUG nova.network.neutron [None req-d2edd17d-ac3d-4b81-90f1-278c62ed416c tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] [instance: de2e2be7-efdb-45a8-842a-640ab9deb1d9] Updating instance_info_cache with network_info: [{"id": "c833a993-4af0-4bdb-90d2-cc82a796a560", "address": "fa:16:3e:71:ab:fe", "network": {"id": "4114ddfe-818f-47a5-bb6f-44a4c22517f6", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-676131672-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dd30584efc254610b3939d1ba3806693", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a485857d-7086-4dcf-9d65-d0dcd177fcb0", "external-id": "nsx-vlan-transportzone-232", "segmentation_id": 232, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc833a993-4a", "ovs_interfaceid": "c833a993-4af0-4bdb-90d2-cc82a796a560", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 826.608431] env[62066]: INFO nova.compute.manager [-] [instance: 050040cb-6ee3-4ad2-960a-fcebb53ac394] Took 1.86 seconds to deallocate network for instance. [ 826.646262] env[62066]: DEBUG nova.network.neutron [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] [instance: 1a4b0637-1a56-41ef-b89b-6b56d24ed206] Updating instance_info_cache with network_info: [{"id": "84936b80-0721-4235-a92a-7bc4adf450d8", "address": "fa:16:3e:f7:5a:62", "network": {"id": "518e5084-b25d-4931-9319-a62810398063", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.233", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "eb52f7069a374c61ae946f052007c6d9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c6365036-aa37-44d2-90d1-ca1c3516ded9", "external-id": "nsx-vlan-transportzone-66", "segmentation_id": 66, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap84936b80-07", "ovs_interfaceid": "84936b80-0721-4235-a92a-7bc4adf450d8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 826.694402] env[62066]: DEBUG nova.scheduler.client.report [None req-02667002-03be-4979-8517-c36490492214 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 826.796524] env[62066]: DEBUG nova.compute.manager [None req-acef35f4-b906-4efb-ae97-e8cdc80e8e5a tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] [instance: 5e96def2-0cbd-4bd9-93f4-6a365a0142b0] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 826.811162] env[62066]: DEBUG oslo_vmware.api [None req-9e052131-3359-4888-9d78-89440b639dcc tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Task: {'id': task-1340846, 'name': PowerOnVM_Task, 'duration_secs': 0.540171} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.811456] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-9e052131-3359-4888-9d78-89440b639dcc tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] [instance: 8ff6fbb9-c90f-498d-9a85-d220a8c2f794] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 826.811699] env[62066]: INFO nova.compute.manager [None req-9e052131-3359-4888-9d78-89440b639dcc tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] [instance: 8ff6fbb9-c90f-498d-9a85-d220a8c2f794] Took 9.68 seconds to spawn the instance on the hypervisor. [ 826.811888] env[62066]: DEBUG nova.compute.manager [None req-9e052131-3359-4888-9d78-89440b639dcc tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] [instance: 8ff6fbb9-c90f-498d-9a85-d220a8c2f794] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 826.812732] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-357032ac-645e-44dc-ba5c-30177aa80c0f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.969135] env[62066]: DEBUG oslo_vmware.api [None req-3d67ea1c-5f75-4620-b677-6f1e0faf1bd6 tempest-ServersV294TestFqdnHostnames-843300434 tempest-ServersV294TestFqdnHostnames-843300434-project-member] Task: {'id': task-1340848, 'name': Rename_Task, 'duration_secs': 0.181389} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.969453] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d67ea1c-5f75-4620-b677-6f1e0faf1bd6 tempest-ServersV294TestFqdnHostnames-843300434 tempest-ServersV294TestFqdnHostnames-843300434-project-member] [instance: 2709d0ad-5dd8-4e3c-b1e6-3d1e2e242380] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 826.969729] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-32688666-4cf0-44ea-815d-66f34aac17a9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.976383] env[62066]: DEBUG oslo_vmware.api [None req-3d67ea1c-5f75-4620-b677-6f1e0faf1bd6 tempest-ServersV294TestFqdnHostnames-843300434 tempest-ServersV294TestFqdnHostnames-843300434-project-member] Waiting for the task: (returnval){ [ 826.976383] env[62066]: value = "task-1340849" [ 826.976383] env[62066]: _type = "Task" [ 826.976383] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.984407] env[62066]: DEBUG oslo_vmware.api [None req-3d67ea1c-5f75-4620-b677-6f1e0faf1bd6 tempest-ServersV294TestFqdnHostnames-843300434 tempest-ServersV294TestFqdnHostnames-843300434-project-member] Task: {'id': task-1340849, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.993334] env[62066]: DEBUG oslo_vmware.api [None req-b8836db2-293e-442c-80f2-42d8bc4c2918 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]529ada2a-d27d-38bb-5a0c-75f261c6a642, 'name': SearchDatastore_Task, 'duration_secs': 0.014642} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.993609] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b8836db2-293e-442c-80f2-42d8bc4c2918 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Releasing lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 826.993883] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-b8836db2-293e-442c-80f2-42d8bc4c2918 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk to [datastore2] 88daaaa6-f385-4161-bc74-9fffabb1145d/88daaaa6-f385-4161-bc74-9fffabb1145d.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 826.994177] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b6f2d133-c9bd-45a5-9e26-be8996b510a9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.001254] env[62066]: DEBUG oslo_vmware.api [None req-b8836db2-293e-442c-80f2-42d8bc4c2918 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Waiting for the task: (returnval){ [ 827.001254] env[62066]: value = "task-1340850" [ 827.001254] env[62066]: _type = "Task" [ 827.001254] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.009968] env[62066]: DEBUG oslo_vmware.api [None req-b8836db2-293e-442c-80f2-42d8bc4c2918 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1340850, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.035804] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d2edd17d-ac3d-4b81-90f1-278c62ed416c tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Releasing lock "refresh_cache-de2e2be7-efdb-45a8-842a-640ab9deb1d9" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 827.036209] env[62066]: DEBUG nova.compute.manager [None req-d2edd17d-ac3d-4b81-90f1-278c62ed416c tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] [instance: de2e2be7-efdb-45a8-842a-640ab9deb1d9] Instance network_info: |[{"id": "c833a993-4af0-4bdb-90d2-cc82a796a560", "address": "fa:16:3e:71:ab:fe", "network": {"id": "4114ddfe-818f-47a5-bb6f-44a4c22517f6", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-676131672-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dd30584efc254610b3939d1ba3806693", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a485857d-7086-4dcf-9d65-d0dcd177fcb0", "external-id": "nsx-vlan-transportzone-232", "segmentation_id": 232, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc833a993-4a", "ovs_interfaceid": "c833a993-4af0-4bdb-90d2-cc82a796a560", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 827.115685] env[62066]: DEBUG oslo_concurrency.lockutils [None req-32db9ed7-c319-413e-978b-0abcfac2cab3 tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 827.149926] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Releasing lock "refresh_cache-1a4b0637-1a56-41ef-b89b-6b56d24ed206" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 827.150168] env[62066]: DEBUG nova.compute.manager [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] [instance: 1a4b0637-1a56-41ef-b89b-6b56d24ed206] Updated the network info_cache for instance {{(pid=62066) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 827.150396] env[62066]: DEBUG oslo_service.periodic_task [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 827.150578] env[62066]: DEBUG oslo_service.periodic_task [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 827.150722] env[62066]: DEBUG oslo_service.periodic_task [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 827.150913] env[62066]: DEBUG oslo_service.periodic_task [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 827.151100] env[62066]: DEBUG oslo_service.periodic_task [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 827.151256] env[62066]: DEBUG oslo_service.periodic_task [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 827.151393] env[62066]: DEBUG nova.compute.manager [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62066) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 827.151548] env[62066]: DEBUG oslo_service.periodic_task [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Running periodic task ComputeManager.update_available_resource {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 827.199876] env[62066]: DEBUG oslo_concurrency.lockutils [None req-02667002-03be-4979-8517-c36490492214 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.432s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 827.200435] env[62066]: DEBUG nova.compute.manager [None req-02667002-03be-4979-8517-c36490492214 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 86a92b61-f3e9-48f9-8ee2-756669d558ef] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 827.203137] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d041ede3-45cc-4fa6-b4cc-7d603e4097a9 tempest-ServerMetadataNegativeTestJSON-1549470887 tempest-ServerMetadataNegativeTestJSON-1549470887-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 14.804s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 827.203299] env[62066]: DEBUG nova.objects.instance [None req-d041ede3-45cc-4fa6-b4cc-7d603e4097a9 tempest-ServerMetadataNegativeTestJSON-1549470887 tempest-ServerMetadataNegativeTestJSON-1549470887-project-member] Lazy-loading 'resources' on Instance uuid 31d47299-83a3-4f27-aeb7-95c4cd36c5ac {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 827.333723] env[62066]: INFO nova.compute.manager [None req-9e052131-3359-4888-9d78-89440b639dcc tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] [instance: 8ff6fbb9-c90f-498d-9a85-d220a8c2f794] Took 31.26 seconds to build instance. [ 827.398675] env[62066]: DEBUG nova.network.neutron [None req-acef35f4-b906-4efb-ae97-e8cdc80e8e5a tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] [instance: 5e96def2-0cbd-4bd9-93f4-6a365a0142b0] Successfully updated port: 0932c7ac-630e-4cef-8991-45b438f39483 {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 827.487520] env[62066]: DEBUG oslo_vmware.api [None req-3d67ea1c-5f75-4620-b677-6f1e0faf1bd6 tempest-ServersV294TestFqdnHostnames-843300434 tempest-ServersV294TestFqdnHostnames-843300434-project-member] Task: {'id': task-1340849, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.511285] env[62066]: DEBUG oslo_vmware.api [None req-b8836db2-293e-442c-80f2-42d8bc4c2918 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1340850, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.655222] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 827.712950] env[62066]: DEBUG nova.compute.utils [None req-02667002-03be-4979-8517-c36490492214 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 827.714979] env[62066]: DEBUG nova.compute.manager [None req-02667002-03be-4979-8517-c36490492214 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 86a92b61-f3e9-48f9-8ee2-756669d558ef] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 827.715402] env[62066]: DEBUG nova.network.neutron [None req-02667002-03be-4979-8517-c36490492214 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 86a92b61-f3e9-48f9-8ee2-756669d558ef] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 827.760594] env[62066]: DEBUG nova.policy [None req-02667002-03be-4979-8517-c36490492214 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8910a229218b4ec5ad72c893badfc598', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6497ab02f327476d8ff81c2ecc0371e5', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 827.838309] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9e052131-3359-4888-9d78-89440b639dcc tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Lock "8ff6fbb9-c90f-498d-9a85-d220a8c2f794" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 102.360s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 827.904351] env[62066]: DEBUG oslo_concurrency.lockutils [None req-acef35f4-b906-4efb-ae97-e8cdc80e8e5a tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Acquiring lock "refresh_cache-5e96def2-0cbd-4bd9-93f4-6a365a0142b0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 827.904482] env[62066]: DEBUG oslo_concurrency.lockutils [None req-acef35f4-b906-4efb-ae97-e8cdc80e8e5a tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Acquired lock "refresh_cache-5e96def2-0cbd-4bd9-93f4-6a365a0142b0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 827.904634] env[62066]: DEBUG nova.network.neutron [None req-acef35f4-b906-4efb-ae97-e8cdc80e8e5a tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] [instance: 5e96def2-0cbd-4bd9-93f4-6a365a0142b0] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 827.958657] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d427bb63-d8d3-4845-9315-ba6334536fbc {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.967266] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ec83ffd-5da0-40fe-9978-b0637276c3e1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.002921] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-742acedc-c135-40b9-a486-1860fce9362a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.019793] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5be3b4ea-76b3-46be-b41d-4dfdfe050ba2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.024319] env[62066]: DEBUG oslo_vmware.api [None req-b8836db2-293e-442c-80f2-42d8bc4c2918 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1340850, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.640316} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.024679] env[62066]: DEBUG oslo_vmware.api [None req-3d67ea1c-5f75-4620-b677-6f1e0faf1bd6 tempest-ServersV294TestFqdnHostnames-843300434 tempest-ServersV294TestFqdnHostnames-843300434-project-member] Task: {'id': task-1340849, 'name': PowerOnVM_Task, 'duration_secs': 0.989618} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.024772] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-b8836db2-293e-442c-80f2-42d8bc4c2918 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk to [datastore2] 88daaaa6-f385-4161-bc74-9fffabb1145d/88daaaa6-f385-4161-bc74-9fffabb1145d.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 828.025016] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-b8836db2-293e-442c-80f2-42d8bc4c2918 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 88daaaa6-f385-4161-bc74-9fffabb1145d] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 828.025315] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d67ea1c-5f75-4620-b677-6f1e0faf1bd6 tempest-ServersV294TestFqdnHostnames-843300434 tempest-ServersV294TestFqdnHostnames-843300434-project-member] [instance: 2709d0ad-5dd8-4e3c-b1e6-3d1e2e242380] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 828.025585] env[62066]: INFO nova.compute.manager [None req-3d67ea1c-5f75-4620-b677-6f1e0faf1bd6 tempest-ServersV294TestFqdnHostnames-843300434 tempest-ServersV294TestFqdnHostnames-843300434-project-member] [instance: 2709d0ad-5dd8-4e3c-b1e6-3d1e2e242380] Took 8.50 seconds to spawn the instance on the hypervisor. [ 828.025793] env[62066]: DEBUG nova.compute.manager [None req-3d67ea1c-5f75-4620-b677-6f1e0faf1bd6 tempest-ServersV294TestFqdnHostnames-843300434 tempest-ServersV294TestFqdnHostnames-843300434-project-member] [instance: 2709d0ad-5dd8-4e3c-b1e6-3d1e2e242380] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 828.026455] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e31a0bd2-abbd-4142-9c45-3f7b4827af08 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.028790] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74c4e6ea-2a35-4a4c-89c2-7d5a58bdd490 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.039515] env[62066]: DEBUG nova.compute.provider_tree [None req-d041ede3-45cc-4fa6-b4cc-7d603e4097a9 tempest-ServerMetadataNegativeTestJSON-1549470887 tempest-ServerMetadataNegativeTestJSON-1549470887-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 828.042938] env[62066]: DEBUG nova.network.neutron [None req-02667002-03be-4979-8517-c36490492214 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 86a92b61-f3e9-48f9-8ee2-756669d558ef] Successfully created port: fd69e6c7-721d-4947-80e1-4db6e9a230b0 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 828.052348] env[62066]: DEBUG oslo_vmware.api [None req-b8836db2-293e-442c-80f2-42d8bc4c2918 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Waiting for the task: (returnval){ [ 828.052348] env[62066]: value = "task-1340851" [ 828.052348] env[62066]: _type = "Task" [ 828.052348] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.062649] env[62066]: DEBUG oslo_vmware.api [None req-b8836db2-293e-442c-80f2-42d8bc4c2918 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1340851, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.215533] env[62066]: DEBUG nova.compute.manager [None req-02667002-03be-4979-8517-c36490492214 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 86a92b61-f3e9-48f9-8ee2-756669d558ef] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 828.438157] env[62066]: DEBUG nova.network.neutron [None req-acef35f4-b906-4efb-ae97-e8cdc80e8e5a tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] [instance: 5e96def2-0cbd-4bd9-93f4-6a365a0142b0] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 828.546084] env[62066]: DEBUG nova.scheduler.client.report [None req-d041ede3-45cc-4fa6-b4cc-7d603e4097a9 tempest-ServerMetadataNegativeTestJSON-1549470887 tempest-ServerMetadataNegativeTestJSON-1549470887-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 828.553186] env[62066]: DEBUG nova.network.neutron [None req-acef35f4-b906-4efb-ae97-e8cdc80e8e5a tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] [instance: 5e96def2-0cbd-4bd9-93f4-6a365a0142b0] Updating instance_info_cache with network_info: [{"id": "0932c7ac-630e-4cef-8991-45b438f39483", "address": "fa:16:3e:ca:08:8d", "network": {"id": "fac1b86c-746c-4b6e-8d7f-21fe6ed404a3", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1827773279-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "59855b1cf91d43e3ba0482f414c4e546", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "88651df2-0506-4f6c-b868-dd30a81f2b1c", "external-id": "nsx-vlan-transportzone-366", "segmentation_id": 366, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0932c7ac-63", "ovs_interfaceid": "0932c7ac-630e-4cef-8991-45b438f39483", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 828.566104] env[62066]: INFO nova.compute.manager [None req-3d67ea1c-5f75-4620-b677-6f1e0faf1bd6 tempest-ServersV294TestFqdnHostnames-843300434 tempest-ServersV294TestFqdnHostnames-843300434-project-member] [instance: 2709d0ad-5dd8-4e3c-b1e6-3d1e2e242380] Took 28.63 seconds to build instance. [ 828.572615] env[62066]: DEBUG oslo_vmware.api [None req-b8836db2-293e-442c-80f2-42d8bc4c2918 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1340851, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.07284} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.572896] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-b8836db2-293e-442c-80f2-42d8bc4c2918 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 88daaaa6-f385-4161-bc74-9fffabb1145d] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 828.573706] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1bcf7c5-af62-42b1-8a66-ca1d74b986a1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.599084] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-b8836db2-293e-442c-80f2-42d8bc4c2918 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 88daaaa6-f385-4161-bc74-9fffabb1145d] Reconfiguring VM instance instance-00000047 to attach disk [datastore2] 88daaaa6-f385-4161-bc74-9fffabb1145d/88daaaa6-f385-4161-bc74-9fffabb1145d.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 828.599772] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-66e4cf5d-924c-4c54-8499-ad3d1137debb {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.620757] env[62066]: DEBUG oslo_vmware.api [None req-b8836db2-293e-442c-80f2-42d8bc4c2918 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Waiting for the task: (returnval){ [ 828.620757] env[62066]: value = "task-1340852" [ 828.620757] env[62066]: _type = "Task" [ 828.620757] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.630112] env[62066]: DEBUG oslo_vmware.api [None req-b8836db2-293e-442c-80f2-42d8bc4c2918 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1340852, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.051308] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d041ede3-45cc-4fa6-b4cc-7d603e4097a9 tempest-ServerMetadataNegativeTestJSON-1549470887 tempest-ServerMetadataNegativeTestJSON-1549470887-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.848s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 829.054228] env[62066]: DEBUG oslo_concurrency.lockutils [None req-423c892e-6435-422f-ab41-7d39bb22efa3 tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 7.127s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 829.054530] env[62066]: DEBUG nova.objects.instance [None req-423c892e-6435-422f-ab41-7d39bb22efa3 tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] Lazy-loading 'resources' on Instance uuid 738ebb2f-7cc6-4d0f-871c-1428c9c21384 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 829.066294] env[62066]: DEBUG oslo_concurrency.lockutils [None req-acef35f4-b906-4efb-ae97-e8cdc80e8e5a tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Releasing lock "refresh_cache-5e96def2-0cbd-4bd9-93f4-6a365a0142b0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 829.066407] env[62066]: DEBUG nova.compute.manager [None req-acef35f4-b906-4efb-ae97-e8cdc80e8e5a tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] [instance: 5e96def2-0cbd-4bd9-93f4-6a365a0142b0] Instance network_info: |[{"id": "0932c7ac-630e-4cef-8991-45b438f39483", "address": "fa:16:3e:ca:08:8d", "network": {"id": "fac1b86c-746c-4b6e-8d7f-21fe6ed404a3", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1827773279-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "59855b1cf91d43e3ba0482f414c4e546", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "88651df2-0506-4f6c-b868-dd30a81f2b1c", "external-id": "nsx-vlan-transportzone-366", "segmentation_id": 366, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0932c7ac-63", "ovs_interfaceid": "0932c7ac-630e-4cef-8991-45b438f39483", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 829.068197] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3d67ea1c-5f75-4620-b677-6f1e0faf1bd6 tempest-ServersV294TestFqdnHostnames-843300434 tempest-ServersV294TestFqdnHostnames-843300434-project-member] Lock "2709d0ad-5dd8-4e3c-b1e6-3d1e2e242380" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 88.638s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 829.080418] env[62066]: INFO nova.scheduler.client.report [None req-d041ede3-45cc-4fa6-b4cc-7d603e4097a9 tempest-ServerMetadataNegativeTestJSON-1549470887 tempest-ServerMetadataNegativeTestJSON-1549470887-project-member] Deleted allocations for instance 31d47299-83a3-4f27-aeb7-95c4cd36c5ac [ 829.133798] env[62066]: DEBUG oslo_vmware.api [None req-b8836db2-293e-442c-80f2-42d8bc4c2918 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1340852, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.178999] env[62066]: DEBUG nova.virt.hardware [None req-d2edd17d-ac3d-4b81-90f1-278c62ed416c tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-23T13:40:41Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-23T13:40:23Z,direct_url=,disk_format='vmdk',id=50ff584c-3b50-4395-af07-3e66769bc9f7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='eb52f7069a374c61ae946f052007c6d9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-23T13:40:24Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 829.178999] env[62066]: DEBUG nova.virt.hardware [None req-d2edd17d-ac3d-4b81-90f1-278c62ed416c tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 829.178999] env[62066]: DEBUG nova.virt.hardware [None req-d2edd17d-ac3d-4b81-90f1-278c62ed416c tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 829.178999] env[62066]: DEBUG nova.virt.hardware [None req-d2edd17d-ac3d-4b81-90f1-278c62ed416c tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 829.178999] env[62066]: DEBUG nova.virt.hardware [None req-d2edd17d-ac3d-4b81-90f1-278c62ed416c tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 829.178999] env[62066]: DEBUG nova.virt.hardware [None req-d2edd17d-ac3d-4b81-90f1-278c62ed416c tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 829.178999] env[62066]: DEBUG nova.virt.hardware [None req-d2edd17d-ac3d-4b81-90f1-278c62ed416c tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 829.178999] env[62066]: DEBUG nova.virt.hardware [None req-d2edd17d-ac3d-4b81-90f1-278c62ed416c tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 829.178999] env[62066]: DEBUG nova.virt.hardware [None req-d2edd17d-ac3d-4b81-90f1-278c62ed416c tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 829.178999] env[62066]: DEBUG nova.virt.hardware [None req-d2edd17d-ac3d-4b81-90f1-278c62ed416c tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 829.178999] env[62066]: DEBUG nova.virt.hardware [None req-d2edd17d-ac3d-4b81-90f1-278c62ed416c tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 829.180956] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37def6d5-c72f-4596-a2f1-ab8c82a24c11 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.191358] env[62066]: DEBUG nova.virt.hardware [None req-acef35f4-b906-4efb-ae97-e8cdc80e8e5a tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-23T13:40:41Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-23T13:40:23Z,direct_url=,disk_format='vmdk',id=50ff584c-3b50-4395-af07-3e66769bc9f7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='eb52f7069a374c61ae946f052007c6d9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-23T13:40:24Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 829.191637] env[62066]: DEBUG nova.virt.hardware [None req-acef35f4-b906-4efb-ae97-e8cdc80e8e5a tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 829.191855] env[62066]: DEBUG nova.virt.hardware [None req-acef35f4-b906-4efb-ae97-e8cdc80e8e5a tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 829.192100] env[62066]: DEBUG nova.virt.hardware [None req-acef35f4-b906-4efb-ae97-e8cdc80e8e5a tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 829.193039] env[62066]: DEBUG nova.virt.hardware [None req-acef35f4-b906-4efb-ae97-e8cdc80e8e5a tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 829.193039] env[62066]: DEBUG nova.virt.hardware [None req-acef35f4-b906-4efb-ae97-e8cdc80e8e5a tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 829.193039] env[62066]: DEBUG nova.virt.hardware [None req-acef35f4-b906-4efb-ae97-e8cdc80e8e5a tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 829.193039] env[62066]: DEBUG nova.virt.hardware [None req-acef35f4-b906-4efb-ae97-e8cdc80e8e5a tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 829.193231] env[62066]: DEBUG nova.virt.hardware [None req-acef35f4-b906-4efb-ae97-e8cdc80e8e5a tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 829.193357] env[62066]: DEBUG nova.virt.hardware [None req-acef35f4-b906-4efb-ae97-e8cdc80e8e5a tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 829.194031] env[62066]: DEBUG nova.virt.hardware [None req-acef35f4-b906-4efb-ae97-e8cdc80e8e5a tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 829.195049] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8c7749a-9e99-4b5f-85d2-37ee3b4f86a4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.201913] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95fcaa9f-3e02-43e2-8a6b-c7c00828b8f0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.211826] env[62066]: DEBUG oslo_vmware.rw_handles [None req-8d3cb2ee-be58-4e07-b64f-9d5c34ab9462 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5227be83-0a35-4646-6e67-879db120d0e5/disk-0.vmdk. {{(pid=62066) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 829.222742] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acb93a35-9dbf-4256-a025-0b9350144161 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.227976] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28adab48-c578-4468-ab1b-4b3c0a8e80df {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.230090] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-d2edd17d-ac3d-4b81-90f1-278c62ed416c tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] [instance: de2e2be7-efdb-45a8-842a-640ab9deb1d9] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:71:ab:fe', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a485857d-7086-4dcf-9d65-d0dcd177fcb0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c833a993-4af0-4bdb-90d2-cc82a796a560', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 829.238262] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-d2edd17d-ac3d-4b81-90f1-278c62ed416c tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Creating folder: Project (dd30584efc254610b3939d1ba3806693). Parent ref: group-v285980. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 829.239669] env[62066]: DEBUG nova.compute.manager [None req-02667002-03be-4979-8517-c36490492214 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 86a92b61-f3e9-48f9-8ee2-756669d558ef] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 829.241791] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1d83ccad-d3aa-42a0-aced-334d183febfc {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.256784] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-acef35f4-b906-4efb-ae97-e8cdc80e8e5a tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] [instance: 5e96def2-0cbd-4bd9-93f4-6a365a0142b0] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ca:08:8d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '88651df2-0506-4f6c-b868-dd30a81f2b1c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0932c7ac-630e-4cef-8991-45b438f39483', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 829.264247] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-acef35f4-b906-4efb-ae97-e8cdc80e8e5a tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Creating folder: Project (59855b1cf91d43e3ba0482f414c4e546). Parent ref: group-v285980. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 829.268909] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-94b24347-f98d-4c83-8547-96b6ae68a68e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.270547] env[62066]: DEBUG oslo_vmware.rw_handles [None req-8d3cb2ee-be58-4e07-b64f-9d5c34ab9462 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5227be83-0a35-4646-6e67-879db120d0e5/disk-0.vmdk is in state: ready. {{(pid=62066) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 829.270712] env[62066]: ERROR oslo_vmware.rw_handles [None req-8d3cb2ee-be58-4e07-b64f-9d5c34ab9462 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5227be83-0a35-4646-6e67-879db120d0e5/disk-0.vmdk due to incomplete transfer. [ 829.270967] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-d2edd17d-ac3d-4b81-90f1-278c62ed416c tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Created folder: Project (dd30584efc254610b3939d1ba3806693) in parent group-v285980. [ 829.271147] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-d2edd17d-ac3d-4b81-90f1-278c62ed416c tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Creating folder: Instances. Parent ref: group-v286053. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 829.271878] env[62066]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-d963269c-e21c-4a8c-b451-a54bae426b83 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.273294] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-671e596c-d2cc-4470-a6f8-8be41d3e37ad {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.278851] env[62066]: DEBUG nova.virt.hardware [None req-02667002-03be-4979-8517-c36490492214 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-23T13:40:41Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-23T13:40:23Z,direct_url=,disk_format='vmdk',id=50ff584c-3b50-4395-af07-3e66769bc9f7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='eb52f7069a374c61ae946f052007c6d9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-23T13:40:24Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 829.279097] env[62066]: DEBUG nova.virt.hardware [None req-02667002-03be-4979-8517-c36490492214 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 829.279261] env[62066]: DEBUG nova.virt.hardware [None req-02667002-03be-4979-8517-c36490492214 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 829.279457] env[62066]: DEBUG nova.virt.hardware [None req-02667002-03be-4979-8517-c36490492214 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 829.279604] env[62066]: DEBUG nova.virt.hardware [None req-02667002-03be-4979-8517-c36490492214 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 829.279753] env[62066]: DEBUG nova.virt.hardware [None req-02667002-03be-4979-8517-c36490492214 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 829.279975] env[62066]: DEBUG nova.virt.hardware [None req-02667002-03be-4979-8517-c36490492214 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 829.280159] env[62066]: DEBUG nova.virt.hardware [None req-02667002-03be-4979-8517-c36490492214 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 829.280330] env[62066]: DEBUG nova.virt.hardware [None req-02667002-03be-4979-8517-c36490492214 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 829.280494] env[62066]: DEBUG nova.virt.hardware [None req-02667002-03be-4979-8517-c36490492214 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 829.280683] env[62066]: DEBUG nova.virt.hardware [None req-02667002-03be-4979-8517-c36490492214 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 829.282731] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c28cfb18-3fd2-4c80-9fe5-ec36da9f698a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.287531] env[62066]: DEBUG oslo_vmware.rw_handles [None req-8d3cb2ee-be58-4e07-b64f-9d5c34ab9462 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5227be83-0a35-4646-6e67-879db120d0e5/disk-0.vmdk. {{(pid=62066) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 829.287723] env[62066]: DEBUG nova.virt.vmwareapi.images [None req-8d3cb2ee-be58-4e07-b64f-9d5c34ab9462 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 49f46244-34fa-48a1-95a2-8e95850f345d] Uploaded image 843e9ee8-31ee-40e2-9f6f-50ffd87c6796 to the Glance image server {{(pid=62066) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 829.289981] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d3cb2ee-be58-4e07-b64f-9d5c34ab9462 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 49f46244-34fa-48a1-95a2-8e95850f345d] Destroying the VM {{(pid=62066) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 829.290280] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-d2edd17d-ac3d-4b81-90f1-278c62ed416c tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Created folder: Instances in parent group-v286053. [ 829.290497] env[62066]: DEBUG oslo.service.loopingcall [None req-d2edd17d-ac3d-4b81-90f1-278c62ed416c tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 829.290687] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-acef35f4-b906-4efb-ae97-e8cdc80e8e5a tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Created folder: Project (59855b1cf91d43e3ba0482f414c4e546) in parent group-v285980. [ 829.290841] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-acef35f4-b906-4efb-ae97-e8cdc80e8e5a tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Creating folder: Instances. Parent ref: group-v286054. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 829.292085] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-51d6c25f-1418-4c07-bdeb-42cad80297e1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.293647] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: de2e2be7-efdb-45a8-842a-640ab9deb1d9] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 829.293838] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-676280b0-9a9a-4a37-86ad-c47ca472f85c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.298017] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b3de658d-833e-4025-8148-2077778a16f5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.317215] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d86cb63-5024-4e29-af54-fdf1d853b5ee {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.321137] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 829.321137] env[62066]: value = "task-1340857" [ 829.321137] env[62066]: _type = "Task" [ 829.321137] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.323327] env[62066]: DEBUG oslo_vmware.api [None req-8d3cb2ee-be58-4e07-b64f-9d5c34ab9462 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Waiting for the task: (returnval){ [ 829.323327] env[62066]: value = "task-1340858" [ 829.323327] env[62066]: _type = "Task" [ 829.323327] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.323567] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-acef35f4-b906-4efb-ae97-e8cdc80e8e5a tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Created folder: Instances in parent group-v286054. [ 829.325051] env[62066]: DEBUG oslo.service.loopingcall [None req-acef35f4-b906-4efb-ae97-e8cdc80e8e5a tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 829.327119] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5e96def2-0cbd-4bd9-93f4-6a365a0142b0] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 829.338899] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9aa759f1-d283-49e9-ba0e-e59f26132e04 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.358754] env[62066]: DEBUG oslo_vmware.api [None req-8d3cb2ee-be58-4e07-b64f-9d5c34ab9462 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': task-1340858, 'name': Destroy_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.363402] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1340857, 'name': CreateVM_Task} progress is 15%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.363620] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 829.363620] env[62066]: value = "task-1340859" [ 829.363620] env[62066]: _type = "Task" [ 829.363620] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.372661] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1340859, 'name': CreateVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.563465] env[62066]: DEBUG nova.network.neutron [None req-02667002-03be-4979-8517-c36490492214 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 86a92b61-f3e9-48f9-8ee2-756669d558ef] Successfully updated port: fd69e6c7-721d-4947-80e1-4db6e9a230b0 {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 829.589363] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d041ede3-45cc-4fa6-b4cc-7d603e4097a9 tempest-ServerMetadataNegativeTestJSON-1549470887 tempest-ServerMetadataNegativeTestJSON-1549470887-project-member] Lock "31d47299-83a3-4f27-aeb7-95c4cd36c5ac" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 20.297s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 829.635760] env[62066]: DEBUG oslo_vmware.api [None req-b8836db2-293e-442c-80f2-42d8bc4c2918 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1340852, 'name': ReconfigVM_Task, 'duration_secs': 0.552343} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.636163] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-b8836db2-293e-442c-80f2-42d8bc4c2918 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 88daaaa6-f385-4161-bc74-9fffabb1145d] Reconfigured VM instance instance-00000047 to attach disk [datastore2] 88daaaa6-f385-4161-bc74-9fffabb1145d/88daaaa6-f385-4161-bc74-9fffabb1145d.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 829.636869] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b53cc8cc-ac75-4349-ab72-a6657de140af {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.645138] env[62066]: DEBUG oslo_vmware.api [None req-b8836db2-293e-442c-80f2-42d8bc4c2918 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Waiting for the task: (returnval){ [ 829.645138] env[62066]: value = "task-1340860" [ 829.645138] env[62066]: _type = "Task" [ 829.645138] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.653871] env[62066]: DEBUG oslo_vmware.api [None req-b8836db2-293e-442c-80f2-42d8bc4c2918 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1340860, 'name': Rename_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.852791] env[62066]: DEBUG oslo_vmware.api [None req-8d3cb2ee-be58-4e07-b64f-9d5c34ab9462 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': task-1340858, 'name': Destroy_Task} progress is 33%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.854635] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1340857, 'name': CreateVM_Task, 'duration_secs': 0.473071} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.855950] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: de2e2be7-efdb-45a8-842a-640ab9deb1d9] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 829.856834] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8c54a21-b293-4400-a8e4-bba1016f675d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.860428] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d2edd17d-ac3d-4b81-90f1-278c62ed416c tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 829.860638] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d2edd17d-ac3d-4b81-90f1-278c62ed416c tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Acquired lock "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 829.860966] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d2edd17d-ac3d-4b81-90f1-278c62ed416c tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 829.861277] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0e75ede6-428d-418f-a74a-c347e226d7e2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.871912] env[62066]: DEBUG oslo_vmware.api [None req-d2edd17d-ac3d-4b81-90f1-278c62ed416c tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Waiting for the task: (returnval){ [ 829.871912] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]52816777-cacf-39a6-2863-79eea77ec6d3" [ 829.871912] env[62066]: _type = "Task" [ 829.871912] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.873270] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cebbb951-7249-4f85-bf28-b4de1b47da42 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.883951] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1340859, 'name': CreateVM_Task} progress is 99%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.889405] env[62066]: DEBUG oslo_vmware.api [None req-d2edd17d-ac3d-4b81-90f1-278c62ed416c tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52816777-cacf-39a6-2863-79eea77ec6d3, 'name': SearchDatastore_Task, 'duration_secs': 0.01666} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.916061] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d2edd17d-ac3d-4b81-90f1-278c62ed416c tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Releasing lock "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 829.916444] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-d2edd17d-ac3d-4b81-90f1-278c62ed416c tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] [instance: de2e2be7-efdb-45a8-842a-640ab9deb1d9] Processing image 50ff584c-3b50-4395-af07-3e66769bc9f7 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 829.916705] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d2edd17d-ac3d-4b81-90f1-278c62ed416c tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 829.916959] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d2edd17d-ac3d-4b81-90f1-278c62ed416c tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Acquired lock "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 829.917051] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-d2edd17d-ac3d-4b81-90f1-278c62ed416c tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 829.917630] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6df9ad90-d445-46ea-b7ab-31e59a22f84f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.920078] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae4fccfc-c2ae-4df7-bd7c-54d5a75584c5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.927872] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7e1b53c-5c94-4fee-b891-266650673f02 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.932696] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-d2edd17d-ac3d-4b81-90f1-278c62ed416c tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 829.932891] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-d2edd17d-ac3d-4b81-90f1-278c62ed416c tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 829.933935] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bfb7a735-e52a-4f50-ad09-43fee74ef3c2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.943868] env[62066]: DEBUG nova.compute.provider_tree [None req-423c892e-6435-422f-ab41-7d39bb22efa3 tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 829.948472] env[62066]: DEBUG oslo_vmware.api [None req-d2edd17d-ac3d-4b81-90f1-278c62ed416c tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Waiting for the task: (returnval){ [ 829.948472] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]527b6a67-6f32-274a-2957-909b55fc1239" [ 829.948472] env[62066]: _type = "Task" [ 829.948472] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.956808] env[62066]: DEBUG oslo_vmware.api [None req-d2edd17d-ac3d-4b81-90f1-278c62ed416c tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]527b6a67-6f32-274a-2957-909b55fc1239, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.001733] env[62066]: DEBUG nova.compute.manager [req-732fa2fe-08c1-4864-9047-5c8e399a2436 req-aaa9c9a8-f091-46ef-ab4a-61dac97c1614 service nova] [instance: 88daaaa6-f385-4161-bc74-9fffabb1145d] Received event network-changed-ba002f90-5218-46f2-b7e1-cd96ae3018c7 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 830.001906] env[62066]: DEBUG nova.compute.manager [req-732fa2fe-08c1-4864-9047-5c8e399a2436 req-aaa9c9a8-f091-46ef-ab4a-61dac97c1614 service nova] [instance: 88daaaa6-f385-4161-bc74-9fffabb1145d] Refreshing instance network info cache due to event network-changed-ba002f90-5218-46f2-b7e1-cd96ae3018c7. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 830.002161] env[62066]: DEBUG oslo_concurrency.lockutils [req-732fa2fe-08c1-4864-9047-5c8e399a2436 req-aaa9c9a8-f091-46ef-ab4a-61dac97c1614 service nova] Acquiring lock "refresh_cache-88daaaa6-f385-4161-bc74-9fffabb1145d" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 830.002454] env[62066]: DEBUG oslo_concurrency.lockutils [req-732fa2fe-08c1-4864-9047-5c8e399a2436 req-aaa9c9a8-f091-46ef-ab4a-61dac97c1614 service nova] Acquired lock "refresh_cache-88daaaa6-f385-4161-bc74-9fffabb1145d" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 830.002555] env[62066]: DEBUG nova.network.neutron [req-732fa2fe-08c1-4864-9047-5c8e399a2436 req-aaa9c9a8-f091-46ef-ab4a-61dac97c1614 service nova] [instance: 88daaaa6-f385-4161-bc74-9fffabb1145d] Refreshing network info cache for port ba002f90-5218-46f2-b7e1-cd96ae3018c7 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 830.066442] env[62066]: DEBUG oslo_concurrency.lockutils [None req-02667002-03be-4979-8517-c36490492214 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Acquiring lock "refresh_cache-86a92b61-f3e9-48f9-8ee2-756669d558ef" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 830.066584] env[62066]: DEBUG oslo_concurrency.lockutils [None req-02667002-03be-4979-8517-c36490492214 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Acquired lock "refresh_cache-86a92b61-f3e9-48f9-8ee2-756669d558ef" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 830.066711] env[62066]: DEBUG nova.network.neutron [None req-02667002-03be-4979-8517-c36490492214 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 86a92b61-f3e9-48f9-8ee2-756669d558ef] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 830.160472] env[62066]: DEBUG oslo_vmware.api [None req-b8836db2-293e-442c-80f2-42d8bc4c2918 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1340860, 'name': Rename_Task, 'duration_secs': 0.222497} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.160813] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-b8836db2-293e-442c-80f2-42d8bc4c2918 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 88daaaa6-f385-4161-bc74-9fffabb1145d] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 830.161207] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9676e437-bc2f-43a6-a9e7-7a73be70b5f5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.169902] env[62066]: DEBUG oslo_vmware.api [None req-b8836db2-293e-442c-80f2-42d8bc4c2918 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Waiting for the task: (returnval){ [ 830.169902] env[62066]: value = "task-1340861" [ 830.169902] env[62066]: _type = "Task" [ 830.169902] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.180356] env[62066]: DEBUG oslo_vmware.api [None req-b8836db2-293e-442c-80f2-42d8bc4c2918 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1340861, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.342835] env[62066]: DEBUG oslo_vmware.api [None req-8d3cb2ee-be58-4e07-b64f-9d5c34ab9462 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': task-1340858, 'name': Destroy_Task, 'duration_secs': 0.675868} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.344171] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-8d3cb2ee-be58-4e07-b64f-9d5c34ab9462 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 49f46244-34fa-48a1-95a2-8e95850f345d] Destroyed the VM [ 830.344603] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-8d3cb2ee-be58-4e07-b64f-9d5c34ab9462 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 49f46244-34fa-48a1-95a2-8e95850f345d] Deleting Snapshot of the VM instance {{(pid=62066) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 830.345459] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-70977913-f279-4cc4-9fd1-67c9861ae765 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.356017] env[62066]: DEBUG oslo_vmware.api [None req-8d3cb2ee-be58-4e07-b64f-9d5c34ab9462 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Waiting for the task: (returnval){ [ 830.356017] env[62066]: value = "task-1340862" [ 830.356017] env[62066]: _type = "Task" [ 830.356017] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.368125] env[62066]: DEBUG oslo_vmware.api [None req-8d3cb2ee-be58-4e07-b64f-9d5c34ab9462 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': task-1340862, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.378184] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1340859, 'name': CreateVM_Task, 'duration_secs': 0.527283} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.378184] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5e96def2-0cbd-4bd9-93f4-6a365a0142b0] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 830.378184] env[62066]: DEBUG oslo_concurrency.lockutils [None req-acef35f4-b906-4efb-ae97-e8cdc80e8e5a tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 830.378184] env[62066]: DEBUG oslo_concurrency.lockutils [None req-acef35f4-b906-4efb-ae97-e8cdc80e8e5a tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Acquired lock "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 830.378184] env[62066]: DEBUG oslo_concurrency.lockutils [None req-acef35f4-b906-4efb-ae97-e8cdc80e8e5a tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 830.378184] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7fc73e4d-3391-42f0-aafb-962285e5387f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.385018] env[62066]: DEBUG oslo_vmware.api [None req-acef35f4-b906-4efb-ae97-e8cdc80e8e5a tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Waiting for the task: (returnval){ [ 830.385018] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]52f933bb-8ffe-d710-50a1-7663921e2966" [ 830.385018] env[62066]: _type = "Task" [ 830.385018] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.393525] env[62066]: DEBUG oslo_vmware.api [None req-acef35f4-b906-4efb-ae97-e8cdc80e8e5a tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52f933bb-8ffe-d710-50a1-7663921e2966, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.449896] env[62066]: DEBUG nova.scheduler.client.report [None req-423c892e-6435-422f-ab41-7d39bb22efa3 tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 830.463864] env[62066]: DEBUG oslo_vmware.api [None req-d2edd17d-ac3d-4b81-90f1-278c62ed416c tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]527b6a67-6f32-274a-2957-909b55fc1239, 'name': SearchDatastore_Task, 'duration_secs': 0.024366} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.465213] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ab140019-7330-47a7-8828-fcf5672598e6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.474039] env[62066]: DEBUG oslo_vmware.api [None req-d2edd17d-ac3d-4b81-90f1-278c62ed416c tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Waiting for the task: (returnval){ [ 830.474039] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]52b614d0-991c-bebd-136f-0f935350f5e1" [ 830.474039] env[62066]: _type = "Task" [ 830.474039] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.482766] env[62066]: DEBUG oslo_vmware.api [None req-d2edd17d-ac3d-4b81-90f1-278c62ed416c tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52b614d0-991c-bebd-136f-0f935350f5e1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.647852] env[62066]: DEBUG nova.network.neutron [None req-02667002-03be-4979-8517-c36490492214 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 86a92b61-f3e9-48f9-8ee2-756669d558ef] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 830.681392] env[62066]: DEBUG oslo_vmware.api [None req-b8836db2-293e-442c-80f2-42d8bc4c2918 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1340861, 'name': PowerOnVM_Task} progress is 1%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.866304] env[62066]: DEBUG oslo_vmware.api [None req-8d3cb2ee-be58-4e07-b64f-9d5c34ab9462 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': task-1340862, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.902155] env[62066]: DEBUG oslo_vmware.api [None req-acef35f4-b906-4efb-ae97-e8cdc80e8e5a tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52f933bb-8ffe-d710-50a1-7663921e2966, 'name': SearchDatastore_Task, 'duration_secs': 0.014004} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.902516] env[62066]: DEBUG oslo_concurrency.lockutils [None req-acef35f4-b906-4efb-ae97-e8cdc80e8e5a tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Releasing lock "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 830.902766] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-acef35f4-b906-4efb-ae97-e8cdc80e8e5a tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] [instance: 5e96def2-0cbd-4bd9-93f4-6a365a0142b0] Processing image 50ff584c-3b50-4395-af07-3e66769bc9f7 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 830.902981] env[62066]: DEBUG oslo_concurrency.lockutils [None req-acef35f4-b906-4efb-ae97-e8cdc80e8e5a tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 830.959089] env[62066]: DEBUG oslo_concurrency.lockutils [None req-423c892e-6435-422f-ab41-7d39bb22efa3 tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.904s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 830.961812] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1e760fa3-f701-40c8-8666-21a2eeae0751 tempest-ServerAddressesTestJSON-1053293248 tempest-ServerAddressesTestJSON-1053293248-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 7.869s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 830.962078] env[62066]: DEBUG nova.objects.instance [None req-1e760fa3-f701-40c8-8666-21a2eeae0751 tempest-ServerAddressesTestJSON-1053293248 tempest-ServerAddressesTestJSON-1053293248-project-member] Lazy-loading 'resources' on Instance uuid 49bc1350-0095-406b-bc68-005eb6b681a8 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 830.986762] env[62066]: DEBUG oslo_vmware.api [None req-d2edd17d-ac3d-4b81-90f1-278c62ed416c tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52b614d0-991c-bebd-136f-0f935350f5e1, 'name': SearchDatastore_Task, 'duration_secs': 0.011673} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.990659] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d2edd17d-ac3d-4b81-90f1-278c62ed416c tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Releasing lock "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 830.991031] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-d2edd17d-ac3d-4b81-90f1-278c62ed416c tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk to [datastore1] de2e2be7-efdb-45a8-842a-640ab9deb1d9/de2e2be7-efdb-45a8-842a-640ab9deb1d9.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 830.991645] env[62066]: DEBUG oslo_concurrency.lockutils [None req-acef35f4-b906-4efb-ae97-e8cdc80e8e5a tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Acquired lock "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 830.992297] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-acef35f4-b906-4efb-ae97-e8cdc80e8e5a tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 830.992468] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6438626c-53c0-4eda-9e4f-418487409e29 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.995482] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ec499c44-ca55-4b2f-86c8-79234bae7024 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.997994] env[62066]: INFO nova.scheduler.client.report [None req-423c892e-6435-422f-ab41-7d39bb22efa3 tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] Deleted allocations for instance 738ebb2f-7cc6-4d0f-871c-1428c9c21384 [ 831.009545] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-acef35f4-b906-4efb-ae97-e8cdc80e8e5a tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 831.009783] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-acef35f4-b906-4efb-ae97-e8cdc80e8e5a tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 831.011567] env[62066]: DEBUG oslo_vmware.api [None req-d2edd17d-ac3d-4b81-90f1-278c62ed416c tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Waiting for the task: (returnval){ [ 831.011567] env[62066]: value = "task-1340863" [ 831.011567] env[62066]: _type = "Task" [ 831.011567] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.012044] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-86236adb-2733-4c5c-af04-84d0f96de278 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.021059] env[62066]: DEBUG oslo_vmware.api [None req-acef35f4-b906-4efb-ae97-e8cdc80e8e5a tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Waiting for the task: (returnval){ [ 831.021059] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]5201e691-0d3a-28f6-57c0-711626608820" [ 831.021059] env[62066]: _type = "Task" [ 831.021059] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.025166] env[62066]: DEBUG oslo_vmware.api [None req-d2edd17d-ac3d-4b81-90f1-278c62ed416c tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Task: {'id': task-1340863, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.033199] env[62066]: DEBUG oslo_vmware.api [None req-acef35f4-b906-4efb-ae97-e8cdc80e8e5a tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]5201e691-0d3a-28f6-57c0-711626608820, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.184996] env[62066]: DEBUG oslo_vmware.api [None req-b8836db2-293e-442c-80f2-42d8bc4c2918 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1340861, 'name': PowerOnVM_Task} progress is 64%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.186296] env[62066]: DEBUG nova.network.neutron [None req-02667002-03be-4979-8517-c36490492214 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 86a92b61-f3e9-48f9-8ee2-756669d558ef] Updating instance_info_cache with network_info: [{"id": "fd69e6c7-721d-4947-80e1-4db6e9a230b0", "address": "fa:16:3e:ac:f0:6d", "network": {"id": "9f3396ce-ee48-4eaf-8cc7-3e1711f37537", "bridge": "br-int", "label": "tempest-ServersTestJSON-2129775093-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6497ab02f327476d8ff81c2ecc0371e5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7ab8d568-adb0-4f3b-b6cc-68413e6546ae", "external-id": "nsx-vlan-transportzone-86", "segmentation_id": 86, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfd69e6c7-72", "ovs_interfaceid": "fd69e6c7-721d-4947-80e1-4db6e9a230b0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 831.187649] env[62066]: DEBUG oslo_concurrency.lockutils [None req-8589d93a-2555-4139-b6bf-f9de6865cf74 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Acquiring lock "49f46244-34fa-48a1-95a2-8e95850f345d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 831.188068] env[62066]: DEBUG oslo_concurrency.lockutils [None req-8589d93a-2555-4139-b6bf-f9de6865cf74 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Lock "49f46244-34fa-48a1-95a2-8e95850f345d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 831.188292] env[62066]: DEBUG oslo_concurrency.lockutils [None req-8589d93a-2555-4139-b6bf-f9de6865cf74 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Acquiring lock "49f46244-34fa-48a1-95a2-8e95850f345d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 831.188491] env[62066]: DEBUG oslo_concurrency.lockutils [None req-8589d93a-2555-4139-b6bf-f9de6865cf74 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Lock "49f46244-34fa-48a1-95a2-8e95850f345d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 831.188666] env[62066]: DEBUG oslo_concurrency.lockutils [None req-8589d93a-2555-4139-b6bf-f9de6865cf74 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Lock "49f46244-34fa-48a1-95a2-8e95850f345d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 831.195119] env[62066]: INFO nova.compute.manager [None req-8589d93a-2555-4139-b6bf-f9de6865cf74 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 49f46244-34fa-48a1-95a2-8e95850f345d] Terminating instance [ 831.197781] env[62066]: DEBUG nova.compute.manager [None req-8589d93a-2555-4139-b6bf-f9de6865cf74 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 49f46244-34fa-48a1-95a2-8e95850f345d] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 831.198015] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-8589d93a-2555-4139-b6bf-f9de6865cf74 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 49f46244-34fa-48a1-95a2-8e95850f345d] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 831.198851] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afcbce62-f633-4344-88e1-6c3059cef9ce {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.208555] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-8589d93a-2555-4139-b6bf-f9de6865cf74 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 49f46244-34fa-48a1-95a2-8e95850f345d] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 831.208858] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7263a7f5-ddae-43eb-a610-a81fef1b46b9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.253832] env[62066]: DEBUG oslo_vmware.rw_handles [None req-41432ebd-8922-4676-9cc7-6eed1922da68 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/520e24d9-6862-b3ad-ba9f-81ea21948905/disk-0.vmdk. {{(pid=62066) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 831.254101] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b82e9fc-9654-4b46-856a-77ad3ab493ed {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.260265] env[62066]: DEBUG oslo_vmware.rw_handles [None req-41432ebd-8922-4676-9cc7-6eed1922da68 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/520e24d9-6862-b3ad-ba9f-81ea21948905/disk-0.vmdk is in state: ready. {{(pid=62066) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 831.260554] env[62066]: ERROR oslo_vmware.rw_handles [None req-41432ebd-8922-4676-9cc7-6eed1922da68 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/520e24d9-6862-b3ad-ba9f-81ea21948905/disk-0.vmdk due to incomplete transfer. [ 831.260711] env[62066]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-7ac06a4c-c0fe-403a-81c4-85b2bb3f7798 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.267897] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-8589d93a-2555-4139-b6bf-f9de6865cf74 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 49f46244-34fa-48a1-95a2-8e95850f345d] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 831.269057] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-8589d93a-2555-4139-b6bf-f9de6865cf74 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 49f46244-34fa-48a1-95a2-8e95850f345d] Deleting contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 831.269057] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-8589d93a-2555-4139-b6bf-f9de6865cf74 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Deleting the datastore file [datastore2] 49f46244-34fa-48a1-95a2-8e95850f345d {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 831.269322] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-39809a8a-30a3-497a-bbff-dccd03f0ad1e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.271256] env[62066]: DEBUG oslo_vmware.rw_handles [None req-41432ebd-8922-4676-9cc7-6eed1922da68 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/520e24d9-6862-b3ad-ba9f-81ea21948905/disk-0.vmdk. {{(pid=62066) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 831.271469] env[62066]: DEBUG nova.virt.vmwareapi.images [None req-41432ebd-8922-4676-9cc7-6eed1922da68 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] [instance: 1e1a2ede-4ad8-4600-851e-6f2046b3f919] Uploaded image b6a3f5de-6b26-4b8a-96cd-dc69c51db39d to the Glance image server {{(pid=62066) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 831.273284] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-41432ebd-8922-4676-9cc7-6eed1922da68 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] [instance: 1e1a2ede-4ad8-4600-851e-6f2046b3f919] Destroying the VM {{(pid=62066) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 831.274123] env[62066]: DEBUG nova.network.neutron [req-732fa2fe-08c1-4864-9047-5c8e399a2436 req-aaa9c9a8-f091-46ef-ab4a-61dac97c1614 service nova] [instance: 88daaaa6-f385-4161-bc74-9fffabb1145d] Updated VIF entry in instance network info cache for port ba002f90-5218-46f2-b7e1-cd96ae3018c7. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 831.274734] env[62066]: DEBUG nova.network.neutron [req-732fa2fe-08c1-4864-9047-5c8e399a2436 req-aaa9c9a8-f091-46ef-ab4a-61dac97c1614 service nova] [instance: 88daaaa6-f385-4161-bc74-9fffabb1145d] Updating instance_info_cache with network_info: [{"id": "ba002f90-5218-46f2-b7e1-cd96ae3018c7", "address": "fa:16:3e:54:27:b5", "network": {"id": "1c6c51e4-96f8-4175-b7f5-96b8a42fcb13", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-2006089809-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d609babc987a42e2a8ddb4bfb9c3b103", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a965790c-2d2f-4c2a-9ee7-745f4d53039b", "external-id": "nsx-vlan-transportzone-708", "segmentation_id": 708, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapba002f90-52", "ovs_interfaceid": "ba002f90-5218-46f2-b7e1-cd96ae3018c7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 831.276695] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-7a1f4246-8757-46c4-8710-17ec6f3af6a0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.284039] env[62066]: DEBUG oslo_vmware.api [None req-8589d93a-2555-4139-b6bf-f9de6865cf74 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Waiting for the task: (returnval){ [ 831.284039] env[62066]: value = "task-1340865" [ 831.284039] env[62066]: _type = "Task" [ 831.284039] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.286496] env[62066]: DEBUG oslo_vmware.api [None req-41432ebd-8922-4676-9cc7-6eed1922da68 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Waiting for the task: (returnval){ [ 831.286496] env[62066]: value = "task-1340866" [ 831.286496] env[62066]: _type = "Task" [ 831.286496] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.298402] env[62066]: DEBUG nova.compute.manager [req-4b838d08-c646-418e-99ad-1ba912c8bec4 req-0929f6cc-4017-4b94-8c04-58fb36680636 service nova] [instance: 5f645eaa-6edc-4362-82ad-38c4c57b2be2] Received event network-changed-65fcb910-75be-4a68-8cec-476f7cc7cd00 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 831.298625] env[62066]: DEBUG nova.compute.manager [req-4b838d08-c646-418e-99ad-1ba912c8bec4 req-0929f6cc-4017-4b94-8c04-58fb36680636 service nova] [instance: 5f645eaa-6edc-4362-82ad-38c4c57b2be2] Refreshing instance network info cache due to event network-changed-65fcb910-75be-4a68-8cec-476f7cc7cd00. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 831.298850] env[62066]: DEBUG oslo_concurrency.lockutils [req-4b838d08-c646-418e-99ad-1ba912c8bec4 req-0929f6cc-4017-4b94-8c04-58fb36680636 service nova] Acquiring lock "refresh_cache-5f645eaa-6edc-4362-82ad-38c4c57b2be2" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 831.298999] env[62066]: DEBUG oslo_concurrency.lockutils [req-4b838d08-c646-418e-99ad-1ba912c8bec4 req-0929f6cc-4017-4b94-8c04-58fb36680636 service nova] Acquired lock "refresh_cache-5f645eaa-6edc-4362-82ad-38c4c57b2be2" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 831.299184] env[62066]: DEBUG nova.network.neutron [req-4b838d08-c646-418e-99ad-1ba912c8bec4 req-0929f6cc-4017-4b94-8c04-58fb36680636 service nova] [instance: 5f645eaa-6edc-4362-82ad-38c4c57b2be2] Refreshing network info cache for port 65fcb910-75be-4a68-8cec-476f7cc7cd00 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 831.308639] env[62066]: DEBUG oslo_vmware.api [None req-8589d93a-2555-4139-b6bf-f9de6865cf74 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': task-1340865, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.311297] env[62066]: DEBUG oslo_vmware.api [None req-41432ebd-8922-4676-9cc7-6eed1922da68 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Task: {'id': task-1340866, 'name': Destroy_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.367049] env[62066]: DEBUG oslo_vmware.api [None req-8d3cb2ee-be58-4e07-b64f-9d5c34ab9462 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': task-1340862, 'name': RemoveSnapshot_Task, 'duration_secs': 0.574161} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.367445] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-8d3cb2ee-be58-4e07-b64f-9d5c34ab9462 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 49f46244-34fa-48a1-95a2-8e95850f345d] Deleted Snapshot of the VM instance {{(pid=62066) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 831.367728] env[62066]: INFO nova.compute.manager [None req-8d3cb2ee-be58-4e07-b64f-9d5c34ab9462 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 49f46244-34fa-48a1-95a2-8e95850f345d] Took 17.40 seconds to snapshot the instance on the hypervisor. [ 831.521853] env[62066]: DEBUG oslo_concurrency.lockutils [None req-423c892e-6435-422f-ab41-7d39bb22efa3 tempest-ServerGroupTestJSON-582792794 tempest-ServerGroupTestJSON-582792794-project-member] Lock "738ebb2f-7cc6-4d0f-871c-1428c9c21384" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 13.729s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 831.528492] env[62066]: DEBUG oslo_vmware.api [None req-d2edd17d-ac3d-4b81-90f1-278c62ed416c tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Task: {'id': task-1340863, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.547062] env[62066]: DEBUG oslo_vmware.api [None req-acef35f4-b906-4efb-ae97-e8cdc80e8e5a tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]5201e691-0d3a-28f6-57c0-711626608820, 'name': SearchDatastore_Task, 'duration_secs': 0.017854} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.550634] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dcb60c4b-bca7-456e-86bb-12b518cf7767 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.558152] env[62066]: DEBUG oslo_vmware.api [None req-acef35f4-b906-4efb-ae97-e8cdc80e8e5a tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Waiting for the task: (returnval){ [ 831.558152] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]522a0c1d-370e-19ca-c880-ef310afd9744" [ 831.558152] env[62066]: _type = "Task" [ 831.558152] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.566407] env[62066]: DEBUG oslo_vmware.api [None req-acef35f4-b906-4efb-ae97-e8cdc80e8e5a tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]522a0c1d-370e-19ca-c880-ef310afd9744, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.691142] env[62066]: DEBUG oslo_vmware.api [None req-b8836db2-293e-442c-80f2-42d8bc4c2918 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1340861, 'name': PowerOnVM_Task} progress is 64%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.692322] env[62066]: DEBUG oslo_concurrency.lockutils [None req-02667002-03be-4979-8517-c36490492214 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Releasing lock "refresh_cache-86a92b61-f3e9-48f9-8ee2-756669d558ef" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 831.692840] env[62066]: DEBUG nova.compute.manager [None req-02667002-03be-4979-8517-c36490492214 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 86a92b61-f3e9-48f9-8ee2-756669d558ef] Instance network_info: |[{"id": "fd69e6c7-721d-4947-80e1-4db6e9a230b0", "address": "fa:16:3e:ac:f0:6d", "network": {"id": "9f3396ce-ee48-4eaf-8cc7-3e1711f37537", "bridge": "br-int", "label": "tempest-ServersTestJSON-2129775093-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6497ab02f327476d8ff81c2ecc0371e5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7ab8d568-adb0-4f3b-b6cc-68413e6546ae", "external-id": "nsx-vlan-transportzone-86", "segmentation_id": 86, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfd69e6c7-72", "ovs_interfaceid": "fd69e6c7-721d-4947-80e1-4db6e9a230b0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 831.693563] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-02667002-03be-4979-8517-c36490492214 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 86a92b61-f3e9-48f9-8ee2-756669d558ef] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ac:f0:6d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7ab8d568-adb0-4f3b-b6cc-68413e6546ae', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'fd69e6c7-721d-4947-80e1-4db6e9a230b0', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 831.706944] env[62066]: DEBUG oslo.service.loopingcall [None req-02667002-03be-4979-8517-c36490492214 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 831.711281] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 86a92b61-f3e9-48f9-8ee2-756669d558ef] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 831.713032] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1d604fb8-e523-4b3f-8aeb-d54684af1023 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.752903] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 831.752903] env[62066]: value = "task-1340867" [ 831.752903] env[62066]: _type = "Task" [ 831.752903] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.769150] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1340867, 'name': CreateVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.773056] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19b0fd2b-33af-403a-8e37-aa7443b02cf6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.784727] env[62066]: DEBUG oslo_concurrency.lockutils [req-732fa2fe-08c1-4864-9047-5c8e399a2436 req-aaa9c9a8-f091-46ef-ab4a-61dac97c1614 service nova] Releasing lock "refresh_cache-88daaaa6-f385-4161-bc74-9fffabb1145d" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 831.795401] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5e187f4-5a1c-426b-85f5-6e123b32c0fe {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.846227] env[62066]: DEBUG oslo_vmware.api [None req-8589d93a-2555-4139-b6bf-f9de6865cf74 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': task-1340865, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.847690] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a16f915d-444e-47b1-9a1d-7e010e29e65d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.854220] env[62066]: DEBUG oslo_vmware.api [None req-41432ebd-8922-4676-9cc7-6eed1922da68 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Task: {'id': task-1340866, 'name': Destroy_Task} progress is 33%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.860326] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ff82186-dbaf-4267-835c-7d651a5ac0c4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.877542] env[62066]: DEBUG nova.compute.manager [None req-8d3cb2ee-be58-4e07-b64f-9d5c34ab9462 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 49f46244-34fa-48a1-95a2-8e95850f345d] Instance disappeared during snapshot {{(pid=62066) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4494}} [ 831.879027] env[62066]: DEBUG nova.compute.provider_tree [None req-1e760fa3-f701-40c8-8666-21a2eeae0751 tempest-ServerAddressesTestJSON-1053293248 tempest-ServerAddressesTestJSON-1053293248-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 831.901784] env[62066]: DEBUG nova.compute.manager [None req-8d3cb2ee-be58-4e07-b64f-9d5c34ab9462 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Image not found during clean up 843e9ee8-31ee-40e2-9f6f-50ffd87c6796 {{(pid=62066) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4500}} [ 832.030315] env[62066]: DEBUG oslo_vmware.api [None req-d2edd17d-ac3d-4b81-90f1-278c62ed416c tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Task: {'id': task-1340863, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.072113] env[62066]: DEBUG oslo_vmware.api [None req-acef35f4-b906-4efb-ae97-e8cdc80e8e5a tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]522a0c1d-370e-19ca-c880-ef310afd9744, 'name': SearchDatastore_Task, 'duration_secs': 0.025605} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.072113] env[62066]: DEBUG oslo_concurrency.lockutils [None req-acef35f4-b906-4efb-ae97-e8cdc80e8e5a tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Releasing lock "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 832.072113] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-acef35f4-b906-4efb-ae97-e8cdc80e8e5a tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk to [datastore1] 5e96def2-0cbd-4bd9-93f4-6a365a0142b0/5e96def2-0cbd-4bd9-93f4-6a365a0142b0.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 832.072366] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c0af3ef7-d410-4351-86b0-1add337bd4d1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.079210] env[62066]: DEBUG oslo_vmware.api [None req-acef35f4-b906-4efb-ae97-e8cdc80e8e5a tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Waiting for the task: (returnval){ [ 832.079210] env[62066]: value = "task-1340868" [ 832.079210] env[62066]: _type = "Task" [ 832.079210] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.088103] env[62066]: DEBUG oslo_vmware.api [None req-acef35f4-b906-4efb-ae97-e8cdc80e8e5a tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Task: {'id': task-1340868, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.189073] env[62066]: DEBUG oslo_vmware.api [None req-b8836db2-293e-442c-80f2-42d8bc4c2918 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1340861, 'name': PowerOnVM_Task, 'duration_secs': 1.867464} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.189378] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-b8836db2-293e-442c-80f2-42d8bc4c2918 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 88daaaa6-f385-4161-bc74-9fffabb1145d] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 832.189779] env[62066]: INFO nova.compute.manager [None req-b8836db2-293e-442c-80f2-42d8bc4c2918 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 88daaaa6-f385-4161-bc74-9fffabb1145d] Took 10.28 seconds to spawn the instance on the hypervisor. [ 832.192443] env[62066]: DEBUG nova.compute.manager [None req-b8836db2-293e-442c-80f2-42d8bc4c2918 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 88daaaa6-f385-4161-bc74-9fffabb1145d] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 832.192443] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d514f6a-4914-44b1-8757-61701333b9af {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.196078] env[62066]: DEBUG nova.network.neutron [req-4b838d08-c646-418e-99ad-1ba912c8bec4 req-0929f6cc-4017-4b94-8c04-58fb36680636 service nova] [instance: 5f645eaa-6edc-4362-82ad-38c4c57b2be2] Updated VIF entry in instance network info cache for port 65fcb910-75be-4a68-8cec-476f7cc7cd00. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 832.196499] env[62066]: DEBUG nova.network.neutron [req-4b838d08-c646-418e-99ad-1ba912c8bec4 req-0929f6cc-4017-4b94-8c04-58fb36680636 service nova] [instance: 5f645eaa-6edc-4362-82ad-38c4c57b2be2] Updating instance_info_cache with network_info: [{"id": "65fcb910-75be-4a68-8cec-476f7cc7cd00", "address": "fa:16:3e:20:29:72", "network": {"id": "b0f4607a-9d59-457e-aabc-5851c90b3366", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-1154549021-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a9b239f4cb3242d0bd1dcd98a2a3bfc0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c02dd284-ab80-451c-93eb-48c8360acb9c", "external-id": "nsx-vlan-transportzone-818", "segmentation_id": 818, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap65fcb910-75", "ovs_interfaceid": "65fcb910-75be-4a68-8cec-476f7cc7cd00", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 832.241441] env[62066]: DEBUG nova.compute.manager [req-97b5a9ff-1955-48e1-b324-f1f25741440b req-8fd77951-be69-4a8c-8790-6001cd148a56 service nova] [instance: de2e2be7-efdb-45a8-842a-640ab9deb1d9] Received event network-vif-plugged-c833a993-4af0-4bdb-90d2-cc82a796a560 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 832.241870] env[62066]: DEBUG oslo_concurrency.lockutils [req-97b5a9ff-1955-48e1-b324-f1f25741440b req-8fd77951-be69-4a8c-8790-6001cd148a56 service nova] Acquiring lock "de2e2be7-efdb-45a8-842a-640ab9deb1d9-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 832.242292] env[62066]: DEBUG oslo_concurrency.lockutils [req-97b5a9ff-1955-48e1-b324-f1f25741440b req-8fd77951-be69-4a8c-8790-6001cd148a56 service nova] Lock "de2e2be7-efdb-45a8-842a-640ab9deb1d9-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 832.242596] env[62066]: DEBUG oslo_concurrency.lockutils [req-97b5a9ff-1955-48e1-b324-f1f25741440b req-8fd77951-be69-4a8c-8790-6001cd148a56 service nova] Lock "de2e2be7-efdb-45a8-842a-640ab9deb1d9-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 832.242888] env[62066]: DEBUG nova.compute.manager [req-97b5a9ff-1955-48e1-b324-f1f25741440b req-8fd77951-be69-4a8c-8790-6001cd148a56 service nova] [instance: de2e2be7-efdb-45a8-842a-640ab9deb1d9] No waiting events found dispatching network-vif-plugged-c833a993-4af0-4bdb-90d2-cc82a796a560 {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 832.243182] env[62066]: WARNING nova.compute.manager [req-97b5a9ff-1955-48e1-b324-f1f25741440b req-8fd77951-be69-4a8c-8790-6001cd148a56 service nova] [instance: de2e2be7-efdb-45a8-842a-640ab9deb1d9] Received unexpected event network-vif-plugged-c833a993-4af0-4bdb-90d2-cc82a796a560 for instance with vm_state building and task_state spawning. [ 832.243458] env[62066]: DEBUG nova.compute.manager [req-97b5a9ff-1955-48e1-b324-f1f25741440b req-8fd77951-be69-4a8c-8790-6001cd148a56 service nova] [instance: de2e2be7-efdb-45a8-842a-640ab9deb1d9] Received event network-changed-c833a993-4af0-4bdb-90d2-cc82a796a560 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 832.243720] env[62066]: DEBUG nova.compute.manager [req-97b5a9ff-1955-48e1-b324-f1f25741440b req-8fd77951-be69-4a8c-8790-6001cd148a56 service nova] [instance: de2e2be7-efdb-45a8-842a-640ab9deb1d9] Refreshing instance network info cache due to event network-changed-c833a993-4af0-4bdb-90d2-cc82a796a560. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 832.244036] env[62066]: DEBUG oslo_concurrency.lockutils [req-97b5a9ff-1955-48e1-b324-f1f25741440b req-8fd77951-be69-4a8c-8790-6001cd148a56 service nova] Acquiring lock "refresh_cache-de2e2be7-efdb-45a8-842a-640ab9deb1d9" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 832.244300] env[62066]: DEBUG oslo_concurrency.lockutils [req-97b5a9ff-1955-48e1-b324-f1f25741440b req-8fd77951-be69-4a8c-8790-6001cd148a56 service nova] Acquired lock "refresh_cache-de2e2be7-efdb-45a8-842a-640ab9deb1d9" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 832.244648] env[62066]: DEBUG nova.network.neutron [req-97b5a9ff-1955-48e1-b324-f1f25741440b req-8fd77951-be69-4a8c-8790-6001cd148a56 service nova] [instance: de2e2be7-efdb-45a8-842a-640ab9deb1d9] Refreshing network info cache for port c833a993-4af0-4bdb-90d2-cc82a796a560 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 832.267799] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1340867, 'name': CreateVM_Task, 'duration_secs': 0.41962} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.267799] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 86a92b61-f3e9-48f9-8ee2-756669d558ef] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 832.267799] env[62066]: DEBUG oslo_concurrency.lockutils [None req-02667002-03be-4979-8517-c36490492214 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 832.267799] env[62066]: DEBUG oslo_concurrency.lockutils [None req-02667002-03be-4979-8517-c36490492214 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Acquired lock "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 832.267799] env[62066]: DEBUG oslo_concurrency.lockutils [None req-02667002-03be-4979-8517-c36490492214 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 832.267799] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9030d35d-30c3-4db6-b056-bb57f2c1123f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.272537] env[62066]: DEBUG oslo_vmware.api [None req-02667002-03be-4979-8517-c36490492214 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Waiting for the task: (returnval){ [ 832.272537] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]52ca4c85-31b5-65fe-d4be-53983c250a08" [ 832.272537] env[62066]: _type = "Task" [ 832.272537] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.285108] env[62066]: DEBUG oslo_concurrency.lockutils [None req-6075adee-b71b-4aa1-9781-77a864c59e59 tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Acquiring lock "5f645eaa-6edc-4362-82ad-38c4c57b2be2" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 832.285660] env[62066]: DEBUG oslo_concurrency.lockutils [None req-6075adee-b71b-4aa1-9781-77a864c59e59 tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Lock "5f645eaa-6edc-4362-82ad-38c4c57b2be2" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 832.285757] env[62066]: DEBUG oslo_concurrency.lockutils [None req-6075adee-b71b-4aa1-9781-77a864c59e59 tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Acquiring lock "5f645eaa-6edc-4362-82ad-38c4c57b2be2-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 832.286960] env[62066]: DEBUG oslo_concurrency.lockutils [None req-6075adee-b71b-4aa1-9781-77a864c59e59 tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Lock "5f645eaa-6edc-4362-82ad-38c4c57b2be2-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 832.286960] env[62066]: DEBUG oslo_concurrency.lockutils [None req-6075adee-b71b-4aa1-9781-77a864c59e59 tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Lock "5f645eaa-6edc-4362-82ad-38c4c57b2be2-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 832.288408] env[62066]: DEBUG oslo_vmware.api [None req-02667002-03be-4979-8517-c36490492214 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52ca4c85-31b5-65fe-d4be-53983c250a08, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.289040] env[62066]: INFO nova.compute.manager [None req-6075adee-b71b-4aa1-9781-77a864c59e59 tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] [instance: 5f645eaa-6edc-4362-82ad-38c4c57b2be2] Terminating instance [ 832.295656] env[62066]: DEBUG nova.compute.manager [None req-6075adee-b71b-4aa1-9781-77a864c59e59 tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] [instance: 5f645eaa-6edc-4362-82ad-38c4c57b2be2] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 832.295978] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-6075adee-b71b-4aa1-9781-77a864c59e59 tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] [instance: 5f645eaa-6edc-4362-82ad-38c4c57b2be2] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 832.297315] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fb27848-0443-4540-9b09-a3b26f3ea8e6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.308395] env[62066]: DEBUG oslo_vmware.api [None req-8589d93a-2555-4139-b6bf-f9de6865cf74 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': task-1340865, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.54947} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.308668] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-6075adee-b71b-4aa1-9781-77a864c59e59 tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] [instance: 5f645eaa-6edc-4362-82ad-38c4c57b2be2] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 832.309273] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-8589d93a-2555-4139-b6bf-f9de6865cf74 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 832.309479] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-8589d93a-2555-4139-b6bf-f9de6865cf74 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 49f46244-34fa-48a1-95a2-8e95850f345d] Deleted contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 832.309671] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-8589d93a-2555-4139-b6bf-f9de6865cf74 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 49f46244-34fa-48a1-95a2-8e95850f345d] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 832.309848] env[62066]: INFO nova.compute.manager [None req-8589d93a-2555-4139-b6bf-f9de6865cf74 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 49f46244-34fa-48a1-95a2-8e95850f345d] Took 1.11 seconds to destroy the instance on the hypervisor. [ 832.310209] env[62066]: DEBUG oslo.service.loopingcall [None req-8589d93a-2555-4139-b6bf-f9de6865cf74 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 832.310290] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2125330d-b1e6-46db-9fcb-e60ec7adf9c8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.314492] env[62066]: DEBUG nova.compute.manager [-] [instance: 49f46244-34fa-48a1-95a2-8e95850f345d] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 832.314592] env[62066]: DEBUG nova.network.neutron [-] [instance: 49f46244-34fa-48a1-95a2-8e95850f345d] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 832.316960] env[62066]: DEBUG oslo_vmware.api [None req-41432ebd-8922-4676-9cc7-6eed1922da68 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Task: {'id': task-1340866, 'name': Destroy_Task, 'duration_secs': 0.85835} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.316960] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-41432ebd-8922-4676-9cc7-6eed1922da68 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] [instance: 1e1a2ede-4ad8-4600-851e-6f2046b3f919] Destroyed the VM [ 832.317112] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-41432ebd-8922-4676-9cc7-6eed1922da68 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] [instance: 1e1a2ede-4ad8-4600-851e-6f2046b3f919] Deleting Snapshot of the VM instance {{(pid=62066) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 832.317831] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-6421dc5a-f909-49a6-b9ff-93d068009a17 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.320816] env[62066]: DEBUG oslo_vmware.api [None req-6075adee-b71b-4aa1-9781-77a864c59e59 tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Waiting for the task: (returnval){ [ 832.320816] env[62066]: value = "task-1340869" [ 832.320816] env[62066]: _type = "Task" [ 832.320816] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.327413] env[62066]: DEBUG oslo_vmware.api [None req-41432ebd-8922-4676-9cc7-6eed1922da68 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Waiting for the task: (returnval){ [ 832.327413] env[62066]: value = "task-1340870" [ 832.327413] env[62066]: _type = "Task" [ 832.327413] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.334668] env[62066]: DEBUG oslo_vmware.api [None req-6075adee-b71b-4aa1-9781-77a864c59e59 tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Task: {'id': task-1340869, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.339746] env[62066]: DEBUG oslo_vmware.api [None req-41432ebd-8922-4676-9cc7-6eed1922da68 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Task: {'id': task-1340870, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.384910] env[62066]: DEBUG nova.scheduler.client.report [None req-1e760fa3-f701-40c8-8666-21a2eeae0751 tempest-ServerAddressesTestJSON-1053293248 tempest-ServerAddressesTestJSON-1053293248-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 832.392953] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e3db032e-002b-4e25-a569-0fed0f11200b tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Acquiring lock "b08f27b1-429e-4d05-9a95-f45fd0cb2904" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 832.392953] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e3db032e-002b-4e25-a569-0fed0f11200b tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Lock "b08f27b1-429e-4d05-9a95-f45fd0cb2904" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 832.530871] env[62066]: DEBUG oslo_vmware.api [None req-d2edd17d-ac3d-4b81-90f1-278c62ed416c tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Task: {'id': task-1340863, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.079408} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.531450] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-d2edd17d-ac3d-4b81-90f1-278c62ed416c tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk to [datastore1] de2e2be7-efdb-45a8-842a-640ab9deb1d9/de2e2be7-efdb-45a8-842a-640ab9deb1d9.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 832.531872] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-d2edd17d-ac3d-4b81-90f1-278c62ed416c tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] [instance: de2e2be7-efdb-45a8-842a-640ab9deb1d9] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 832.532282] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2bacfe32-8c80-49a8-953a-8531273efba2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.542427] env[62066]: DEBUG oslo_vmware.api [None req-d2edd17d-ac3d-4b81-90f1-278c62ed416c tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Waiting for the task: (returnval){ [ 832.542427] env[62066]: value = "task-1340871" [ 832.542427] env[62066]: _type = "Task" [ 832.542427] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.549945] env[62066]: DEBUG oslo_vmware.api [None req-d2edd17d-ac3d-4b81-90f1-278c62ed416c tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Task: {'id': task-1340871, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.594045] env[62066]: DEBUG oslo_vmware.api [None req-acef35f4-b906-4efb-ae97-e8cdc80e8e5a tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Task: {'id': task-1340868, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.707338] env[62066]: DEBUG oslo_concurrency.lockutils [req-4b838d08-c646-418e-99ad-1ba912c8bec4 req-0929f6cc-4017-4b94-8c04-58fb36680636 service nova] Releasing lock "refresh_cache-5f645eaa-6edc-4362-82ad-38c4c57b2be2" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 832.716331] env[62066]: INFO nova.compute.manager [None req-b8836db2-293e-442c-80f2-42d8bc4c2918 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 88daaaa6-f385-4161-bc74-9fffabb1145d] Took 28.17 seconds to build instance. [ 832.784520] env[62066]: DEBUG oslo_vmware.api [None req-02667002-03be-4979-8517-c36490492214 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52ca4c85-31b5-65fe-d4be-53983c250a08, 'name': SearchDatastore_Task, 'duration_secs': 0.019503} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.784912] env[62066]: DEBUG oslo_concurrency.lockutils [None req-02667002-03be-4979-8517-c36490492214 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Releasing lock "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 832.785182] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-02667002-03be-4979-8517-c36490492214 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 86a92b61-f3e9-48f9-8ee2-756669d558ef] Processing image 50ff584c-3b50-4395-af07-3e66769bc9f7 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 832.785448] env[62066]: DEBUG oslo_concurrency.lockutils [None req-02667002-03be-4979-8517-c36490492214 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 832.785688] env[62066]: DEBUG oslo_concurrency.lockutils [None req-02667002-03be-4979-8517-c36490492214 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Acquired lock "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 832.785898] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-02667002-03be-4979-8517-c36490492214 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 832.786220] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1acfd804-1fee-4441-970b-52f665e2e525 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.803310] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-02667002-03be-4979-8517-c36490492214 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 832.803570] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-02667002-03be-4979-8517-c36490492214 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 832.804399] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-31eaa66e-b826-4466-8282-b146a83442cd {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.812096] env[62066]: DEBUG oslo_vmware.api [None req-02667002-03be-4979-8517-c36490492214 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Waiting for the task: (returnval){ [ 832.812096] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]52ea71a6-c306-6b2c-223f-0c606e8b2bab" [ 832.812096] env[62066]: _type = "Task" [ 832.812096] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.825233] env[62066]: DEBUG oslo_vmware.api [None req-02667002-03be-4979-8517-c36490492214 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52ea71a6-c306-6b2c-223f-0c606e8b2bab, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.838975] env[62066]: DEBUG oslo_vmware.api [None req-6075adee-b71b-4aa1-9781-77a864c59e59 tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Task: {'id': task-1340869, 'name': PowerOffVM_Task, 'duration_secs': 0.196379} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.842821] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-6075adee-b71b-4aa1-9781-77a864c59e59 tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] [instance: 5f645eaa-6edc-4362-82ad-38c4c57b2be2] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 832.843122] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-6075adee-b71b-4aa1-9781-77a864c59e59 tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] [instance: 5f645eaa-6edc-4362-82ad-38c4c57b2be2] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 832.843558] env[62066]: DEBUG oslo_vmware.api [None req-41432ebd-8922-4676-9cc7-6eed1922da68 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Task: {'id': task-1340870, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.843693] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-681e0270-8e46-4d5d-ab5f-e352fb399578 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.893321] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1e760fa3-f701-40c8-8666-21a2eeae0751 tempest-ServerAddressesTestJSON-1053293248 tempest-ServerAddressesTestJSON-1053293248-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.931s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 832.897160] env[62066]: DEBUG oslo_concurrency.lockutils [None req-32db9ed7-c319-413e-978b-0abcfac2cab3 tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.781s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 832.897160] env[62066]: DEBUG nova.objects.instance [None req-32db9ed7-c319-413e-978b-0abcfac2cab3 tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] Lazy-loading 'resources' on Instance uuid 050040cb-6ee3-4ad2-960a-fcebb53ac394 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 832.897759] env[62066]: DEBUG nova.compute.manager [None req-e3db032e-002b-4e25-a569-0fed0f11200b tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: b08f27b1-429e-4d05-9a95-f45fd0cb2904] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 832.920654] env[62066]: INFO nova.scheduler.client.report [None req-1e760fa3-f701-40c8-8666-21a2eeae0751 tempest-ServerAddressesTestJSON-1053293248 tempest-ServerAddressesTestJSON-1053293248-project-member] Deleted allocations for instance 49bc1350-0095-406b-bc68-005eb6b681a8 [ 833.053105] env[62066]: DEBUG oslo_vmware.api [None req-d2edd17d-ac3d-4b81-90f1-278c62ed416c tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Task: {'id': task-1340871, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.122212} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.055711] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-d2edd17d-ac3d-4b81-90f1-278c62ed416c tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] [instance: de2e2be7-efdb-45a8-842a-640ab9deb1d9] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 833.057599] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e76eea4-4396-4b80-a353-c8f2cd1e3f64 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.083218] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-d2edd17d-ac3d-4b81-90f1-278c62ed416c tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] [instance: de2e2be7-efdb-45a8-842a-640ab9deb1d9] Reconfiguring VM instance instance-00000048 to attach disk [datastore1] de2e2be7-efdb-45a8-842a-640ab9deb1d9/de2e2be7-efdb-45a8-842a-640ab9deb1d9.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 833.084023] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-16b06d61-2fbb-4f88-8d2d-31ba630035f5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.113743] env[62066]: DEBUG oslo_vmware.api [None req-acef35f4-b906-4efb-ae97-e8cdc80e8e5a tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Task: {'id': task-1340868, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.869716} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.115070] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-acef35f4-b906-4efb-ae97-e8cdc80e8e5a tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk to [datastore1] 5e96def2-0cbd-4bd9-93f4-6a365a0142b0/5e96def2-0cbd-4bd9-93f4-6a365a0142b0.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 833.115311] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-acef35f4-b906-4efb-ae97-e8cdc80e8e5a tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] [instance: 5e96def2-0cbd-4bd9-93f4-6a365a0142b0] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 833.115641] env[62066]: DEBUG oslo_vmware.api [None req-d2edd17d-ac3d-4b81-90f1-278c62ed416c tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Waiting for the task: (returnval){ [ 833.115641] env[62066]: value = "task-1340873" [ 833.115641] env[62066]: _type = "Task" [ 833.115641] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.115848] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ac3bf1f1-acde-4613-9aef-90ac4878eaa1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.128596] env[62066]: DEBUG oslo_vmware.api [None req-d2edd17d-ac3d-4b81-90f1-278c62ed416c tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Task: {'id': task-1340873, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.130443] env[62066]: DEBUG oslo_vmware.api [None req-acef35f4-b906-4efb-ae97-e8cdc80e8e5a tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Waiting for the task: (returnval){ [ 833.130443] env[62066]: value = "task-1340874" [ 833.130443] env[62066]: _type = "Task" [ 833.130443] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.141337] env[62066]: DEBUG oslo_vmware.api [None req-acef35f4-b906-4efb-ae97-e8cdc80e8e5a tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Task: {'id': task-1340874, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.220332] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b8836db2-293e-442c-80f2-42d8bc4c2918 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Lock "88daaaa6-f385-4161-bc74-9fffabb1145d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 90.043s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 833.307757] env[62066]: DEBUG nova.network.neutron [req-97b5a9ff-1955-48e1-b324-f1f25741440b req-8fd77951-be69-4a8c-8790-6001cd148a56 service nova] [instance: de2e2be7-efdb-45a8-842a-640ab9deb1d9] Updated VIF entry in instance network info cache for port c833a993-4af0-4bdb-90d2-cc82a796a560. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 833.307757] env[62066]: DEBUG nova.network.neutron [req-97b5a9ff-1955-48e1-b324-f1f25741440b req-8fd77951-be69-4a8c-8790-6001cd148a56 service nova] [instance: de2e2be7-efdb-45a8-842a-640ab9deb1d9] Updating instance_info_cache with network_info: [{"id": "c833a993-4af0-4bdb-90d2-cc82a796a560", "address": "fa:16:3e:71:ab:fe", "network": {"id": "4114ddfe-818f-47a5-bb6f-44a4c22517f6", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-676131672-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dd30584efc254610b3939d1ba3806693", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a485857d-7086-4dcf-9d65-d0dcd177fcb0", "external-id": "nsx-vlan-transportzone-232", "segmentation_id": 232, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc833a993-4a", "ovs_interfaceid": "c833a993-4af0-4bdb-90d2-cc82a796a560", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 833.324465] env[62066]: DEBUG oslo_vmware.api [None req-02667002-03be-4979-8517-c36490492214 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52ea71a6-c306-6b2c-223f-0c606e8b2bab, 'name': SearchDatastore_Task, 'duration_secs': 0.060761} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.325307] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c1653e55-f3c2-403a-9f55-927ec5707091 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.330610] env[62066]: DEBUG oslo_vmware.api [None req-02667002-03be-4979-8517-c36490492214 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Waiting for the task: (returnval){ [ 833.330610] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]5284ce35-0867-cebb-d429-95711f1b2e98" [ 833.330610] env[62066]: _type = "Task" [ 833.330610] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.341369] env[62066]: DEBUG oslo_vmware.api [None req-41432ebd-8922-4676-9cc7-6eed1922da68 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Task: {'id': task-1340870, 'name': RemoveSnapshot_Task} progress is 70%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.347189] env[62066]: DEBUG oslo_vmware.api [None req-02667002-03be-4979-8517-c36490492214 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]5284ce35-0867-cebb-d429-95711f1b2e98, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.373488] env[62066]: INFO nova.compute.manager [None req-28a644a7-dc09-4460-bd7c-0b89c8effe40 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 88daaaa6-f385-4161-bc74-9fffabb1145d] Rebuilding instance [ 833.428366] env[62066]: DEBUG nova.compute.manager [None req-28a644a7-dc09-4460-bd7c-0b89c8effe40 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 88daaaa6-f385-4161-bc74-9fffabb1145d] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 833.431442] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32a543ee-ca7b-4878-9612-42c4c9b947ff {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.437510] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e3db032e-002b-4e25-a569-0fed0f11200b tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 833.438367] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1e760fa3-f701-40c8-8666-21a2eeae0751 tempest-ServerAddressesTestJSON-1053293248 tempest-ServerAddressesTestJSON-1053293248-project-member] Lock "49bc1350-0095-406b-bc68-005eb6b681a8" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 13.619s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 833.498238] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-6075adee-b71b-4aa1-9781-77a864c59e59 tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] [instance: 5f645eaa-6edc-4362-82ad-38c4c57b2be2] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 833.498480] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-6075adee-b71b-4aa1-9781-77a864c59e59 tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] [instance: 5f645eaa-6edc-4362-82ad-38c4c57b2be2] Deleting contents of the VM from datastore datastore1 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 833.498670] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-6075adee-b71b-4aa1-9781-77a864c59e59 tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Deleting the datastore file [datastore1] 5f645eaa-6edc-4362-82ad-38c4c57b2be2 {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 833.499037] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b7561dd8-ac58-4f8e-8242-430414dc8b5e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.509040] env[62066]: DEBUG oslo_vmware.api [None req-6075adee-b71b-4aa1-9781-77a864c59e59 tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Waiting for the task: (returnval){ [ 833.509040] env[62066]: value = "task-1340875" [ 833.509040] env[62066]: _type = "Task" [ 833.509040] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.517454] env[62066]: DEBUG oslo_vmware.api [None req-6075adee-b71b-4aa1-9781-77a864c59e59 tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Task: {'id': task-1340875, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.524672] env[62066]: DEBUG nova.network.neutron [-] [instance: 49f46244-34fa-48a1-95a2-8e95850f345d] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 833.631311] env[62066]: DEBUG oslo_vmware.api [None req-d2edd17d-ac3d-4b81-90f1-278c62ed416c tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Task: {'id': task-1340873, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.641441] env[62066]: DEBUG oslo_vmware.api [None req-acef35f4-b906-4efb-ae97-e8cdc80e8e5a tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Task: {'id': task-1340874, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071253} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.641558] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-acef35f4-b906-4efb-ae97-e8cdc80e8e5a tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] [instance: 5e96def2-0cbd-4bd9-93f4-6a365a0142b0] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 833.642411] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e63f25e6-7b7d-4ecb-a054-af9d6b3ada1a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.646478] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d56df669-5457-4006-b116-17180c04418f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.667274] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-acef35f4-b906-4efb-ae97-e8cdc80e8e5a tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] [instance: 5e96def2-0cbd-4bd9-93f4-6a365a0142b0] Reconfiguring VM instance instance-00000049 to attach disk [datastore1] 5e96def2-0cbd-4bd9-93f4-6a365a0142b0/5e96def2-0cbd-4bd9-93f4-6a365a0142b0.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 833.668171] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e6a06e7e-7c0a-4897-9574-3371d5a61ed1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.688015] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-977411f0-d522-406e-9539-89a4936a3ef0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.692828] env[62066]: DEBUG oslo_vmware.api [None req-acef35f4-b906-4efb-ae97-e8cdc80e8e5a tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Waiting for the task: (returnval){ [ 833.692828] env[62066]: value = "task-1340876" [ 833.692828] env[62066]: _type = "Task" [ 833.692828] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.721751] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-491b03b8-9e40-4cf6-bd24-e685d6be54d3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.727477] env[62066]: DEBUG oslo_vmware.api [None req-acef35f4-b906-4efb-ae97-e8cdc80e8e5a tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Task: {'id': task-1340876, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.732522] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1e994dd-9464-4209-9449-18a846db45a8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.746149] env[62066]: DEBUG nova.compute.provider_tree [None req-32db9ed7-c319-413e-978b-0abcfac2cab3 tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 833.809233] env[62066]: DEBUG oslo_concurrency.lockutils [req-97b5a9ff-1955-48e1-b324-f1f25741440b req-8fd77951-be69-4a8c-8790-6001cd148a56 service nova] Releasing lock "refresh_cache-de2e2be7-efdb-45a8-842a-640ab9deb1d9" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 833.809588] env[62066]: DEBUG nova.compute.manager [req-97b5a9ff-1955-48e1-b324-f1f25741440b req-8fd77951-be69-4a8c-8790-6001cd148a56 service nova] [instance: 050040cb-6ee3-4ad2-960a-fcebb53ac394] Received event network-vif-deleted-d329136d-bd14-4bee-b09a-39ef97845251 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 833.810086] env[62066]: DEBUG nova.compute.manager [req-97b5a9ff-1955-48e1-b324-f1f25741440b req-8fd77951-be69-4a8c-8790-6001cd148a56 service nova] [instance: 050040cb-6ee3-4ad2-960a-fcebb53ac394] Received event network-vif-deleted-72521b90-587a-4f59-b744-919ef3087539 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 833.810284] env[62066]: DEBUG nova.compute.manager [req-97b5a9ff-1955-48e1-b324-f1f25741440b req-8fd77951-be69-4a8c-8790-6001cd148a56 service nova] [instance: 5e96def2-0cbd-4bd9-93f4-6a365a0142b0] Received event network-vif-plugged-0932c7ac-630e-4cef-8991-45b438f39483 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 833.810538] env[62066]: DEBUG oslo_concurrency.lockutils [req-97b5a9ff-1955-48e1-b324-f1f25741440b req-8fd77951-be69-4a8c-8790-6001cd148a56 service nova] Acquiring lock "5e96def2-0cbd-4bd9-93f4-6a365a0142b0-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 833.810786] env[62066]: DEBUG oslo_concurrency.lockutils [req-97b5a9ff-1955-48e1-b324-f1f25741440b req-8fd77951-be69-4a8c-8790-6001cd148a56 service nova] Lock "5e96def2-0cbd-4bd9-93f4-6a365a0142b0-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 833.811558] env[62066]: DEBUG oslo_concurrency.lockutils [req-97b5a9ff-1955-48e1-b324-f1f25741440b req-8fd77951-be69-4a8c-8790-6001cd148a56 service nova] Lock "5e96def2-0cbd-4bd9-93f4-6a365a0142b0-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 833.811982] env[62066]: DEBUG nova.compute.manager [req-97b5a9ff-1955-48e1-b324-f1f25741440b req-8fd77951-be69-4a8c-8790-6001cd148a56 service nova] [instance: 5e96def2-0cbd-4bd9-93f4-6a365a0142b0] No waiting events found dispatching network-vif-plugged-0932c7ac-630e-4cef-8991-45b438f39483 {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 833.811982] env[62066]: WARNING nova.compute.manager [req-97b5a9ff-1955-48e1-b324-f1f25741440b req-8fd77951-be69-4a8c-8790-6001cd148a56 service nova] [instance: 5e96def2-0cbd-4bd9-93f4-6a365a0142b0] Received unexpected event network-vif-plugged-0932c7ac-630e-4cef-8991-45b438f39483 for instance with vm_state building and task_state spawning. [ 833.812201] env[62066]: DEBUG nova.compute.manager [req-97b5a9ff-1955-48e1-b324-f1f25741440b req-8fd77951-be69-4a8c-8790-6001cd148a56 service nova] [instance: 5e96def2-0cbd-4bd9-93f4-6a365a0142b0] Received event network-changed-0932c7ac-630e-4cef-8991-45b438f39483 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 833.812369] env[62066]: DEBUG nova.compute.manager [req-97b5a9ff-1955-48e1-b324-f1f25741440b req-8fd77951-be69-4a8c-8790-6001cd148a56 service nova] [instance: 5e96def2-0cbd-4bd9-93f4-6a365a0142b0] Refreshing instance network info cache due to event network-changed-0932c7ac-630e-4cef-8991-45b438f39483. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 833.812654] env[62066]: DEBUG oslo_concurrency.lockutils [req-97b5a9ff-1955-48e1-b324-f1f25741440b req-8fd77951-be69-4a8c-8790-6001cd148a56 service nova] Acquiring lock "refresh_cache-5e96def2-0cbd-4bd9-93f4-6a365a0142b0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 833.814332] env[62066]: DEBUG oslo_concurrency.lockutils [req-97b5a9ff-1955-48e1-b324-f1f25741440b req-8fd77951-be69-4a8c-8790-6001cd148a56 service nova] Acquired lock "refresh_cache-5e96def2-0cbd-4bd9-93f4-6a365a0142b0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 833.814565] env[62066]: DEBUG nova.network.neutron [req-97b5a9ff-1955-48e1-b324-f1f25741440b req-8fd77951-be69-4a8c-8790-6001cd148a56 service nova] [instance: 5e96def2-0cbd-4bd9-93f4-6a365a0142b0] Refreshing network info cache for port 0932c7ac-630e-4cef-8991-45b438f39483 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 833.847031] env[62066]: DEBUG oslo_vmware.api [None req-02667002-03be-4979-8517-c36490492214 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]5284ce35-0867-cebb-d429-95711f1b2e98, 'name': SearchDatastore_Task, 'duration_secs': 0.048568} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.850205] env[62066]: DEBUG oslo_concurrency.lockutils [None req-02667002-03be-4979-8517-c36490492214 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Releasing lock "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 833.850837] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-02667002-03be-4979-8517-c36490492214 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk to [datastore1] 86a92b61-f3e9-48f9-8ee2-756669d558ef/86a92b61-f3e9-48f9-8ee2-756669d558ef.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 833.850837] env[62066]: DEBUG oslo_vmware.api [None req-41432ebd-8922-4676-9cc7-6eed1922da68 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Task: {'id': task-1340870, 'name': RemoveSnapshot_Task} progress is 70%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.851102] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b9fea770-7807-42f9-92a1-47f2964e7d7f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.859098] env[62066]: DEBUG oslo_vmware.api [None req-02667002-03be-4979-8517-c36490492214 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Waiting for the task: (returnval){ [ 833.859098] env[62066]: value = "task-1340877" [ 833.859098] env[62066]: _type = "Task" [ 833.859098] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.868512] env[62066]: DEBUG oslo_vmware.api [None req-02667002-03be-4979-8517-c36490492214 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': task-1340877, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.948535] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-28a644a7-dc09-4460-bd7c-0b89c8effe40 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 88daaaa6-f385-4161-bc74-9fffabb1145d] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 833.948903] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fd87fc69-2316-4dfa-90ba-21e27254b1c1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.958963] env[62066]: DEBUG oslo_vmware.api [None req-28a644a7-dc09-4460-bd7c-0b89c8effe40 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Waiting for the task: (returnval){ [ 833.958963] env[62066]: value = "task-1340878" [ 833.958963] env[62066]: _type = "Task" [ 833.958963] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.971022] env[62066]: DEBUG oslo_vmware.api [None req-28a644a7-dc09-4460-bd7c-0b89c8effe40 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1340878, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.021888] env[62066]: DEBUG oslo_vmware.api [None req-6075adee-b71b-4aa1-9781-77a864c59e59 tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Task: {'id': task-1340875, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.027523] env[62066]: INFO nova.compute.manager [-] [instance: 49f46244-34fa-48a1-95a2-8e95850f345d] Took 1.71 seconds to deallocate network for instance. [ 834.140021] env[62066]: DEBUG oslo_vmware.api [None req-d2edd17d-ac3d-4b81-90f1-278c62ed416c tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Task: {'id': task-1340873, 'name': ReconfigVM_Task, 'duration_secs': 0.889971} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.140021] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-d2edd17d-ac3d-4b81-90f1-278c62ed416c tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] [instance: de2e2be7-efdb-45a8-842a-640ab9deb1d9] Reconfigured VM instance instance-00000048 to attach disk [datastore1] de2e2be7-efdb-45a8-842a-640ab9deb1d9/de2e2be7-efdb-45a8-842a-640ab9deb1d9.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 834.140021] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-443ff688-b0f5-4705-b361-7d98a314f1aa {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.146770] env[62066]: DEBUG oslo_vmware.api [None req-d2edd17d-ac3d-4b81-90f1-278c62ed416c tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Waiting for the task: (returnval){ [ 834.146770] env[62066]: value = "task-1340879" [ 834.146770] env[62066]: _type = "Task" [ 834.146770] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.157422] env[62066]: DEBUG oslo_vmware.api [None req-d2edd17d-ac3d-4b81-90f1-278c62ed416c tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Task: {'id': task-1340879, 'name': Rename_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.204100] env[62066]: DEBUG oslo_vmware.api [None req-acef35f4-b906-4efb-ae97-e8cdc80e8e5a tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Task: {'id': task-1340876, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.249492] env[62066]: DEBUG nova.scheduler.client.report [None req-32db9ed7-c319-413e-978b-0abcfac2cab3 tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 834.272197] env[62066]: DEBUG nova.compute.manager [req-19b0a469-9f2b-40b8-b227-d6b653e8b5e5 req-ba2dfaf0-20e4-474b-9da2-60f30cd8c25a service nova] [instance: 49f46244-34fa-48a1-95a2-8e95850f345d] Received event network-vif-deleted-213f129a-7b70-4807-ac40-621d02749750 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 834.345441] env[62066]: DEBUG oslo_vmware.api [None req-41432ebd-8922-4676-9cc7-6eed1922da68 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Task: {'id': task-1340870, 'name': RemoveSnapshot_Task, 'duration_secs': 1.891697} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.345796] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-41432ebd-8922-4676-9cc7-6eed1922da68 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] [instance: 1e1a2ede-4ad8-4600-851e-6f2046b3f919] Deleted Snapshot of the VM instance {{(pid=62066) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 834.346132] env[62066]: INFO nova.compute.manager [None req-41432ebd-8922-4676-9cc7-6eed1922da68 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] [instance: 1e1a2ede-4ad8-4600-851e-6f2046b3f919] Took 17.35 seconds to snapshot the instance on the hypervisor. [ 834.368433] env[62066]: DEBUG oslo_vmware.api [None req-02667002-03be-4979-8517-c36490492214 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': task-1340877, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.474522] env[62066]: DEBUG oslo_vmware.api [None req-28a644a7-dc09-4460-bd7c-0b89c8effe40 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1340878, 'name': PowerOffVM_Task, 'duration_secs': 0.336302} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.475193] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-28a644a7-dc09-4460-bd7c-0b89c8effe40 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 88daaaa6-f385-4161-bc74-9fffabb1145d] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 834.475541] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-28a644a7-dc09-4460-bd7c-0b89c8effe40 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 88daaaa6-f385-4161-bc74-9fffabb1145d] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 834.476700] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a4dd003-429a-4457-8afd-0a5880fa6acb {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.485565] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-28a644a7-dc09-4460-bd7c-0b89c8effe40 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 88daaaa6-f385-4161-bc74-9fffabb1145d] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 834.485927] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-bd56daec-ea9e-4da6-a8d0-72b481e8036b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.522113] env[62066]: DEBUG oslo_vmware.api [None req-6075adee-b71b-4aa1-9781-77a864c59e59 tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Task: {'id': task-1340875, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.610438} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.522429] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-6075adee-b71b-4aa1-9781-77a864c59e59 tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 834.523556] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-6075adee-b71b-4aa1-9781-77a864c59e59 tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] [instance: 5f645eaa-6edc-4362-82ad-38c4c57b2be2] Deleted contents of the VM from datastore datastore1 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 834.523818] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-6075adee-b71b-4aa1-9781-77a864c59e59 tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] [instance: 5f645eaa-6edc-4362-82ad-38c4c57b2be2] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 834.524248] env[62066]: INFO nova.compute.manager [None req-6075adee-b71b-4aa1-9781-77a864c59e59 tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] [instance: 5f645eaa-6edc-4362-82ad-38c4c57b2be2] Took 2.23 seconds to destroy the instance on the hypervisor. [ 834.524547] env[62066]: DEBUG oslo.service.loopingcall [None req-6075adee-b71b-4aa1-9781-77a864c59e59 tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 834.524773] env[62066]: DEBUG nova.compute.manager [-] [instance: 5f645eaa-6edc-4362-82ad-38c4c57b2be2] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 834.524861] env[62066]: DEBUG nova.network.neutron [-] [instance: 5f645eaa-6edc-4362-82ad-38c4c57b2be2] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 834.535034] env[62066]: DEBUG oslo_concurrency.lockutils [None req-8589d93a-2555-4139-b6bf-f9de6865cf74 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 834.657875] env[62066]: DEBUG oslo_vmware.api [None req-d2edd17d-ac3d-4b81-90f1-278c62ed416c tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Task: {'id': task-1340879, 'name': Rename_Task, 'duration_secs': 0.208477} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.658206] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-d2edd17d-ac3d-4b81-90f1-278c62ed416c tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] [instance: de2e2be7-efdb-45a8-842a-640ab9deb1d9] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 834.658516] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-32d9a4a1-88c8-4eec-a291-1923ea2e7cbd {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.666553] env[62066]: DEBUG oslo_vmware.api [None req-d2edd17d-ac3d-4b81-90f1-278c62ed416c tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Waiting for the task: (returnval){ [ 834.666553] env[62066]: value = "task-1340881" [ 834.666553] env[62066]: _type = "Task" [ 834.666553] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.677731] env[62066]: DEBUG oslo_vmware.api [None req-d2edd17d-ac3d-4b81-90f1-278c62ed416c tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Task: {'id': task-1340881, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.704567] env[62066]: DEBUG oslo_vmware.api [None req-acef35f4-b906-4efb-ae97-e8cdc80e8e5a tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Task: {'id': task-1340876, 'name': ReconfigVM_Task, 'duration_secs': 0.805702} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.704885] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-acef35f4-b906-4efb-ae97-e8cdc80e8e5a tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] [instance: 5e96def2-0cbd-4bd9-93f4-6a365a0142b0] Reconfigured VM instance instance-00000049 to attach disk [datastore1] 5e96def2-0cbd-4bd9-93f4-6a365a0142b0/5e96def2-0cbd-4bd9-93f4-6a365a0142b0.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 834.705529] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-148d66e6-1b33-49ba-8d6c-6d821fa10f79 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.713543] env[62066]: DEBUG oslo_vmware.api [None req-acef35f4-b906-4efb-ae97-e8cdc80e8e5a tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Waiting for the task: (returnval){ [ 834.713543] env[62066]: value = "task-1340882" [ 834.713543] env[62066]: _type = "Task" [ 834.713543] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.721808] env[62066]: DEBUG oslo_vmware.api [None req-acef35f4-b906-4efb-ae97-e8cdc80e8e5a tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Task: {'id': task-1340882, 'name': Rename_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.755962] env[62066]: DEBUG oslo_concurrency.lockutils [None req-32db9ed7-c319-413e-978b-0abcfac2cab3 tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.859s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 834.760551] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 7.105s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 834.760757] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 834.760918] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62066) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 834.761334] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e3db032e-002b-4e25-a569-0fed0f11200b tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.324s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 834.764681] env[62066]: INFO nova.compute.claims [None req-e3db032e-002b-4e25-a569-0fed0f11200b tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: b08f27b1-429e-4d05-9a95-f45fd0cb2904] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 834.766153] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd296d87-2a92-4158-af43-b7c7aba60966 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.776360] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16e4156c-07f6-4c71-9975-c6834ff879e6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.795751] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec54faac-d786-49a5-a93d-2d8d1adfb589 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.803747] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27f0ab09-15ef-41da-9dc8-c6f2716e957d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.811356] env[62066]: INFO nova.scheduler.client.report [None req-32db9ed7-c319-413e-978b-0abcfac2cab3 tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] Deleted allocations for instance 050040cb-6ee3-4ad2-960a-fcebb53ac394 [ 834.842578] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181435MB free_disk=153GB free_vcpus=48 pci_devices=None {{(pid=62066) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 834.842818] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 834.874376] env[62066]: DEBUG oslo_vmware.api [None req-02667002-03be-4979-8517-c36490492214 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': task-1340877, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.872921} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.874984] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-02667002-03be-4979-8517-c36490492214 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk to [datastore1] 86a92b61-f3e9-48f9-8ee2-756669d558ef/86a92b61-f3e9-48f9-8ee2-756669d558ef.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 834.875340] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-02667002-03be-4979-8517-c36490492214 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 86a92b61-f3e9-48f9-8ee2-756669d558ef] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 834.875502] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8947d270-4acf-40b3-a46d-66b5e7790a0d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.889262] env[62066]: DEBUG oslo_vmware.api [None req-02667002-03be-4979-8517-c36490492214 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Waiting for the task: (returnval){ [ 834.889262] env[62066]: value = "task-1340883" [ 834.889262] env[62066]: _type = "Task" [ 834.889262] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.895735] env[62066]: DEBUG oslo_vmware.api [None req-02667002-03be-4979-8517-c36490492214 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': task-1340883, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.996434] env[62066]: DEBUG nova.network.neutron [req-97b5a9ff-1955-48e1-b324-f1f25741440b req-8fd77951-be69-4a8c-8790-6001cd148a56 service nova] [instance: 5e96def2-0cbd-4bd9-93f4-6a365a0142b0] Updated VIF entry in instance network info cache for port 0932c7ac-630e-4cef-8991-45b438f39483. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 834.996818] env[62066]: DEBUG nova.network.neutron [req-97b5a9ff-1955-48e1-b324-f1f25741440b req-8fd77951-be69-4a8c-8790-6001cd148a56 service nova] [instance: 5e96def2-0cbd-4bd9-93f4-6a365a0142b0] Updating instance_info_cache with network_info: [{"id": "0932c7ac-630e-4cef-8991-45b438f39483", "address": "fa:16:3e:ca:08:8d", "network": {"id": "fac1b86c-746c-4b6e-8d7f-21fe6ed404a3", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1827773279-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "59855b1cf91d43e3ba0482f414c4e546", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "88651df2-0506-4f6c-b868-dd30a81f2b1c", "external-id": "nsx-vlan-transportzone-366", "segmentation_id": 366, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0932c7ac-63", "ovs_interfaceid": "0932c7ac-630e-4cef-8991-45b438f39483", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 835.019154] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-28a644a7-dc09-4460-bd7c-0b89c8effe40 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 88daaaa6-f385-4161-bc74-9fffabb1145d] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 835.019376] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-28a644a7-dc09-4460-bd7c-0b89c8effe40 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 88daaaa6-f385-4161-bc74-9fffabb1145d] Deleting contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 835.019663] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-28a644a7-dc09-4460-bd7c-0b89c8effe40 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Deleting the datastore file [datastore2] 88daaaa6-f385-4161-bc74-9fffabb1145d {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 835.019942] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5a556d01-f9a6-44a8-83fb-beb35cc9a578 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.027966] env[62066]: DEBUG oslo_vmware.api [None req-28a644a7-dc09-4460-bd7c-0b89c8effe40 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Waiting for the task: (returnval){ [ 835.027966] env[62066]: value = "task-1340884" [ 835.027966] env[62066]: _type = "Task" [ 835.027966] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.038525] env[62066]: DEBUG oslo_vmware.api [None req-28a644a7-dc09-4460-bd7c-0b89c8effe40 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1340884, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.177223] env[62066]: DEBUG oslo_vmware.api [None req-d2edd17d-ac3d-4b81-90f1-278c62ed416c tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Task: {'id': task-1340881, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.226800] env[62066]: DEBUG oslo_vmware.api [None req-acef35f4-b906-4efb-ae97-e8cdc80e8e5a tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Task: {'id': task-1340882, 'name': Rename_Task, 'duration_secs': 0.188125} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.230024] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-acef35f4-b906-4efb-ae97-e8cdc80e8e5a tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] [instance: 5e96def2-0cbd-4bd9-93f4-6a365a0142b0] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 835.230024] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-29c7d311-ccb3-4292-bf08-23cb47f58cef {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.233620] env[62066]: DEBUG oslo_vmware.api [None req-acef35f4-b906-4efb-ae97-e8cdc80e8e5a tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Waiting for the task: (returnval){ [ 835.233620] env[62066]: value = "task-1340885" [ 835.233620] env[62066]: _type = "Task" [ 835.233620] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.245220] env[62066]: DEBUG oslo_vmware.api [None req-acef35f4-b906-4efb-ae97-e8cdc80e8e5a tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Task: {'id': task-1340885, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.353229] env[62066]: DEBUG oslo_concurrency.lockutils [None req-32db9ed7-c319-413e-978b-0abcfac2cab3 tempest-ServersTestMultiNic-1729475152 tempest-ServersTestMultiNic-1729475152-project-member] Lock "050040cb-6ee3-4ad2-960a-fcebb53ac394" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.765s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 835.397729] env[62066]: DEBUG oslo_vmware.api [None req-02667002-03be-4979-8517-c36490492214 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': task-1340883, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070774} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.398343] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-02667002-03be-4979-8517-c36490492214 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 86a92b61-f3e9-48f9-8ee2-756669d558ef] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 835.399447] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-879f8385-504e-4fcb-b9d4-785b1f323784 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.424355] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-02667002-03be-4979-8517-c36490492214 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 86a92b61-f3e9-48f9-8ee2-756669d558ef] Reconfiguring VM instance instance-0000004a to attach disk [datastore1] 86a92b61-f3e9-48f9-8ee2-756669d558ef/86a92b61-f3e9-48f9-8ee2-756669d558ef.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 835.424757] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-05d814b8-1b12-4a2a-8521-3a2015cd59f5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.445775] env[62066]: DEBUG oslo_vmware.api [None req-02667002-03be-4979-8517-c36490492214 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Waiting for the task: (returnval){ [ 835.445775] env[62066]: value = "task-1340886" [ 835.445775] env[62066]: _type = "Task" [ 835.445775] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.453836] env[62066]: DEBUG oslo_vmware.api [None req-02667002-03be-4979-8517-c36490492214 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': task-1340886, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.499805] env[62066]: DEBUG oslo_concurrency.lockutils [req-97b5a9ff-1955-48e1-b324-f1f25741440b req-8fd77951-be69-4a8c-8790-6001cd148a56 service nova] Releasing lock "refresh_cache-5e96def2-0cbd-4bd9-93f4-6a365a0142b0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 835.500983] env[62066]: DEBUG nova.compute.manager [req-97b5a9ff-1955-48e1-b324-f1f25741440b req-8fd77951-be69-4a8c-8790-6001cd148a56 service nova] [instance: 86a92b61-f3e9-48f9-8ee2-756669d558ef] Received event network-vif-plugged-fd69e6c7-721d-4947-80e1-4db6e9a230b0 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 835.500983] env[62066]: DEBUG oslo_concurrency.lockutils [req-97b5a9ff-1955-48e1-b324-f1f25741440b req-8fd77951-be69-4a8c-8790-6001cd148a56 service nova] Acquiring lock "86a92b61-f3e9-48f9-8ee2-756669d558ef-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 835.500983] env[62066]: DEBUG oslo_concurrency.lockutils [req-97b5a9ff-1955-48e1-b324-f1f25741440b req-8fd77951-be69-4a8c-8790-6001cd148a56 service nova] Lock "86a92b61-f3e9-48f9-8ee2-756669d558ef-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 835.500983] env[62066]: DEBUG oslo_concurrency.lockutils [req-97b5a9ff-1955-48e1-b324-f1f25741440b req-8fd77951-be69-4a8c-8790-6001cd148a56 service nova] Lock "86a92b61-f3e9-48f9-8ee2-756669d558ef-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 835.500983] env[62066]: DEBUG nova.compute.manager [req-97b5a9ff-1955-48e1-b324-f1f25741440b req-8fd77951-be69-4a8c-8790-6001cd148a56 service nova] [instance: 86a92b61-f3e9-48f9-8ee2-756669d558ef] No waiting events found dispatching network-vif-plugged-fd69e6c7-721d-4947-80e1-4db6e9a230b0 {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 835.501241] env[62066]: WARNING nova.compute.manager [req-97b5a9ff-1955-48e1-b324-f1f25741440b req-8fd77951-be69-4a8c-8790-6001cd148a56 service nova] [instance: 86a92b61-f3e9-48f9-8ee2-756669d558ef] Received unexpected event network-vif-plugged-fd69e6c7-721d-4947-80e1-4db6e9a230b0 for instance with vm_state building and task_state spawning. [ 835.501423] env[62066]: DEBUG nova.compute.manager [req-97b5a9ff-1955-48e1-b324-f1f25741440b req-8fd77951-be69-4a8c-8790-6001cd148a56 service nova] [instance: 86a92b61-f3e9-48f9-8ee2-756669d558ef] Received event network-changed-fd69e6c7-721d-4947-80e1-4db6e9a230b0 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 835.501500] env[62066]: DEBUG nova.compute.manager [req-97b5a9ff-1955-48e1-b324-f1f25741440b req-8fd77951-be69-4a8c-8790-6001cd148a56 service nova] [instance: 86a92b61-f3e9-48f9-8ee2-756669d558ef] Refreshing instance network info cache due to event network-changed-fd69e6c7-721d-4947-80e1-4db6e9a230b0. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 835.501651] env[62066]: DEBUG oslo_concurrency.lockutils [req-97b5a9ff-1955-48e1-b324-f1f25741440b req-8fd77951-be69-4a8c-8790-6001cd148a56 service nova] Acquiring lock "refresh_cache-86a92b61-f3e9-48f9-8ee2-756669d558ef" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 835.501828] env[62066]: DEBUG oslo_concurrency.lockutils [req-97b5a9ff-1955-48e1-b324-f1f25741440b req-8fd77951-be69-4a8c-8790-6001cd148a56 service nova] Acquired lock "refresh_cache-86a92b61-f3e9-48f9-8ee2-756669d558ef" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 835.502016] env[62066]: DEBUG nova.network.neutron [req-97b5a9ff-1955-48e1-b324-f1f25741440b req-8fd77951-be69-4a8c-8790-6001cd148a56 service nova] [instance: 86a92b61-f3e9-48f9-8ee2-756669d558ef] Refreshing network info cache for port fd69e6c7-721d-4947-80e1-4db6e9a230b0 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 835.539603] env[62066]: DEBUG oslo_vmware.api [None req-28a644a7-dc09-4460-bd7c-0b89c8effe40 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1340884, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.450986} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.539930] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-28a644a7-dc09-4460-bd7c-0b89c8effe40 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 835.540066] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-28a644a7-dc09-4460-bd7c-0b89c8effe40 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 88daaaa6-f385-4161-bc74-9fffabb1145d] Deleted contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 835.540249] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-28a644a7-dc09-4460-bd7c-0b89c8effe40 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 88daaaa6-f385-4161-bc74-9fffabb1145d] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 835.678440] env[62066]: DEBUG oslo_vmware.api [None req-d2edd17d-ac3d-4b81-90f1-278c62ed416c tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Task: {'id': task-1340881, 'name': PowerOnVM_Task, 'duration_secs': 0.523613} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.678812] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-d2edd17d-ac3d-4b81-90f1-278c62ed416c tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] [instance: de2e2be7-efdb-45a8-842a-640ab9deb1d9] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 835.679071] env[62066]: INFO nova.compute.manager [None req-d2edd17d-ac3d-4b81-90f1-278c62ed416c tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] [instance: de2e2be7-efdb-45a8-842a-640ab9deb1d9] Took 11.29 seconds to spawn the instance on the hypervisor. [ 835.679317] env[62066]: DEBUG nova.compute.manager [None req-d2edd17d-ac3d-4b81-90f1-278c62ed416c tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] [instance: de2e2be7-efdb-45a8-842a-640ab9deb1d9] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 835.680241] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23820fa7-dcbf-4a88-a37d-2ce608ee89bf {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.745738] env[62066]: DEBUG oslo_vmware.api [None req-acef35f4-b906-4efb-ae97-e8cdc80e8e5a tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Task: {'id': task-1340885, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.956359] env[62066]: DEBUG oslo_vmware.api [None req-02667002-03be-4979-8517-c36490492214 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': task-1340886, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.012053] env[62066]: DEBUG nova.network.neutron [-] [instance: 5f645eaa-6edc-4362-82ad-38c4c57b2be2] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 836.060954] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5519cede-8245-49b8-a685-7c12c870135a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.069863] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cec9fb3-fac4-4b08-a0a7-110a5acbef41 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.113406] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-451a2e4b-6348-428a-a6b0-79b6f108d2f1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.125738] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1cdd577-7b07-4411-8799-2453906aa66a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.142181] env[62066]: DEBUG nova.compute.provider_tree [None req-e3db032e-002b-4e25-a569-0fed0f11200b tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 836.204721] env[62066]: INFO nova.compute.manager [None req-d2edd17d-ac3d-4b81-90f1-278c62ed416c tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] [instance: de2e2be7-efdb-45a8-842a-640ab9deb1d9] Took 29.33 seconds to build instance. [ 836.247684] env[62066]: DEBUG oslo_vmware.api [None req-acef35f4-b906-4efb-ae97-e8cdc80e8e5a tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Task: {'id': task-1340885, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.272543] env[62066]: DEBUG nova.compute.manager [None req-838d4b27-f790-4e25-a20b-eab8ecc5b3e3 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] [instance: 1a4b0637-1a56-41ef-b89b-6b56d24ed206] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 836.272543] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-980c900a-eea6-4e36-8771-8a993b3e57ff {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.301571] env[62066]: DEBUG nova.compute.manager [req-5d265556-953a-49e1-9af3-740537e40a31 req-6139570c-9134-4e6f-8fce-56ccff073c2b service nova] [instance: 5f645eaa-6edc-4362-82ad-38c4c57b2be2] Received event network-vif-deleted-65fcb910-75be-4a68-8cec-476f7cc7cd00 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 836.456920] env[62066]: DEBUG oslo_vmware.api [None req-02667002-03be-4979-8517-c36490492214 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': task-1340886, 'name': ReconfigVM_Task, 'duration_secs': 0.56716} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.457290] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-02667002-03be-4979-8517-c36490492214 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 86a92b61-f3e9-48f9-8ee2-756669d558ef] Reconfigured VM instance instance-0000004a to attach disk [datastore1] 86a92b61-f3e9-48f9-8ee2-756669d558ef/86a92b61-f3e9-48f9-8ee2-756669d558ef.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 836.458034] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f06a7492-e0a2-41a4-b071-382bd8c8c84b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.465344] env[62066]: DEBUG oslo_vmware.api [None req-02667002-03be-4979-8517-c36490492214 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Waiting for the task: (returnval){ [ 836.465344] env[62066]: value = "task-1340887" [ 836.465344] env[62066]: _type = "Task" [ 836.465344] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.482461] env[62066]: DEBUG oslo_vmware.api [None req-02667002-03be-4979-8517-c36490492214 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': task-1340887, 'name': Rename_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.514957] env[62066]: INFO nova.compute.manager [-] [instance: 5f645eaa-6edc-4362-82ad-38c4c57b2be2] Took 1.99 seconds to deallocate network for instance. [ 836.524076] env[62066]: DEBUG nova.network.neutron [req-97b5a9ff-1955-48e1-b324-f1f25741440b req-8fd77951-be69-4a8c-8790-6001cd148a56 service nova] [instance: 86a92b61-f3e9-48f9-8ee2-756669d558ef] Updated VIF entry in instance network info cache for port fd69e6c7-721d-4947-80e1-4db6e9a230b0. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 836.524076] env[62066]: DEBUG nova.network.neutron [req-97b5a9ff-1955-48e1-b324-f1f25741440b req-8fd77951-be69-4a8c-8790-6001cd148a56 service nova] [instance: 86a92b61-f3e9-48f9-8ee2-756669d558ef] Updating instance_info_cache with network_info: [{"id": "fd69e6c7-721d-4947-80e1-4db6e9a230b0", "address": "fa:16:3e:ac:f0:6d", "network": {"id": "9f3396ce-ee48-4eaf-8cc7-3e1711f37537", "bridge": "br-int", "label": "tempest-ServersTestJSON-2129775093-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6497ab02f327476d8ff81c2ecc0371e5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7ab8d568-adb0-4f3b-b6cc-68413e6546ae", "external-id": "nsx-vlan-transportzone-86", "segmentation_id": 86, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfd69e6c7-72", "ovs_interfaceid": "fd69e6c7-721d-4947-80e1-4db6e9a230b0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 836.577773] env[62066]: DEBUG nova.virt.hardware [None req-28a644a7-dc09-4460-bd7c-0b89c8effe40 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-23T13:40:41Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-23T13:40:23Z,direct_url=,disk_format='vmdk',id=50ff584c-3b50-4395-af07-3e66769bc9f7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='eb52f7069a374c61ae946f052007c6d9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-23T13:40:24Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 836.578213] env[62066]: DEBUG nova.virt.hardware [None req-28a644a7-dc09-4460-bd7c-0b89c8effe40 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 836.578417] env[62066]: DEBUG nova.virt.hardware [None req-28a644a7-dc09-4460-bd7c-0b89c8effe40 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 836.578623] env[62066]: DEBUG nova.virt.hardware [None req-28a644a7-dc09-4460-bd7c-0b89c8effe40 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 836.578776] env[62066]: DEBUG nova.virt.hardware [None req-28a644a7-dc09-4460-bd7c-0b89c8effe40 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 836.578923] env[62066]: DEBUG nova.virt.hardware [None req-28a644a7-dc09-4460-bd7c-0b89c8effe40 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 836.579179] env[62066]: DEBUG nova.virt.hardware [None req-28a644a7-dc09-4460-bd7c-0b89c8effe40 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 836.579352] env[62066]: DEBUG nova.virt.hardware [None req-28a644a7-dc09-4460-bd7c-0b89c8effe40 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 836.579559] env[62066]: DEBUG nova.virt.hardware [None req-28a644a7-dc09-4460-bd7c-0b89c8effe40 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 836.579973] env[62066]: DEBUG nova.virt.hardware [None req-28a644a7-dc09-4460-bd7c-0b89c8effe40 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 836.580273] env[62066]: DEBUG nova.virt.hardware [None req-28a644a7-dc09-4460-bd7c-0b89c8effe40 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 836.582273] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-633806fd-7a04-438f-ac95-fa548b4fc917 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.592777] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-498e58de-e10a-4909-98c2-13603e70f9d9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.610390] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-28a644a7-dc09-4460-bd7c-0b89c8effe40 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 88daaaa6-f385-4161-bc74-9fffabb1145d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:54:27:b5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a965790c-2d2f-4c2a-9ee7-745f4d53039b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ba002f90-5218-46f2-b7e1-cd96ae3018c7', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 836.620405] env[62066]: DEBUG oslo.service.loopingcall [None req-28a644a7-dc09-4460-bd7c-0b89c8effe40 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 836.620569] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 88daaaa6-f385-4161-bc74-9fffabb1145d] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 836.620886] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a92c7cae-a858-4760-8963-2f28ce53a26d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.646197] env[62066]: DEBUG nova.scheduler.client.report [None req-e3db032e-002b-4e25-a569-0fed0f11200b tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 836.651486] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 836.651486] env[62066]: value = "task-1340888" [ 836.651486] env[62066]: _type = "Task" [ 836.651486] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.659881] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1340888, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.706966] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d2edd17d-ac3d-4b81-90f1-278c62ed416c tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Lock "de2e2be7-efdb-45a8-842a-640ab9deb1d9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 86.944s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 836.752158] env[62066]: DEBUG oslo_vmware.api [None req-acef35f4-b906-4efb-ae97-e8cdc80e8e5a tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Task: {'id': task-1340885, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.784089] env[62066]: INFO nova.compute.manager [None req-838d4b27-f790-4e25-a20b-eab8ecc5b3e3 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] [instance: 1a4b0637-1a56-41ef-b89b-6b56d24ed206] instance snapshotting [ 836.788019] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ad88830-4220-4778-84e2-7224734e6293 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.819533] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1433e30c-5f95-4f43-b88d-d9a23c5962d7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.982183] env[62066]: DEBUG oslo_vmware.api [None req-02667002-03be-4979-8517-c36490492214 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': task-1340887, 'name': Rename_Task, 'duration_secs': 0.322694} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.982183] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-02667002-03be-4979-8517-c36490492214 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 86a92b61-f3e9-48f9-8ee2-756669d558ef] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 836.982183] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-161a5b86-1081-4bbd-8393-10671134703d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.987835] env[62066]: DEBUG oslo_vmware.api [None req-02667002-03be-4979-8517-c36490492214 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Waiting for the task: (returnval){ [ 836.987835] env[62066]: value = "task-1340889" [ 836.987835] env[62066]: _type = "Task" [ 836.987835] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.998975] env[62066]: DEBUG oslo_vmware.api [None req-02667002-03be-4979-8517-c36490492214 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': task-1340889, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.026699] env[62066]: DEBUG oslo_concurrency.lockutils [None req-6075adee-b71b-4aa1-9781-77a864c59e59 tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 837.027213] env[62066]: DEBUG oslo_concurrency.lockutils [req-97b5a9ff-1955-48e1-b324-f1f25741440b req-8fd77951-be69-4a8c-8790-6001cd148a56 service nova] Releasing lock "refresh_cache-86a92b61-f3e9-48f9-8ee2-756669d558ef" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 837.027463] env[62066]: DEBUG nova.compute.manager [req-97b5a9ff-1955-48e1-b324-f1f25741440b req-8fd77951-be69-4a8c-8790-6001cd148a56 service nova] [instance: 8ff6fbb9-c90f-498d-9a85-d220a8c2f794] Received event network-changed-8c324833-7bf9-4007-85aa-5b16871f63c7 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 837.027758] env[62066]: DEBUG nova.compute.manager [req-97b5a9ff-1955-48e1-b324-f1f25741440b req-8fd77951-be69-4a8c-8790-6001cd148a56 service nova] [instance: 8ff6fbb9-c90f-498d-9a85-d220a8c2f794] Refreshing instance network info cache due to event network-changed-8c324833-7bf9-4007-85aa-5b16871f63c7. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 837.027942] env[62066]: DEBUG oslo_concurrency.lockutils [req-97b5a9ff-1955-48e1-b324-f1f25741440b req-8fd77951-be69-4a8c-8790-6001cd148a56 service nova] Acquiring lock "refresh_cache-8ff6fbb9-c90f-498d-9a85-d220a8c2f794" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 837.028418] env[62066]: DEBUG oslo_concurrency.lockutils [req-97b5a9ff-1955-48e1-b324-f1f25741440b req-8fd77951-be69-4a8c-8790-6001cd148a56 service nova] Acquired lock "refresh_cache-8ff6fbb9-c90f-498d-9a85-d220a8c2f794" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 837.028531] env[62066]: DEBUG nova.network.neutron [req-97b5a9ff-1955-48e1-b324-f1f25741440b req-8fd77951-be69-4a8c-8790-6001cd148a56 service nova] [instance: 8ff6fbb9-c90f-498d-9a85-d220a8c2f794] Refreshing network info cache for port 8c324833-7bf9-4007-85aa-5b16871f63c7 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 837.152125] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e3db032e-002b-4e25-a569-0fed0f11200b tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.390s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 837.153508] env[62066]: DEBUG nova.compute.manager [None req-e3db032e-002b-4e25-a569-0fed0f11200b tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: b08f27b1-429e-4d05-9a95-f45fd0cb2904] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 837.158114] env[62066]: DEBUG oslo_concurrency.lockutils [None req-8589d93a-2555-4139-b6bf-f9de6865cf74 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.623s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 837.158782] env[62066]: DEBUG nova.objects.instance [None req-8589d93a-2555-4139-b6bf-f9de6865cf74 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Lazy-loading 'resources' on Instance uuid 49f46244-34fa-48a1-95a2-8e95850f345d {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 837.174382] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1340888, 'name': CreateVM_Task, 'duration_secs': 0.365697} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.174382] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 88daaaa6-f385-4161-bc74-9fffabb1145d] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 837.174526] env[62066]: DEBUG oslo_concurrency.lockutils [None req-28a644a7-dc09-4460-bd7c-0b89c8effe40 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 837.174651] env[62066]: DEBUG oslo_concurrency.lockutils [None req-28a644a7-dc09-4460-bd7c-0b89c8effe40 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Acquired lock "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 837.176224] env[62066]: DEBUG oslo_concurrency.lockutils [None req-28a644a7-dc09-4460-bd7c-0b89c8effe40 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 837.176224] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7d229f92-5318-44e5-b136-b3c803c64a4d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.181727] env[62066]: DEBUG oslo_vmware.api [None req-28a644a7-dc09-4460-bd7c-0b89c8effe40 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Waiting for the task: (returnval){ [ 837.181727] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]52557d17-d118-61c9-585c-0977178a75e6" [ 837.181727] env[62066]: _type = "Task" [ 837.181727] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.194147] env[62066]: DEBUG oslo_vmware.api [None req-28a644a7-dc09-4460-bd7c-0b89c8effe40 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52557d17-d118-61c9-585c-0977178a75e6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.247961] env[62066]: DEBUG oslo_vmware.api [None req-acef35f4-b906-4efb-ae97-e8cdc80e8e5a tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Task: {'id': task-1340885, 'name': PowerOnVM_Task, 'duration_secs': 1.887918} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.251524] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-acef35f4-b906-4efb-ae97-e8cdc80e8e5a tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] [instance: 5e96def2-0cbd-4bd9-93f4-6a365a0142b0] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 837.251524] env[62066]: INFO nova.compute.manager [None req-acef35f4-b906-4efb-ae97-e8cdc80e8e5a tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] [instance: 5e96def2-0cbd-4bd9-93f4-6a365a0142b0] Took 10.45 seconds to spawn the instance on the hypervisor. [ 837.251524] env[62066]: DEBUG nova.compute.manager [None req-acef35f4-b906-4efb-ae97-e8cdc80e8e5a tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] [instance: 5e96def2-0cbd-4bd9-93f4-6a365a0142b0] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 837.251524] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f13295f-c61b-4b23-a9db-a321f5261c28 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.283650] env[62066]: DEBUG nova.compute.manager [req-b1aa4ea4-13cb-40e7-808e-2515c2ce9cda req-aebb7172-322a-4e02-928b-0d82d01f1571 service nova] [instance: de2e2be7-efdb-45a8-842a-640ab9deb1d9] Received event network-changed-c833a993-4af0-4bdb-90d2-cc82a796a560 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 837.283650] env[62066]: DEBUG nova.compute.manager [req-b1aa4ea4-13cb-40e7-808e-2515c2ce9cda req-aebb7172-322a-4e02-928b-0d82d01f1571 service nova] [instance: de2e2be7-efdb-45a8-842a-640ab9deb1d9] Refreshing instance network info cache due to event network-changed-c833a993-4af0-4bdb-90d2-cc82a796a560. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 837.284263] env[62066]: DEBUG oslo_concurrency.lockutils [req-b1aa4ea4-13cb-40e7-808e-2515c2ce9cda req-aebb7172-322a-4e02-928b-0d82d01f1571 service nova] Acquiring lock "refresh_cache-de2e2be7-efdb-45a8-842a-640ab9deb1d9" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 837.284263] env[62066]: DEBUG oslo_concurrency.lockutils [req-b1aa4ea4-13cb-40e7-808e-2515c2ce9cda req-aebb7172-322a-4e02-928b-0d82d01f1571 service nova] Acquired lock "refresh_cache-de2e2be7-efdb-45a8-842a-640ab9deb1d9" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 837.284263] env[62066]: DEBUG nova.network.neutron [req-b1aa4ea4-13cb-40e7-808e-2515c2ce9cda req-aebb7172-322a-4e02-928b-0d82d01f1571 service nova] [instance: de2e2be7-efdb-45a8-842a-640ab9deb1d9] Refreshing network info cache for port c833a993-4af0-4bdb-90d2-cc82a796a560 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 837.332427] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-838d4b27-f790-4e25-a20b-eab8ecc5b3e3 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] [instance: 1a4b0637-1a56-41ef-b89b-6b56d24ed206] Creating Snapshot of the VM instance {{(pid=62066) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 837.332427] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-87c4793a-fc7c-4747-9c58-0bb4b1a6786e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.341254] env[62066]: DEBUG oslo_vmware.api [None req-838d4b27-f790-4e25-a20b-eab8ecc5b3e3 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Waiting for the task: (returnval){ [ 837.341254] env[62066]: value = "task-1340890" [ 837.341254] env[62066]: _type = "Task" [ 837.341254] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.353453] env[62066]: DEBUG oslo_vmware.api [None req-838d4b27-f790-4e25-a20b-eab8ecc5b3e3 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Task: {'id': task-1340890, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.467505] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f63339cf-eb8a-4d19-b0c5-1ac7744e7482 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Acquiring lock "d77e3405-d158-4a8c-9f54-d5fbd26fed48" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 837.467505] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f63339cf-eb8a-4d19-b0c5-1ac7744e7482 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Lock "d77e3405-d158-4a8c-9f54-d5fbd26fed48" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 837.502222] env[62066]: DEBUG oslo_vmware.api [None req-02667002-03be-4979-8517-c36490492214 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': task-1340889, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.672818] env[62066]: DEBUG nova.compute.utils [None req-e3db032e-002b-4e25-a569-0fed0f11200b tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 837.676451] env[62066]: DEBUG nova.compute.manager [None req-e3db032e-002b-4e25-a569-0fed0f11200b tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: b08f27b1-429e-4d05-9a95-f45fd0cb2904] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 837.676620] env[62066]: DEBUG nova.network.neutron [None req-e3db032e-002b-4e25-a569-0fed0f11200b tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: b08f27b1-429e-4d05-9a95-f45fd0cb2904] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 837.699783] env[62066]: DEBUG oslo_vmware.api [None req-28a644a7-dc09-4460-bd7c-0b89c8effe40 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52557d17-d118-61c9-585c-0977178a75e6, 'name': SearchDatastore_Task, 'duration_secs': 0.017545} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.699946] env[62066]: DEBUG oslo_concurrency.lockutils [None req-28a644a7-dc09-4460-bd7c-0b89c8effe40 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Releasing lock "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 837.700698] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-28a644a7-dc09-4460-bd7c-0b89c8effe40 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 88daaaa6-f385-4161-bc74-9fffabb1145d] Processing image 50ff584c-3b50-4395-af07-3e66769bc9f7 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 837.700698] env[62066]: DEBUG oslo_concurrency.lockutils [None req-28a644a7-dc09-4460-bd7c-0b89c8effe40 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 837.700698] env[62066]: DEBUG oslo_concurrency.lockutils [None req-28a644a7-dc09-4460-bd7c-0b89c8effe40 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Acquired lock "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 837.700698] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-28a644a7-dc09-4460-bd7c-0b89c8effe40 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 837.704060] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-22a8df6c-bd26-4ce6-aeb3-57ec2f7b4965 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.715125] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-28a644a7-dc09-4460-bd7c-0b89c8effe40 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 837.715995] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-28a644a7-dc09-4460-bd7c-0b89c8effe40 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 837.716306] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ef64ef4c-a655-44dc-a9de-82cadb72ec24 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.728610] env[62066]: DEBUG oslo_vmware.api [None req-28a644a7-dc09-4460-bd7c-0b89c8effe40 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Waiting for the task: (returnval){ [ 837.728610] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]52db4aa2-e9c3-aeb6-41c0-a75fd5265916" [ 837.728610] env[62066]: _type = "Task" [ 837.728610] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.736014] env[62066]: DEBUG oslo_concurrency.lockutils [None req-22d0b1dd-f08e-49e4-a262-9fed18806112 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Acquiring lock "3120c9a5-0558-4e83-b0fb-994a989398fb" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 837.736730] env[62066]: DEBUG oslo_concurrency.lockutils [None req-22d0b1dd-f08e-49e4-a262-9fed18806112 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Lock "3120c9a5-0558-4e83-b0fb-994a989398fb" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 837.748969] env[62066]: DEBUG oslo_vmware.api [None req-28a644a7-dc09-4460-bd7c-0b89c8effe40 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52db4aa2-e9c3-aeb6-41c0-a75fd5265916, 'name': SearchDatastore_Task, 'duration_secs': 0.010872} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.749467] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-39f3d05d-22f9-42e1-a150-200d97f4a98a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.755741] env[62066]: DEBUG oslo_vmware.api [None req-28a644a7-dc09-4460-bd7c-0b89c8effe40 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Waiting for the task: (returnval){ [ 837.755741] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]528f8611-d2fb-8bd0-ee60-633f8a1bf74d" [ 837.755741] env[62066]: _type = "Task" [ 837.755741] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.780078] env[62066]: DEBUG oslo_vmware.api [None req-28a644a7-dc09-4460-bd7c-0b89c8effe40 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]528f8611-d2fb-8bd0-ee60-633f8a1bf74d, 'name': SearchDatastore_Task, 'duration_secs': 0.012694} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.789912] env[62066]: DEBUG oslo_concurrency.lockutils [None req-28a644a7-dc09-4460-bd7c-0b89c8effe40 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Releasing lock "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 837.790047] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-28a644a7-dc09-4460-bd7c-0b89c8effe40 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk to [datastore1] 88daaaa6-f385-4161-bc74-9fffabb1145d/88daaaa6-f385-4161-bc74-9fffabb1145d.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 837.793541] env[62066]: INFO nova.compute.manager [None req-acef35f4-b906-4efb-ae97-e8cdc80e8e5a tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] [instance: 5e96def2-0cbd-4bd9-93f4-6a365a0142b0] Took 28.15 seconds to build instance. [ 837.794734] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f3b1630f-6099-4470-abe2-2b6bf09c1d56 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.807147] env[62066]: DEBUG oslo_vmware.api [None req-28a644a7-dc09-4460-bd7c-0b89c8effe40 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Waiting for the task: (returnval){ [ 837.807147] env[62066]: value = "task-1340891" [ 837.807147] env[62066]: _type = "Task" [ 837.807147] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.821422] env[62066]: DEBUG oslo_vmware.api [None req-28a644a7-dc09-4460-bd7c-0b89c8effe40 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1340891, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.864828] env[62066]: DEBUG oslo_vmware.api [None req-838d4b27-f790-4e25-a20b-eab8ecc5b3e3 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Task: {'id': task-1340890, 'name': CreateSnapshot_Task, 'duration_secs': 0.492873} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.865163] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-838d4b27-f790-4e25-a20b-eab8ecc5b3e3 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] [instance: 1a4b0637-1a56-41ef-b89b-6b56d24ed206] Created Snapshot of the VM instance {{(pid=62066) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 837.866116] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e082116-7cc4-4041-86f9-47309ae0e950 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.974117] env[62066]: DEBUG nova.compute.manager [None req-f63339cf-eb8a-4d19-b0c5-1ac7744e7482 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] [instance: d77e3405-d158-4a8c-9f54-d5fbd26fed48] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 837.997675] env[62066]: DEBUG nova.network.neutron [req-97b5a9ff-1955-48e1-b324-f1f25741440b req-8fd77951-be69-4a8c-8790-6001cd148a56 service nova] [instance: 8ff6fbb9-c90f-498d-9a85-d220a8c2f794] Updated VIF entry in instance network info cache for port 8c324833-7bf9-4007-85aa-5b16871f63c7. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 837.997783] env[62066]: DEBUG nova.network.neutron [req-97b5a9ff-1955-48e1-b324-f1f25741440b req-8fd77951-be69-4a8c-8790-6001cd148a56 service nova] [instance: 8ff6fbb9-c90f-498d-9a85-d220a8c2f794] Updating instance_info_cache with network_info: [{"id": "8c324833-7bf9-4007-85aa-5b16871f63c7", "address": "fa:16:3e:00:85:34", "network": {"id": "0e17b5e2-526e-4d4c-a463-5e34408044cf", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1588316661-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.184", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1f05d8a66d0c4a479e49a947e4b4bbff", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ffcecdaa-a7b8-49fc-9371-dbdb7744688e", "external-id": "nsx-vlan-transportzone-994", "segmentation_id": 994, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8c324833-7b", "ovs_interfaceid": "8c324833-7bf9-4007-85aa-5b16871f63c7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 838.008787] env[62066]: DEBUG oslo_vmware.api [None req-02667002-03be-4979-8517-c36490492214 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': task-1340889, 'name': PowerOnVM_Task, 'duration_secs': 0.827276} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.009065] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-02667002-03be-4979-8517-c36490492214 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 86a92b61-f3e9-48f9-8ee2-756669d558ef] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 838.009975] env[62066]: INFO nova.compute.manager [None req-02667002-03be-4979-8517-c36490492214 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 86a92b61-f3e9-48f9-8ee2-756669d558ef] Took 8.77 seconds to spawn the instance on the hypervisor. [ 838.009975] env[62066]: DEBUG nova.compute.manager [None req-02667002-03be-4979-8517-c36490492214 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 86a92b61-f3e9-48f9-8ee2-756669d558ef] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 838.011892] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ccb711a-59ea-4408-ae26-ed4c1af2d351 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.014841] env[62066]: DEBUG nova.policy [None req-e3db032e-002b-4e25-a569-0fed0f11200b tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '60e2bc61839b4299912c53137bc23fbc', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '75f0352c852947369474c9e05766c584', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 838.056884] env[62066]: INFO nova.compute.manager [None req-63dd5fc5-276d-49c0-99ba-71a261a8efa4 tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] Rebuilding instance [ 838.059546] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-407888a3-e349-487d-af96-9fe6d21a6001 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.072732] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9a5e246-164e-416e-8ddb-1bf0aad7e191 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.108416] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe924b50-d393-4bf9-9f41-a31075833eac {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.117220] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a4c083d-5d1a-41bf-8c13-d5d8d958c264 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.140226] env[62066]: DEBUG nova.compute.provider_tree [None req-8589d93a-2555-4139-b6bf-f9de6865cf74 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 838.149913] env[62066]: DEBUG nova.compute.manager [None req-63dd5fc5-276d-49c0-99ba-71a261a8efa4 tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 838.150850] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8055abaf-d1da-415f-8a43-f9cba2184ac6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.180969] env[62066]: DEBUG nova.compute.manager [None req-e3db032e-002b-4e25-a569-0fed0f11200b tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: b08f27b1-429e-4d05-9a95-f45fd0cb2904] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 838.241155] env[62066]: DEBUG nova.compute.manager [None req-22d0b1dd-f08e-49e4-a262-9fed18806112 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] [instance: 3120c9a5-0558-4e83-b0fb-994a989398fb] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 838.301205] env[62066]: DEBUG oslo_concurrency.lockutils [None req-acef35f4-b906-4efb-ae97-e8cdc80e8e5a tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Lock "5e96def2-0cbd-4bd9-93f4-6a365a0142b0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 45.194s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 838.314740] env[62066]: DEBUG nova.network.neutron [req-b1aa4ea4-13cb-40e7-808e-2515c2ce9cda req-aebb7172-322a-4e02-928b-0d82d01f1571 service nova] [instance: de2e2be7-efdb-45a8-842a-640ab9deb1d9] Updated VIF entry in instance network info cache for port c833a993-4af0-4bdb-90d2-cc82a796a560. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 838.315116] env[62066]: DEBUG nova.network.neutron [req-b1aa4ea4-13cb-40e7-808e-2515c2ce9cda req-aebb7172-322a-4e02-928b-0d82d01f1571 service nova] [instance: de2e2be7-efdb-45a8-842a-640ab9deb1d9] Updating instance_info_cache with network_info: [{"id": "c833a993-4af0-4bdb-90d2-cc82a796a560", "address": "fa:16:3e:71:ab:fe", "network": {"id": "4114ddfe-818f-47a5-bb6f-44a4c22517f6", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-676131672-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.157", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dd30584efc254610b3939d1ba3806693", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a485857d-7086-4dcf-9d65-d0dcd177fcb0", "external-id": "nsx-vlan-transportzone-232", "segmentation_id": 232, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc833a993-4a", "ovs_interfaceid": "c833a993-4af0-4bdb-90d2-cc82a796a560", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 838.320046] env[62066]: DEBUG oslo_vmware.api [None req-28a644a7-dc09-4460-bd7c-0b89c8effe40 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1340891, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.392321] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-838d4b27-f790-4e25-a20b-eab8ecc5b3e3 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] [instance: 1a4b0637-1a56-41ef-b89b-6b56d24ed206] Creating linked-clone VM from snapshot {{(pid=62066) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 838.392945] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-4b958010-30ef-4ba3-b9ed-37c29b21ebc6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.403187] env[62066]: DEBUG oslo_vmware.api [None req-838d4b27-f790-4e25-a20b-eab8ecc5b3e3 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Waiting for the task: (returnval){ [ 838.403187] env[62066]: value = "task-1340892" [ 838.403187] env[62066]: _type = "Task" [ 838.403187] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.413820] env[62066]: DEBUG oslo_vmware.api [None req-838d4b27-f790-4e25-a20b-eab8ecc5b3e3 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Task: {'id': task-1340892, 'name': CloneVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.460017] env[62066]: DEBUG nova.network.neutron [None req-e3db032e-002b-4e25-a569-0fed0f11200b tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: b08f27b1-429e-4d05-9a95-f45fd0cb2904] Successfully created port: 62d600d9-8ad3-43f8-a2b0-050747916fdb {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 838.505477] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f63339cf-eb8a-4d19-b0c5-1ac7744e7482 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 838.506025] env[62066]: DEBUG oslo_concurrency.lockutils [req-97b5a9ff-1955-48e1-b324-f1f25741440b req-8fd77951-be69-4a8c-8790-6001cd148a56 service nova] Releasing lock "refresh_cache-8ff6fbb9-c90f-498d-9a85-d220a8c2f794" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 838.506277] env[62066]: DEBUG nova.compute.manager [req-97b5a9ff-1955-48e1-b324-f1f25741440b req-8fd77951-be69-4a8c-8790-6001cd148a56 service nova] [instance: 2709d0ad-5dd8-4e3c-b1e6-3d1e2e242380] Received event network-changed-c5cdc663-1e55-4238-bad4-7fb406b34a36 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 838.507857] env[62066]: DEBUG nova.compute.manager [req-97b5a9ff-1955-48e1-b324-f1f25741440b req-8fd77951-be69-4a8c-8790-6001cd148a56 service nova] [instance: 2709d0ad-5dd8-4e3c-b1e6-3d1e2e242380] Refreshing instance network info cache due to event network-changed-c5cdc663-1e55-4238-bad4-7fb406b34a36. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 838.507857] env[62066]: DEBUG oslo_concurrency.lockutils [req-97b5a9ff-1955-48e1-b324-f1f25741440b req-8fd77951-be69-4a8c-8790-6001cd148a56 service nova] Acquiring lock "refresh_cache-2709d0ad-5dd8-4e3c-b1e6-3d1e2e242380" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 838.507857] env[62066]: DEBUG oslo_concurrency.lockutils [req-97b5a9ff-1955-48e1-b324-f1f25741440b req-8fd77951-be69-4a8c-8790-6001cd148a56 service nova] Acquired lock "refresh_cache-2709d0ad-5dd8-4e3c-b1e6-3d1e2e242380" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 838.507857] env[62066]: DEBUG nova.network.neutron [req-97b5a9ff-1955-48e1-b324-f1f25741440b req-8fd77951-be69-4a8c-8790-6001cd148a56 service nova] [instance: 2709d0ad-5dd8-4e3c-b1e6-3d1e2e242380] Refreshing network info cache for port c5cdc663-1e55-4238-bad4-7fb406b34a36 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 838.538027] env[62066]: INFO nova.compute.manager [None req-02667002-03be-4979-8517-c36490492214 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 86a92b61-f3e9-48f9-8ee2-756669d558ef] Took 27.16 seconds to build instance. [ 838.644349] env[62066]: DEBUG nova.scheduler.client.report [None req-8589d93a-2555-4139-b6bf-f9de6865cf74 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 838.664940] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-63dd5fc5-276d-49c0-99ba-71a261a8efa4 tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 838.665443] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4a105460-6d95-42e4-951b-741bafb8c305 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.674552] env[62066]: DEBUG oslo_vmware.api [None req-63dd5fc5-276d-49c0-99ba-71a261a8efa4 tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] Waiting for the task: (returnval){ [ 838.674552] env[62066]: value = "task-1340893" [ 838.674552] env[62066]: _type = "Task" [ 838.674552] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.686067] env[62066]: DEBUG oslo_vmware.api [None req-63dd5fc5-276d-49c0-99ba-71a261a8efa4 tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] Task: {'id': task-1340893, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.765862] env[62066]: DEBUG oslo_concurrency.lockutils [None req-22d0b1dd-f08e-49e4-a262-9fed18806112 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 838.820476] env[62066]: DEBUG oslo_vmware.api [None req-28a644a7-dc09-4460-bd7c-0b89c8effe40 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1340891, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.645556} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.820648] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-28a644a7-dc09-4460-bd7c-0b89c8effe40 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk to [datastore1] 88daaaa6-f385-4161-bc74-9fffabb1145d/88daaaa6-f385-4161-bc74-9fffabb1145d.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 838.820881] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-28a644a7-dc09-4460-bd7c-0b89c8effe40 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 88daaaa6-f385-4161-bc74-9fffabb1145d] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 838.821420] env[62066]: DEBUG oslo_concurrency.lockutils [req-b1aa4ea4-13cb-40e7-808e-2515c2ce9cda req-aebb7172-322a-4e02-928b-0d82d01f1571 service nova] Releasing lock "refresh_cache-de2e2be7-efdb-45a8-842a-640ab9deb1d9" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 838.821800] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4042190b-0a5d-490f-b6ef-689ed84a784a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.829511] env[62066]: DEBUG oslo_vmware.api [None req-28a644a7-dc09-4460-bd7c-0b89c8effe40 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Waiting for the task: (returnval){ [ 838.829511] env[62066]: value = "task-1340894" [ 838.829511] env[62066]: _type = "Task" [ 838.829511] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.848800] env[62066]: DEBUG oslo_vmware.api [None req-28a644a7-dc09-4460-bd7c-0b89c8effe40 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1340894, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.920021] env[62066]: DEBUG oslo_vmware.api [None req-838d4b27-f790-4e25-a20b-eab8ecc5b3e3 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Task: {'id': task-1340892, 'name': CloneVM_Task} progress is 94%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.042398] env[62066]: DEBUG oslo_concurrency.lockutils [None req-02667002-03be-4979-8517-c36490492214 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Lock "86a92b61-f3e9-48f9-8ee2-756669d558ef" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 34.054s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 839.127575] env[62066]: DEBUG oslo_concurrency.lockutils [None req-73e4bc46-8266-447d-984c-3ca2b99ffd23 tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] Acquiring lock "0bc31a8b-a5fc-4f6e-a4c0-024bf4260755" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 839.127970] env[62066]: DEBUG oslo_concurrency.lockutils [None req-73e4bc46-8266-447d-984c-3ca2b99ffd23 tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] Lock "0bc31a8b-a5fc-4f6e-a4c0-024bf4260755" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 839.152443] env[62066]: DEBUG oslo_concurrency.lockutils [None req-8589d93a-2555-4139-b6bf-f9de6865cf74 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.992s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 839.153876] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 4.311s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 839.184459] env[62066]: DEBUG oslo_vmware.api [None req-63dd5fc5-276d-49c0-99ba-71a261a8efa4 tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] Task: {'id': task-1340893, 'name': PowerOffVM_Task, 'duration_secs': 0.336055} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.185641] env[62066]: INFO nova.scheduler.client.report [None req-8589d93a-2555-4139-b6bf-f9de6865cf74 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Deleted allocations for instance 49f46244-34fa-48a1-95a2-8e95850f345d [ 839.186995] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-63dd5fc5-276d-49c0-99ba-71a261a8efa4 tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 839.187977] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-63dd5fc5-276d-49c0-99ba-71a261a8efa4 tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 839.188351] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1c2a1190-6a10-46ad-846d-dae9c3eac8a3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.194864] env[62066]: DEBUG nova.compute.manager [None req-e3db032e-002b-4e25-a569-0fed0f11200b tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: b08f27b1-429e-4d05-9a95-f45fd0cb2904] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 839.208585] env[62066]: DEBUG oslo_vmware.api [None req-63dd5fc5-276d-49c0-99ba-71a261a8efa4 tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] Waiting for the task: (returnval){ [ 839.208585] env[62066]: value = "task-1340895" [ 839.208585] env[62066]: _type = "Task" [ 839.208585] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.230112] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-63dd5fc5-276d-49c0-99ba-71a261a8efa4 tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] VM already powered off {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 839.230430] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-63dd5fc5-276d-49c0-99ba-71a261a8efa4 tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] Volume detach. Driver type: vmdk {{(pid=62066) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 839.230728] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-63dd5fc5-276d-49c0-99ba-71a261a8efa4 tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-285999', 'volume_id': '26ceca45-1cae-48c9-a567-e7cc4f9970c1', 'name': 'volume-26ceca45-1cae-48c9-a567-e7cc4f9970c1', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'daffaf51-4c45-44aa-8fc2-4db066a09971', 'attached_at': '', 'detached_at': '', 'volume_id': '26ceca45-1cae-48c9-a567-e7cc4f9970c1', 'serial': '26ceca45-1cae-48c9-a567-e7cc4f9970c1'} {{(pid=62066) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 839.232140] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ac4bf29-be9d-460b-99e8-48d81492d008 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.241935] env[62066]: DEBUG nova.virt.hardware [None req-e3db032e-002b-4e25-a569-0fed0f11200b tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-23T13:40:41Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-23T13:40:23Z,direct_url=,disk_format='vmdk',id=50ff584c-3b50-4395-af07-3e66769bc9f7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='eb52f7069a374c61ae946f052007c6d9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-23T13:40:24Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 839.242244] env[62066]: DEBUG nova.virt.hardware [None req-e3db032e-002b-4e25-a569-0fed0f11200b tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 839.242415] env[62066]: DEBUG nova.virt.hardware [None req-e3db032e-002b-4e25-a569-0fed0f11200b tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 839.243424] env[62066]: DEBUG nova.virt.hardware [None req-e3db032e-002b-4e25-a569-0fed0f11200b tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 839.243424] env[62066]: DEBUG nova.virt.hardware [None req-e3db032e-002b-4e25-a569-0fed0f11200b tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 839.243424] env[62066]: DEBUG nova.virt.hardware [None req-e3db032e-002b-4e25-a569-0fed0f11200b tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 839.243424] env[62066]: DEBUG nova.virt.hardware [None req-e3db032e-002b-4e25-a569-0fed0f11200b tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 839.243424] env[62066]: DEBUG nova.virt.hardware [None req-e3db032e-002b-4e25-a569-0fed0f11200b tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 839.243424] env[62066]: DEBUG nova.virt.hardware [None req-e3db032e-002b-4e25-a569-0fed0f11200b tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 839.243956] env[62066]: DEBUG nova.virt.hardware [None req-e3db032e-002b-4e25-a569-0fed0f11200b tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 839.243956] env[62066]: DEBUG nova.virt.hardware [None req-e3db032e-002b-4e25-a569-0fed0f11200b tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 839.244940] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc772488-6525-4927-877e-af9221cdde81 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.266019] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a7d0fa3-3d09-4582-b46d-6bc762209c8f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.270198] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad900a9f-6cbb-4551-8fb8-e7079ab2af53 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.276787] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75b33f23-4474-453a-8aad-bbf834c3a06c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.304860] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3a43169-fa5c-45b2-8eff-06255a5a6cde {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.322904] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-63dd5fc5-276d-49c0-99ba-71a261a8efa4 tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] The volume has not been displaced from its original location: [datastore2] volume-26ceca45-1cae-48c9-a567-e7cc4f9970c1/volume-26ceca45-1cae-48c9-a567-e7cc4f9970c1.vmdk. No consolidation needed. {{(pid=62066) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 839.328350] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-63dd5fc5-276d-49c0-99ba-71a261a8efa4 tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] Reconfiguring VM instance instance-0000003c to detach disk 2000 {{(pid=62066) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 839.328549] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-15ac48e6-1893-4eb6-8756-c936b5da83e9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.353050] env[62066]: DEBUG oslo_vmware.api [None req-28a644a7-dc09-4460-bd7c-0b89c8effe40 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1340894, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.150317} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.354474] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-28a644a7-dc09-4460-bd7c-0b89c8effe40 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 88daaaa6-f385-4161-bc74-9fffabb1145d] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 839.354772] env[62066]: DEBUG oslo_vmware.api [None req-63dd5fc5-276d-49c0-99ba-71a261a8efa4 tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] Waiting for the task: (returnval){ [ 839.354772] env[62066]: value = "task-1340896" [ 839.354772] env[62066]: _type = "Task" [ 839.354772] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.355808] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87d571e6-d241-4aa3-8942-a78281675a91 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.382990] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-28a644a7-dc09-4460-bd7c-0b89c8effe40 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 88daaaa6-f385-4161-bc74-9fffabb1145d] Reconfiguring VM instance instance-00000047 to attach disk [datastore1] 88daaaa6-f385-4161-bc74-9fffabb1145d/88daaaa6-f385-4161-bc74-9fffabb1145d.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 839.389102] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ed1279a3-ecf2-4d18-99c0-c7c0fe413310 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.403448] env[62066]: DEBUG oslo_vmware.api [None req-63dd5fc5-276d-49c0-99ba-71a261a8efa4 tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] Task: {'id': task-1340896, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.409439] env[62066]: DEBUG oslo_vmware.api [None req-28a644a7-dc09-4460-bd7c-0b89c8effe40 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Waiting for the task: (returnval){ [ 839.409439] env[62066]: value = "task-1340897" [ 839.409439] env[62066]: _type = "Task" [ 839.409439] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.415765] env[62066]: DEBUG oslo_vmware.api [None req-838d4b27-f790-4e25-a20b-eab8ecc5b3e3 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Task: {'id': task-1340892, 'name': CloneVM_Task} progress is 94%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.422232] env[62066]: DEBUG oslo_vmware.api [None req-28a644a7-dc09-4460-bd7c-0b89c8effe40 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1340897, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.506123] env[62066]: DEBUG nova.network.neutron [req-97b5a9ff-1955-48e1-b324-f1f25741440b req-8fd77951-be69-4a8c-8790-6001cd148a56 service nova] [instance: 2709d0ad-5dd8-4e3c-b1e6-3d1e2e242380] Updated VIF entry in instance network info cache for port c5cdc663-1e55-4238-bad4-7fb406b34a36. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 839.506704] env[62066]: DEBUG nova.network.neutron [req-97b5a9ff-1955-48e1-b324-f1f25741440b req-8fd77951-be69-4a8c-8790-6001cd148a56 service nova] [instance: 2709d0ad-5dd8-4e3c-b1e6-3d1e2e242380] Updating instance_info_cache with network_info: [{"id": "c5cdc663-1e55-4238-bad4-7fb406b34a36", "address": "fa:16:3e:42:1a:81", "network": {"id": "2df566a5-1f6f-4665-a06d-1441530a4f80", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-1008410317-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.148", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d1785af300d545e6913c0e91148f4563", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b4d548e7-d762-406a-bb2d-dc7168a8ca67", "external-id": "nsx-vlan-transportzone-796", "segmentation_id": 796, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc5cdc663-1e", "ovs_interfaceid": "c5cdc663-1e55-4238-bad4-7fb406b34a36", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 839.563251] env[62066]: INFO nova.compute.manager [None req-ef8a6667-cf2e-41c6-b3b6-d096c3290f66 tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] [instance: 5e96def2-0cbd-4bd9-93f4-6a365a0142b0] Rescuing [ 839.565395] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ef8a6667-cf2e-41c6-b3b6-d096c3290f66 tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Acquiring lock "refresh_cache-5e96def2-0cbd-4bd9-93f4-6a365a0142b0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 839.565395] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ef8a6667-cf2e-41c6-b3b6-d096c3290f66 tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Acquired lock "refresh_cache-5e96def2-0cbd-4bd9-93f4-6a365a0142b0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 839.565395] env[62066]: DEBUG nova.network.neutron [None req-ef8a6667-cf2e-41c6-b3b6-d096c3290f66 tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] [instance: 5e96def2-0cbd-4bd9-93f4-6a365a0142b0] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 839.575247] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e0d7e39c-c972-4829-872c-c4830b26c06b tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Acquiring lock "86a92b61-f3e9-48f9-8ee2-756669d558ef" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 839.575247] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e0d7e39c-c972-4829-872c-c4830b26c06b tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Lock "86a92b61-f3e9-48f9-8ee2-756669d558ef" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 839.575247] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e0d7e39c-c972-4829-872c-c4830b26c06b tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Acquiring lock "86a92b61-f3e9-48f9-8ee2-756669d558ef-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 839.575247] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e0d7e39c-c972-4829-872c-c4830b26c06b tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Lock "86a92b61-f3e9-48f9-8ee2-756669d558ef-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 839.575247] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e0d7e39c-c972-4829-872c-c4830b26c06b tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Lock "86a92b61-f3e9-48f9-8ee2-756669d558ef-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 839.575247] env[62066]: INFO nova.compute.manager [None req-e0d7e39c-c972-4829-872c-c4830b26c06b tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 86a92b61-f3e9-48f9-8ee2-756669d558ef] Terminating instance [ 839.580324] env[62066]: DEBUG nova.compute.manager [None req-e0d7e39c-c972-4829-872c-c4830b26c06b tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 86a92b61-f3e9-48f9-8ee2-756669d558ef] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 839.581268] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-e0d7e39c-c972-4829-872c-c4830b26c06b tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 86a92b61-f3e9-48f9-8ee2-756669d558ef] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 839.583185] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a475c8d4-5bf5-4fd3-ad15-353838ad8149 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.592266] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-e0d7e39c-c972-4829-872c-c4830b26c06b tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 86a92b61-f3e9-48f9-8ee2-756669d558ef] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 839.592794] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b447f5a2-a192-4c96-83ce-45747f24ae30 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.602994] env[62066]: DEBUG oslo_vmware.api [None req-e0d7e39c-c972-4829-872c-c4830b26c06b tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Waiting for the task: (returnval){ [ 839.602994] env[62066]: value = "task-1340898" [ 839.602994] env[62066]: _type = "Task" [ 839.602994] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.615967] env[62066]: DEBUG oslo_vmware.api [None req-e0d7e39c-c972-4829-872c-c4830b26c06b tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': task-1340898, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.630791] env[62066]: DEBUG nova.compute.manager [None req-73e4bc46-8266-447d-984c-3ca2b99ffd23 tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] [instance: 0bc31a8b-a5fc-4f6e-a4c0-024bf4260755] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 839.701599] env[62066]: DEBUG oslo_concurrency.lockutils [None req-8589d93a-2555-4139-b6bf-f9de6865cf74 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Lock "49f46244-34fa-48a1-95a2-8e95850f345d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.513s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 839.871980] env[62066]: DEBUG oslo_vmware.api [None req-63dd5fc5-276d-49c0-99ba-71a261a8efa4 tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] Task: {'id': task-1340896, 'name': ReconfigVM_Task, 'duration_secs': 0.396129} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.872721] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-63dd5fc5-276d-49c0-99ba-71a261a8efa4 tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] Reconfigured VM instance instance-0000003c to detach disk 2000 {{(pid=62066) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 839.880147] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ce3b8fec-0e93-49a2-a92f-d9d17bdd7d2d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.898050] env[62066]: DEBUG oslo_vmware.api [None req-63dd5fc5-276d-49c0-99ba-71a261a8efa4 tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] Waiting for the task: (returnval){ [ 839.898050] env[62066]: value = "task-1340899" [ 839.898050] env[62066]: _type = "Task" [ 839.898050] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.907061] env[62066]: DEBUG oslo_vmware.api [None req-63dd5fc5-276d-49c0-99ba-71a261a8efa4 tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] Task: {'id': task-1340899, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.923644] env[62066]: DEBUG oslo_vmware.api [None req-838d4b27-f790-4e25-a20b-eab8ecc5b3e3 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Task: {'id': task-1340892, 'name': CloneVM_Task} progress is 95%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.928050] env[62066]: DEBUG oslo_vmware.api [None req-28a644a7-dc09-4460-bd7c-0b89c8effe40 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1340897, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.011804] env[62066]: DEBUG oslo_concurrency.lockutils [req-97b5a9ff-1955-48e1-b324-f1f25741440b req-8fd77951-be69-4a8c-8790-6001cd148a56 service nova] Releasing lock "refresh_cache-2709d0ad-5dd8-4e3c-b1e6-3d1e2e242380" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 840.116496] env[62066]: DEBUG oslo_vmware.api [None req-e0d7e39c-c972-4829-872c-c4830b26c06b tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': task-1340898, 'name': PowerOffVM_Task, 'duration_secs': 0.293551} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.116496] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-e0d7e39c-c972-4829-872c-c4830b26c06b tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 86a92b61-f3e9-48f9-8ee2-756669d558ef] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 840.116496] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-e0d7e39c-c972-4829-872c-c4830b26c06b tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 86a92b61-f3e9-48f9-8ee2-756669d558ef] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 840.116496] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-60ea9434-f4cf-47cd-8f68-4fcf1d96e9fc {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.163255] env[62066]: DEBUG oslo_concurrency.lockutils [None req-73e4bc46-8266-447d-984c-3ca2b99ffd23 tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 840.187783] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-e0d7e39c-c972-4829-872c-c4830b26c06b tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 86a92b61-f3e9-48f9-8ee2-756669d558ef] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 840.187994] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-e0d7e39c-c972-4829-872c-c4830b26c06b tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 86a92b61-f3e9-48f9-8ee2-756669d558ef] Deleting contents of the VM from datastore datastore1 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 840.188972] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-e0d7e39c-c972-4829-872c-c4830b26c06b tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Deleting the datastore file [datastore1] 86a92b61-f3e9-48f9-8ee2-756669d558ef {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 840.188972] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-62a68431-2b22-4480-9b20-1157b3b343c2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.194486] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Instance 1a4b0637-1a56-41ef-b89b-6b56d24ed206 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 840.194621] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Instance 1e1a2ede-4ad8-4600-851e-6f2046b3f919 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 840.194738] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Instance daffaf51-4c45-44aa-8fc2-4db066a09971 actively managed on this compute host and has allocations in placement: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 840.194848] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Instance 02fb3e92-5dd6-4b1e-a6e4-d60d3fc85b07 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 840.194957] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Instance 459f432b-e817-405f-9c3e-dfa5d957788c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 840.195104] env[62066]: WARNING nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Instance 5f645eaa-6edc-4362-82ad-38c4c57b2be2 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 840.195216] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Instance 8ff6fbb9-c90f-498d-9a85-d220a8c2f794 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 840.195324] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Instance 2709d0ad-5dd8-4e3c-b1e6-3d1e2e242380 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 840.195429] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Instance 88daaaa6-f385-4161-bc74-9fffabb1145d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 840.195532] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Instance de2e2be7-efdb-45a8-842a-640ab9deb1d9 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 840.195633] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Instance 5e96def2-0cbd-4bd9-93f4-6a365a0142b0 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 840.195733] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Instance 86a92b61-f3e9-48f9-8ee2-756669d558ef actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 840.195833] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Instance b08f27b1-429e-4d05-9a95-f45fd0cb2904 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 840.198854] env[62066]: DEBUG oslo_vmware.api [None req-e0d7e39c-c972-4829-872c-c4830b26c06b tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Waiting for the task: (returnval){ [ 840.198854] env[62066]: value = "task-1340901" [ 840.198854] env[62066]: _type = "Task" [ 840.198854] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.210353] env[62066]: DEBUG oslo_vmware.api [None req-e0d7e39c-c972-4829-872c-c4830b26c06b tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': task-1340901, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.301151] env[62066]: DEBUG nova.network.neutron [None req-ef8a6667-cf2e-41c6-b3b6-d096c3290f66 tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] [instance: 5e96def2-0cbd-4bd9-93f4-6a365a0142b0] Updating instance_info_cache with network_info: [{"id": "0932c7ac-630e-4cef-8991-45b438f39483", "address": "fa:16:3e:ca:08:8d", "network": {"id": "fac1b86c-746c-4b6e-8d7f-21fe6ed404a3", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1827773279-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "59855b1cf91d43e3ba0482f414c4e546", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "88651df2-0506-4f6c-b868-dd30a81f2b1c", "external-id": "nsx-vlan-transportzone-366", "segmentation_id": 366, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0932c7ac-63", "ovs_interfaceid": "0932c7ac-630e-4cef-8991-45b438f39483", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 840.408247] env[62066]: DEBUG oslo_vmware.api [None req-63dd5fc5-276d-49c0-99ba-71a261a8efa4 tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] Task: {'id': task-1340899, 'name': ReconfigVM_Task, 'duration_secs': 0.125632} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.411535] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-63dd5fc5-276d-49c0-99ba-71a261a8efa4 tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-285999', 'volume_id': '26ceca45-1cae-48c9-a567-e7cc4f9970c1', 'name': 'volume-26ceca45-1cae-48c9-a567-e7cc4f9970c1', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'daffaf51-4c45-44aa-8fc2-4db066a09971', 'attached_at': '', 'detached_at': '', 'volume_id': '26ceca45-1cae-48c9-a567-e7cc4f9970c1', 'serial': '26ceca45-1cae-48c9-a567-e7cc4f9970c1'} {{(pid=62066) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 840.411731] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-63dd5fc5-276d-49c0-99ba-71a261a8efa4 tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 840.412577] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d8b95f7-d1bd-484a-abc0-dc446a914051 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.423335] env[62066]: DEBUG oslo_vmware.api [None req-838d4b27-f790-4e25-a20b-eab8ecc5b3e3 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Task: {'id': task-1340892, 'name': CloneVM_Task, 'duration_secs': 1.650516} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.427967] env[62066]: INFO nova.virt.vmwareapi.vmops [None req-838d4b27-f790-4e25-a20b-eab8ecc5b3e3 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] [instance: 1a4b0637-1a56-41ef-b89b-6b56d24ed206] Created linked-clone VM from snapshot [ 840.428283] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-63dd5fc5-276d-49c0-99ba-71a261a8efa4 tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 840.428525] env[62066]: DEBUG oslo_vmware.api [None req-28a644a7-dc09-4460-bd7c-0b89c8effe40 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1340897, 'name': ReconfigVM_Task, 'duration_secs': 0.565807} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.429223] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da4cea7e-0106-4821-8b3d-8603ec74f853 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.431646] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8d4858eb-37bd-4c04-b16f-7702ee42a8b2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.433030] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-28a644a7-dc09-4460-bd7c-0b89c8effe40 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 88daaaa6-f385-4161-bc74-9fffabb1145d] Reconfigured VM instance instance-00000047 to attach disk [datastore1] 88daaaa6-f385-4161-bc74-9fffabb1145d/88daaaa6-f385-4161-bc74-9fffabb1145d.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 840.433618] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-740fb844-446d-4a34-aeca-a5fbaebd2d6e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.440651] env[62066]: DEBUG nova.virt.vmwareapi.images [None req-838d4b27-f790-4e25-a20b-eab8ecc5b3e3 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] [instance: 1a4b0637-1a56-41ef-b89b-6b56d24ed206] Uploading image 70769f31-9c4a-4989-973b-7d69d5ea0e47 {{(pid=62066) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 840.445891] env[62066]: DEBUG oslo_vmware.api [None req-28a644a7-dc09-4460-bd7c-0b89c8effe40 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Waiting for the task: (returnval){ [ 840.445891] env[62066]: value = "task-1340903" [ 840.445891] env[62066]: _type = "Task" [ 840.445891] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.460245] env[62066]: DEBUG oslo_vmware.api [None req-28a644a7-dc09-4460-bd7c-0b89c8effe40 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1340903, 'name': Rename_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.480593] env[62066]: DEBUG oslo_vmware.rw_handles [None req-838d4b27-f790-4e25-a20b-eab8ecc5b3e3 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 840.480593] env[62066]: value = "vm-286062" [ 840.480593] env[62066]: _type = "VirtualMachine" [ 840.480593] env[62066]: }. {{(pid=62066) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 840.480593] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-5f19ffd7-e66c-4c9f-b9e2-675d23f62e88 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.490347] env[62066]: DEBUG oslo_vmware.rw_handles [None req-838d4b27-f790-4e25-a20b-eab8ecc5b3e3 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Lease: (returnval){ [ 840.490347] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]520a9402-f3e7-406d-79f1-f6377e3f35df" [ 840.490347] env[62066]: _type = "HttpNfcLease" [ 840.490347] env[62066]: } obtained for exporting VM: (result){ [ 840.490347] env[62066]: value = "vm-286062" [ 840.490347] env[62066]: _type = "VirtualMachine" [ 840.490347] env[62066]: }. {{(pid=62066) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 840.490347] env[62066]: DEBUG oslo_vmware.api [None req-838d4b27-f790-4e25-a20b-eab8ecc5b3e3 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Waiting for the lease: (returnval){ [ 840.490347] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]520a9402-f3e7-406d-79f1-f6377e3f35df" [ 840.490347] env[62066]: _type = "HttpNfcLease" [ 840.490347] env[62066]: } to be ready. {{(pid=62066) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 840.498495] env[62066]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 840.498495] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]520a9402-f3e7-406d-79f1-f6377e3f35df" [ 840.498495] env[62066]: _type = "HttpNfcLease" [ 840.498495] env[62066]: } is initializing. {{(pid=62066) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 840.563963] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-63dd5fc5-276d-49c0-99ba-71a261a8efa4 tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 840.564073] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-63dd5fc5-276d-49c0-99ba-71a261a8efa4 tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] Deleting contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 840.564190] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-63dd5fc5-276d-49c0-99ba-71a261a8efa4 tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] Deleting the datastore file [datastore2] daffaf51-4c45-44aa-8fc2-4db066a09971 {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 840.564475] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-dfeaf249-4dd2-417f-8a30-3c08c168c8d6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.572157] env[62066]: DEBUG oslo_vmware.api [None req-63dd5fc5-276d-49c0-99ba-71a261a8efa4 tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] Waiting for the task: (returnval){ [ 840.572157] env[62066]: value = "task-1340905" [ 840.572157] env[62066]: _type = "Task" [ 840.572157] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.581688] env[62066]: DEBUG oslo_vmware.api [None req-63dd5fc5-276d-49c0-99ba-71a261a8efa4 tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] Task: {'id': task-1340905, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.598193] env[62066]: DEBUG nova.network.neutron [None req-e3db032e-002b-4e25-a569-0fed0f11200b tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: b08f27b1-429e-4d05-9a95-f45fd0cb2904] Successfully updated port: 62d600d9-8ad3-43f8-a2b0-050747916fdb {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 840.617982] env[62066]: DEBUG nova.compute.manager [req-e443a968-6e93-4d02-a89c-342cf33d9124 req-fa37a904-cfa4-4c00-a503-40847ce82ff5 service nova] [instance: b08f27b1-429e-4d05-9a95-f45fd0cb2904] Received event network-vif-plugged-62d600d9-8ad3-43f8-a2b0-050747916fdb {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 840.618216] env[62066]: DEBUG oslo_concurrency.lockutils [req-e443a968-6e93-4d02-a89c-342cf33d9124 req-fa37a904-cfa4-4c00-a503-40847ce82ff5 service nova] Acquiring lock "b08f27b1-429e-4d05-9a95-f45fd0cb2904-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 840.618463] env[62066]: DEBUG oslo_concurrency.lockutils [req-e443a968-6e93-4d02-a89c-342cf33d9124 req-fa37a904-cfa4-4c00-a503-40847ce82ff5 service nova] Lock "b08f27b1-429e-4d05-9a95-f45fd0cb2904-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 840.618735] env[62066]: DEBUG oslo_concurrency.lockutils [req-e443a968-6e93-4d02-a89c-342cf33d9124 req-fa37a904-cfa4-4c00-a503-40847ce82ff5 service nova] Lock "b08f27b1-429e-4d05-9a95-f45fd0cb2904-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 840.618844] env[62066]: DEBUG nova.compute.manager [req-e443a968-6e93-4d02-a89c-342cf33d9124 req-fa37a904-cfa4-4c00-a503-40847ce82ff5 service nova] [instance: b08f27b1-429e-4d05-9a95-f45fd0cb2904] No waiting events found dispatching network-vif-plugged-62d600d9-8ad3-43f8-a2b0-050747916fdb {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 840.619012] env[62066]: WARNING nova.compute.manager [req-e443a968-6e93-4d02-a89c-342cf33d9124 req-fa37a904-cfa4-4c00-a503-40847ce82ff5 service nova] [instance: b08f27b1-429e-4d05-9a95-f45fd0cb2904] Received unexpected event network-vif-plugged-62d600d9-8ad3-43f8-a2b0-050747916fdb for instance with vm_state building and task_state spawning. [ 840.703840] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Instance d77e3405-d158-4a8c-9f54-d5fbd26fed48 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 840.715775] env[62066]: DEBUG oslo_vmware.api [None req-e0d7e39c-c972-4829-872c-c4830b26c06b tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': task-1340901, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.492047} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.716936] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-e0d7e39c-c972-4829-872c-c4830b26c06b tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 840.716936] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-e0d7e39c-c972-4829-872c-c4830b26c06b tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 86a92b61-f3e9-48f9-8ee2-756669d558ef] Deleted contents of the VM from datastore datastore1 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 840.717133] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-e0d7e39c-c972-4829-872c-c4830b26c06b tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 86a92b61-f3e9-48f9-8ee2-756669d558ef] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 840.717361] env[62066]: INFO nova.compute.manager [None req-e0d7e39c-c972-4829-872c-c4830b26c06b tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 86a92b61-f3e9-48f9-8ee2-756669d558ef] Took 1.14 seconds to destroy the instance on the hypervisor. [ 840.717564] env[62066]: DEBUG oslo.service.loopingcall [None req-e0d7e39c-c972-4829-872c-c4830b26c06b tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 840.717760] env[62066]: DEBUG nova.compute.manager [-] [instance: 86a92b61-f3e9-48f9-8ee2-756669d558ef] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 840.717852] env[62066]: DEBUG nova.network.neutron [-] [instance: 86a92b61-f3e9-48f9-8ee2-756669d558ef] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 840.807905] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ef8a6667-cf2e-41c6-b3b6-d096c3290f66 tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Releasing lock "refresh_cache-5e96def2-0cbd-4bd9-93f4-6a365a0142b0" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 840.962498] env[62066]: DEBUG oslo_vmware.api [None req-28a644a7-dc09-4460-bd7c-0b89c8effe40 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1340903, 'name': Rename_Task, 'duration_secs': 0.243872} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.963142] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-28a644a7-dc09-4460-bd7c-0b89c8effe40 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 88daaaa6-f385-4161-bc74-9fffabb1145d] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 840.963543] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8a066a04-cfc2-42a2-aa46-84afce8caa47 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.970737] env[62066]: DEBUG oslo_vmware.api [None req-28a644a7-dc09-4460-bd7c-0b89c8effe40 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Waiting for the task: (returnval){ [ 840.970737] env[62066]: value = "task-1340906" [ 840.970737] env[62066]: _type = "Task" [ 840.970737] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.984692] env[62066]: DEBUG oslo_vmware.api [None req-28a644a7-dc09-4460-bd7c-0b89c8effe40 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1340906, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.996282] env[62066]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 840.996282] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]520a9402-f3e7-406d-79f1-f6377e3f35df" [ 840.996282] env[62066]: _type = "HttpNfcLease" [ 840.996282] env[62066]: } is ready. {{(pid=62066) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 840.996590] env[62066]: DEBUG oslo_vmware.rw_handles [None req-838d4b27-f790-4e25-a20b-eab8ecc5b3e3 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 840.996590] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]520a9402-f3e7-406d-79f1-f6377e3f35df" [ 840.996590] env[62066]: _type = "HttpNfcLease" [ 840.996590] env[62066]: }. {{(pid=62066) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 840.997347] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1d71589-4284-48f6-8c8d-25c70ea83a8d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.008372] env[62066]: DEBUG oslo_vmware.rw_handles [None req-838d4b27-f790-4e25-a20b-eab8ecc5b3e3 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52e7a16c-6dba-69f6-aeb6-0f9e2db1f0fa/disk-0.vmdk from lease info. {{(pid=62066) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 841.008583] env[62066]: DEBUG oslo_vmware.rw_handles [None req-838d4b27-f790-4e25-a20b-eab8ecc5b3e3 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52e7a16c-6dba-69f6-aeb6-0f9e2db1f0fa/disk-0.vmdk for reading. {{(pid=62066) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 841.083321] env[62066]: DEBUG oslo_vmware.api [None req-63dd5fc5-276d-49c0-99ba-71a261a8efa4 tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] Task: {'id': task-1340905, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.242433} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.083321] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-63dd5fc5-276d-49c0-99ba-71a261a8efa4 tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 841.083636] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-63dd5fc5-276d-49c0-99ba-71a261a8efa4 tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] Deleted contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 841.083858] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-63dd5fc5-276d-49c0-99ba-71a261a8efa4 tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 841.101160] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e3db032e-002b-4e25-a569-0fed0f11200b tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Acquiring lock "refresh_cache-b08f27b1-429e-4d05-9a95-f45fd0cb2904" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 841.101310] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e3db032e-002b-4e25-a569-0fed0f11200b tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Acquired lock "refresh_cache-b08f27b1-429e-4d05-9a95-f45fd0cb2904" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 841.101446] env[62066]: DEBUG nova.network.neutron [None req-e3db032e-002b-4e25-a569-0fed0f11200b tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: b08f27b1-429e-4d05-9a95-f45fd0cb2904] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 841.131097] env[62066]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-b2854f96-ba1f-4444-891f-213e49626b9b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.145753] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-63dd5fc5-276d-49c0-99ba-71a261a8efa4 tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] Volume detach. Driver type: vmdk {{(pid=62066) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 841.146056] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-278fdd7f-f85e-4b45-9567-c3e9cf6deec1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.159392] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90cfc2f6-c0ed-4326-9e6d-bdd5210ee90f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.190593] env[62066]: ERROR nova.compute.manager [None req-63dd5fc5-276d-49c0-99ba-71a261a8efa4 tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] Failed to detach volume 26ceca45-1cae-48c9-a567-e7cc4f9970c1 from /dev/sda: nova.exception.InstanceNotFound: Instance daffaf51-4c45-44aa-8fc2-4db066a09971 could not be found. [ 841.190593] env[62066]: ERROR nova.compute.manager [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] Traceback (most recent call last): [ 841.190593] env[62066]: ERROR nova.compute.manager [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] File "/opt/stack/nova/nova/compute/manager.py", line 4142, in _do_rebuild_instance [ 841.190593] env[62066]: ERROR nova.compute.manager [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] self.driver.rebuild(**kwargs) [ 841.190593] env[62066]: ERROR nova.compute.manager [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] File "/opt/stack/nova/nova/virt/driver.py", line 493, in rebuild [ 841.190593] env[62066]: ERROR nova.compute.manager [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] raise NotImplementedError() [ 841.190593] env[62066]: ERROR nova.compute.manager [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] NotImplementedError [ 841.190593] env[62066]: ERROR nova.compute.manager [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] [ 841.190593] env[62066]: ERROR nova.compute.manager [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] During handling of the above exception, another exception occurred: [ 841.190593] env[62066]: ERROR nova.compute.manager [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] [ 841.190593] env[62066]: ERROR nova.compute.manager [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] Traceback (most recent call last): [ 841.190593] env[62066]: ERROR nova.compute.manager [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] File "/opt/stack/nova/nova/compute/manager.py", line 3565, in _detach_root_volume [ 841.190593] env[62066]: ERROR nova.compute.manager [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] self.driver.detach_volume(context, old_connection_info, [ 841.190593] env[62066]: ERROR nova.compute.manager [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 559, in detach_volume [ 841.190593] env[62066]: ERROR nova.compute.manager [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] return self._volumeops.detach_volume(connection_info, instance) [ 841.190593] env[62066]: ERROR nova.compute.manager [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 649, in detach_volume [ 841.190593] env[62066]: ERROR nova.compute.manager [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] self._detach_volume_vmdk(connection_info, instance) [ 841.190593] env[62066]: ERROR nova.compute.manager [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 569, in _detach_volume_vmdk [ 841.190593] env[62066]: ERROR nova.compute.manager [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] vm_ref = vm_util.get_vm_ref(self._session, instance) [ 841.190593] env[62066]: ERROR nova.compute.manager [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1145, in get_vm_ref [ 841.190593] env[62066]: ERROR nova.compute.manager [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] stable_ref.fetch_moref(session) [ 841.190593] env[62066]: ERROR nova.compute.manager [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1136, in fetch_moref [ 841.190593] env[62066]: ERROR nova.compute.manager [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] raise exception.InstanceNotFound(instance_id=self._uuid) [ 841.190593] env[62066]: ERROR nova.compute.manager [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] nova.exception.InstanceNotFound: Instance daffaf51-4c45-44aa-8fc2-4db066a09971 could not be found. [ 841.190593] env[62066]: ERROR nova.compute.manager [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] [ 841.212028] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Instance 3120c9a5-0558-4e83-b0fb-994a989398fb has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 841.353766] env[62066]: DEBUG nova.compute.utils [None req-63dd5fc5-276d-49c0-99ba-71a261a8efa4 tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] Build of instance daffaf51-4c45-44aa-8fc2-4db066a09971 aborted: Failed to rebuild volume backed instance. {{(pid=62066) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 841.355434] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-ef8a6667-cf2e-41c6-b3b6-d096c3290f66 tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] [instance: 5e96def2-0cbd-4bd9-93f4-6a365a0142b0] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 841.355434] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ca95e5f8-4b7a-440f-9ef0-861035eda7e2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.357788] env[62066]: ERROR nova.compute.manager [None req-63dd5fc5-276d-49c0-99ba-71a261a8efa4 tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] Setting instance vm_state to ERROR: nova.exception.BuildAbortException: Build of instance daffaf51-4c45-44aa-8fc2-4db066a09971 aborted: Failed to rebuild volume backed instance. [ 841.357788] env[62066]: ERROR nova.compute.manager [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] Traceback (most recent call last): [ 841.357788] env[62066]: ERROR nova.compute.manager [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] File "/opt/stack/nova/nova/compute/manager.py", line 4142, in _do_rebuild_instance [ 841.357788] env[62066]: ERROR nova.compute.manager [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] self.driver.rebuild(**kwargs) [ 841.357788] env[62066]: ERROR nova.compute.manager [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] File "/opt/stack/nova/nova/virt/driver.py", line 493, in rebuild [ 841.357788] env[62066]: ERROR nova.compute.manager [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] raise NotImplementedError() [ 841.357788] env[62066]: ERROR nova.compute.manager [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] NotImplementedError [ 841.357788] env[62066]: ERROR nova.compute.manager [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] [ 841.357788] env[62066]: ERROR nova.compute.manager [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] During handling of the above exception, another exception occurred: [ 841.357788] env[62066]: ERROR nova.compute.manager [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] [ 841.357788] env[62066]: ERROR nova.compute.manager [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] Traceback (most recent call last): [ 841.357788] env[62066]: ERROR nova.compute.manager [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] File "/opt/stack/nova/nova/compute/manager.py", line 3600, in _rebuild_volume_backed_instance [ 841.357788] env[62066]: ERROR nova.compute.manager [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] self._detach_root_volume(context, instance, root_bdm) [ 841.357788] env[62066]: ERROR nova.compute.manager [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] File "/opt/stack/nova/nova/compute/manager.py", line 3579, in _detach_root_volume [ 841.357788] env[62066]: ERROR nova.compute.manager [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] with excutils.save_and_reraise_exception(): [ 841.357788] env[62066]: ERROR nova.compute.manager [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 841.357788] env[62066]: ERROR nova.compute.manager [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] self.force_reraise() [ 841.357788] env[62066]: ERROR nova.compute.manager [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 841.357788] env[62066]: ERROR nova.compute.manager [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] raise self.value [ 841.357788] env[62066]: ERROR nova.compute.manager [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] File "/opt/stack/nova/nova/compute/manager.py", line 3565, in _detach_root_volume [ 841.357788] env[62066]: ERROR nova.compute.manager [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] self.driver.detach_volume(context, old_connection_info, [ 841.357788] env[62066]: ERROR nova.compute.manager [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 559, in detach_volume [ 841.357788] env[62066]: ERROR nova.compute.manager [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] return self._volumeops.detach_volume(connection_info, instance) [ 841.357788] env[62066]: ERROR nova.compute.manager [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 649, in detach_volume [ 841.357788] env[62066]: ERROR nova.compute.manager [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] self._detach_volume_vmdk(connection_info, instance) [ 841.357788] env[62066]: ERROR nova.compute.manager [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 569, in _detach_volume_vmdk [ 841.357788] env[62066]: ERROR nova.compute.manager [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] vm_ref = vm_util.get_vm_ref(self._session, instance) [ 841.357788] env[62066]: ERROR nova.compute.manager [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1145, in get_vm_ref [ 841.357788] env[62066]: ERROR nova.compute.manager [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] stable_ref.fetch_moref(session) [ 841.357788] env[62066]: ERROR nova.compute.manager [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1136, in fetch_moref [ 841.357788] env[62066]: ERROR nova.compute.manager [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] raise exception.InstanceNotFound(instance_id=self._uuid) [ 841.357788] env[62066]: ERROR nova.compute.manager [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] nova.exception.InstanceNotFound: Instance daffaf51-4c45-44aa-8fc2-4db066a09971 could not be found. [ 841.357788] env[62066]: ERROR nova.compute.manager [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] [ 841.357788] env[62066]: ERROR nova.compute.manager [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] During handling of the above exception, another exception occurred: [ 841.357788] env[62066]: ERROR nova.compute.manager [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] [ 841.357788] env[62066]: ERROR nova.compute.manager [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] Traceback (most recent call last): [ 841.357788] env[62066]: ERROR nova.compute.manager [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] File "/opt/stack/nova/nova/compute/manager.py", line 10865, in _error_out_instance_on_exception [ 841.357788] env[62066]: ERROR nova.compute.manager [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] yield [ 841.357788] env[62066]: ERROR nova.compute.manager [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] File "/opt/stack/nova/nova/compute/manager.py", line 3868, in rebuild_instance [ 841.357788] env[62066]: ERROR nova.compute.manager [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] self._do_rebuild_instance_with_claim( [ 841.359090] env[62066]: ERROR nova.compute.manager [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] File "/opt/stack/nova/nova/compute/manager.py", line 3954, in _do_rebuild_instance_with_claim [ 841.359090] env[62066]: ERROR nova.compute.manager [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] self._do_rebuild_instance( [ 841.359090] env[62066]: ERROR nova.compute.manager [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] File "/opt/stack/nova/nova/compute/manager.py", line 4146, in _do_rebuild_instance [ 841.359090] env[62066]: ERROR nova.compute.manager [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] self._rebuild_default_impl(**kwargs) [ 841.359090] env[62066]: ERROR nova.compute.manager [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] File "/opt/stack/nova/nova/compute/manager.py", line 3723, in _rebuild_default_impl [ 841.359090] env[62066]: ERROR nova.compute.manager [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] self._rebuild_volume_backed_instance( [ 841.359090] env[62066]: ERROR nova.compute.manager [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] File "/opt/stack/nova/nova/compute/manager.py", line 3615, in _rebuild_volume_backed_instance [ 841.359090] env[62066]: ERROR nova.compute.manager [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] raise exception.BuildAbortException( [ 841.359090] env[62066]: ERROR nova.compute.manager [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] nova.exception.BuildAbortException: Build of instance daffaf51-4c45-44aa-8fc2-4db066a09971 aborted: Failed to rebuild volume backed instance. [ 841.359090] env[62066]: ERROR nova.compute.manager [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] [ 841.367196] env[62066]: DEBUG oslo_vmware.api [None req-ef8a6667-cf2e-41c6-b3b6-d096c3290f66 tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Waiting for the task: (returnval){ [ 841.367196] env[62066]: value = "task-1340907" [ 841.367196] env[62066]: _type = "Task" [ 841.367196] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.381738] env[62066]: DEBUG oslo_vmware.api [None req-ef8a6667-cf2e-41c6-b3b6-d096c3290f66 tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Task: {'id': task-1340907, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.481822] env[62066]: DEBUG oslo_vmware.api [None req-28a644a7-dc09-4460-bd7c-0b89c8effe40 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1340906, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.572043] env[62066]: DEBUG nova.network.neutron [-] [instance: 86a92b61-f3e9-48f9-8ee2-756669d558ef] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 841.638733] env[62066]: DEBUG nova.network.neutron [None req-e3db032e-002b-4e25-a569-0fed0f11200b tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: b08f27b1-429e-4d05-9a95-f45fd0cb2904] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 841.714011] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Instance 0bc31a8b-a5fc-4f6e-a4c0-024bf4260755 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 841.714596] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Total usable vcpus: 48, total allocated vcpus: 12 {{(pid=62066) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 841.715694] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2816MB phys_disk=200GB used_disk=11GB total_vcpus=48 used_vcpus=12 pci_stats=[] {{(pid=62066) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 841.823486] env[62066]: DEBUG nova.network.neutron [None req-e3db032e-002b-4e25-a569-0fed0f11200b tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: b08f27b1-429e-4d05-9a95-f45fd0cb2904] Updating instance_info_cache with network_info: [{"id": "62d600d9-8ad3-43f8-a2b0-050747916fdb", "address": "fa:16:3e:51:fd:46", "network": {"id": "a3852490-7a1f-499c-813a-46cbf52318da", "bridge": "br-int", "label": "tempest-ImagesTestJSON-735840736-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "75f0352c852947369474c9e05766c584", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "50171613-b419-45e3-9ada-fcb6cd921428", "external-id": "nsx-vlan-transportzone-914", "segmentation_id": 914, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap62d600d9-8a", "ovs_interfaceid": "62d600d9-8ad3-43f8-a2b0-050747916fdb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 841.884176] env[62066]: DEBUG oslo_vmware.api [None req-ef8a6667-cf2e-41c6-b3b6-d096c3290f66 tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Task: {'id': task-1340907, 'name': PowerOffVM_Task, 'duration_secs': 0.213657} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.885662] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-ef8a6667-cf2e-41c6-b3b6-d096c3290f66 tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] [instance: 5e96def2-0cbd-4bd9-93f4-6a365a0142b0] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 841.886247] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed9361e3-6e0f-414e-8471-12aed32ec2df {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.912522] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94e46828-9efa-4091-a883-3652250a33fb {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.954676] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-ef8a6667-cf2e-41c6-b3b6-d096c3290f66 tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] [instance: 5e96def2-0cbd-4bd9-93f4-6a365a0142b0] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 841.955242] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-48fc6b44-c631-49b4-b97c-b94b577b4c4b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.964172] env[62066]: DEBUG oslo_vmware.api [None req-ef8a6667-cf2e-41c6-b3b6-d096c3290f66 tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Waiting for the task: (returnval){ [ 841.964172] env[62066]: value = "task-1340908" [ 841.964172] env[62066]: _type = "Task" [ 841.964172] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.980829] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-ef8a6667-cf2e-41c6-b3b6-d096c3290f66 tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] [instance: 5e96def2-0cbd-4bd9-93f4-6a365a0142b0] VM already powered off {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 841.981337] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-ef8a6667-cf2e-41c6-b3b6-d096c3290f66 tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] [instance: 5e96def2-0cbd-4bd9-93f4-6a365a0142b0] Processing image 50ff584c-3b50-4395-af07-3e66769bc9f7 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 841.982655] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ef8a6667-cf2e-41c6-b3b6-d096c3290f66 tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 841.982655] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ef8a6667-cf2e-41c6-b3b6-d096c3290f66 tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Acquired lock "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 841.982655] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-ef8a6667-cf2e-41c6-b3b6-d096c3290f66 tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 841.985214] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a3a18a19-0531-4902-ba8d-15cc8c66826d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.994571] env[62066]: DEBUG oslo_vmware.api [None req-28a644a7-dc09-4460-bd7c-0b89c8effe40 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1340906, 'name': PowerOnVM_Task, 'duration_secs': 0.655268} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.996377] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-28a644a7-dc09-4460-bd7c-0b89c8effe40 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 88daaaa6-f385-4161-bc74-9fffabb1145d] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 841.996846] env[62066]: DEBUG nova.compute.manager [None req-28a644a7-dc09-4460-bd7c-0b89c8effe40 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 88daaaa6-f385-4161-bc74-9fffabb1145d] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 841.997316] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-ef8a6667-cf2e-41c6-b3b6-d096c3290f66 tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 841.997561] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-ef8a6667-cf2e-41c6-b3b6-d096c3290f66 tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 841.999102] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf9642f8-0443-4091-9534-97f4415c6769 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.004067] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-eea2b441-ba5f-4dd3-8698-7abe9ecfbab2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.007831] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c06f8f81-2eaa-4e58-b7f1-398b4b32d921 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.014328] env[62066]: DEBUG oslo_vmware.api [None req-ef8a6667-cf2e-41c6-b3b6-d096c3290f66 tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Waiting for the task: (returnval){ [ 842.014328] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]52580b5e-b6a5-07db-d336-312f3a13a7ce" [ 842.014328] env[62066]: _type = "Task" [ 842.014328] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.024679] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f73c85c9-01f2-4cd7-aaf7-a58eabf77f16 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.034137] env[62066]: DEBUG oslo_vmware.api [None req-ef8a6667-cf2e-41c6-b3b6-d096c3290f66 tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52580b5e-b6a5-07db-d336-312f3a13a7ce, 'name': SearchDatastore_Task, 'duration_secs': 0.008775} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.063391] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8d09e169-e1dc-486c-ac84-80daa1b36ed6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.066621] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-756c1146-36e9-4a10-bb3c-b027e374c7d3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.073912] env[62066]: INFO nova.compute.manager [-] [instance: 86a92b61-f3e9-48f9-8ee2-756669d558ef] Took 1.36 seconds to deallocate network for instance. [ 842.074333] env[62066]: DEBUG oslo_vmware.api [None req-ef8a6667-cf2e-41c6-b3b6-d096c3290f66 tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Waiting for the task: (returnval){ [ 842.074333] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]52af630e-b69c-5ab9-ca16-17a6fb8c1275" [ 842.074333] env[62066]: _type = "Task" [ 842.074333] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.086718] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-445cff43-a62f-4a70-9b3f-ce2ce110f284 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.097988] env[62066]: DEBUG oslo_vmware.api [None req-ef8a6667-cf2e-41c6-b3b6-d096c3290f66 tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52af630e-b69c-5ab9-ca16-17a6fb8c1275, 'name': SearchDatastore_Task, 'duration_secs': 0.018303} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.108027] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ef8a6667-cf2e-41c6-b3b6-d096c3290f66 tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Releasing lock "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 842.108109] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-ef8a6667-cf2e-41c6-b3b6-d096c3290f66 tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Copying virtual disk from [datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk to [datastore1] 5e96def2-0cbd-4bd9-93f4-6a365a0142b0/50ff584c-3b50-4395-af07-3e66769bc9f7-rescue.vmdk. {{(pid=62066) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 842.108644] env[62066]: DEBUG nova.compute.provider_tree [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 842.110542] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-21b87080-d49f-4eb5-aa5c-97990b52d5e1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.118450] env[62066]: DEBUG oslo_vmware.api [None req-ef8a6667-cf2e-41c6-b3b6-d096c3290f66 tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Waiting for the task: (returnval){ [ 842.118450] env[62066]: value = "task-1340909" [ 842.118450] env[62066]: _type = "Task" [ 842.118450] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.128611] env[62066]: DEBUG oslo_vmware.api [None req-ef8a6667-cf2e-41c6-b3b6-d096c3290f66 tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Task: {'id': task-1340909, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.329444] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e3db032e-002b-4e25-a569-0fed0f11200b tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Releasing lock "refresh_cache-b08f27b1-429e-4d05-9a95-f45fd0cb2904" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 842.330147] env[62066]: DEBUG nova.compute.manager [None req-e3db032e-002b-4e25-a569-0fed0f11200b tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: b08f27b1-429e-4d05-9a95-f45fd0cb2904] Instance network_info: |[{"id": "62d600d9-8ad3-43f8-a2b0-050747916fdb", "address": "fa:16:3e:51:fd:46", "network": {"id": "a3852490-7a1f-499c-813a-46cbf52318da", "bridge": "br-int", "label": "tempest-ImagesTestJSON-735840736-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "75f0352c852947369474c9e05766c584", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "50171613-b419-45e3-9ada-fcb6cd921428", "external-id": "nsx-vlan-transportzone-914", "segmentation_id": 914, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap62d600d9-8a", "ovs_interfaceid": "62d600d9-8ad3-43f8-a2b0-050747916fdb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 842.331019] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-e3db032e-002b-4e25-a569-0fed0f11200b tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: b08f27b1-429e-4d05-9a95-f45fd0cb2904] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:51:fd:46', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '50171613-b419-45e3-9ada-fcb6cd921428', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '62d600d9-8ad3-43f8-a2b0-050747916fdb', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 842.341593] env[62066]: DEBUG oslo.service.loopingcall [None req-e3db032e-002b-4e25-a569-0fed0f11200b tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 842.342051] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b08f27b1-429e-4d05-9a95-f45fd0cb2904] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 842.342599] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a44b8e57-402e-4c4d-89e0-a1843d38ab3c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.377421] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 842.377421] env[62066]: value = "task-1340910" [ 842.377421] env[62066]: _type = "Task" [ 842.377421] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.389102] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1340910, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.538375] env[62066]: DEBUG oslo_concurrency.lockutils [None req-28a644a7-dc09-4460-bd7c-0b89c8effe40 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 842.595022] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e0d7e39c-c972-4829-872c-c4830b26c06b tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 842.614170] env[62066]: DEBUG nova.scheduler.client.report [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 842.634853] env[62066]: DEBUG oslo_vmware.api [None req-ef8a6667-cf2e-41c6-b3b6-d096c3290f66 tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Task: {'id': task-1340909, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.658599] env[62066]: DEBUG nova.compute.manager [req-22c69ada-eeaa-44ad-9830-b8e171c9f91e req-595929fe-d3e7-40f1-8d16-e9e986d2911f service nova] [instance: b08f27b1-429e-4d05-9a95-f45fd0cb2904] Received event network-changed-62d600d9-8ad3-43f8-a2b0-050747916fdb {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 842.660829] env[62066]: DEBUG nova.compute.manager [req-22c69ada-eeaa-44ad-9830-b8e171c9f91e req-595929fe-d3e7-40f1-8d16-e9e986d2911f service nova] [instance: b08f27b1-429e-4d05-9a95-f45fd0cb2904] Refreshing instance network info cache due to event network-changed-62d600d9-8ad3-43f8-a2b0-050747916fdb. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 842.660985] env[62066]: DEBUG oslo_concurrency.lockutils [req-22c69ada-eeaa-44ad-9830-b8e171c9f91e req-595929fe-d3e7-40f1-8d16-e9e986d2911f service nova] Acquiring lock "refresh_cache-b08f27b1-429e-4d05-9a95-f45fd0cb2904" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 842.661159] env[62066]: DEBUG oslo_concurrency.lockutils [req-22c69ada-eeaa-44ad-9830-b8e171c9f91e req-595929fe-d3e7-40f1-8d16-e9e986d2911f service nova] Acquired lock "refresh_cache-b08f27b1-429e-4d05-9a95-f45fd0cb2904" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 842.661329] env[62066]: DEBUG nova.network.neutron [req-22c69ada-eeaa-44ad-9830-b8e171c9f91e req-595929fe-d3e7-40f1-8d16-e9e986d2911f service nova] [instance: b08f27b1-429e-4d05-9a95-f45fd0cb2904] Refreshing network info cache for port 62d600d9-8ad3-43f8-a2b0-050747916fdb {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 842.890786] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1340910, 'name': CreateVM_Task, 'duration_secs': 0.472087} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.891124] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b08f27b1-429e-4d05-9a95-f45fd0cb2904] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 842.892422] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e3db032e-002b-4e25-a569-0fed0f11200b tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 842.892720] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e3db032e-002b-4e25-a569-0fed0f11200b tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Acquired lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 842.893288] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e3db032e-002b-4e25-a569-0fed0f11200b tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 842.893562] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-803700e5-58ab-4033-93ed-3f45d3affd58 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.899081] env[62066]: DEBUG oslo_vmware.api [None req-e3db032e-002b-4e25-a569-0fed0f11200b tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Waiting for the task: (returnval){ [ 842.899081] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]521b466c-9983-ea52-4823-2bedef8f7ecf" [ 842.899081] env[62066]: _type = "Task" [ 842.899081] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.908535] env[62066]: DEBUG oslo_vmware.api [None req-e3db032e-002b-4e25-a569-0fed0f11200b tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]521b466c-9983-ea52-4823-2bedef8f7ecf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.125643] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62066) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 843.126042] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.972s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 843.129949] env[62066]: DEBUG oslo_concurrency.lockutils [None req-6075adee-b71b-4aa1-9781-77a864c59e59 tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.103s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 843.130192] env[62066]: DEBUG oslo_concurrency.lockutils [None req-6075adee-b71b-4aa1-9781-77a864c59e59 tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 843.132404] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f63339cf-eb8a-4d19-b0c5-1ac7744e7482 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.627s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 843.134114] env[62066]: INFO nova.compute.claims [None req-f63339cf-eb8a-4d19-b0c5-1ac7744e7482 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] [instance: d77e3405-d158-4a8c-9f54-d5fbd26fed48] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 843.137075] env[62066]: DEBUG oslo_vmware.api [None req-ef8a6667-cf2e-41c6-b3b6-d096c3290f66 tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Task: {'id': task-1340909, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.56083} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.137778] env[62066]: INFO nova.virt.vmwareapi.ds_util [None req-ef8a6667-cf2e-41c6-b3b6-d096c3290f66 tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Copied virtual disk from [datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk to [datastore1] 5e96def2-0cbd-4bd9-93f4-6a365a0142b0/50ff584c-3b50-4395-af07-3e66769bc9f7-rescue.vmdk. [ 843.138975] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d9f9d1e-0db1-4cdb-b0d7-1b8e5749691d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.167361] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-ef8a6667-cf2e-41c6-b3b6-d096c3290f66 tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] [instance: 5e96def2-0cbd-4bd9-93f4-6a365a0142b0] Reconfiguring VM instance instance-00000049 to attach disk [datastore1] 5e96def2-0cbd-4bd9-93f4-6a365a0142b0/50ff584c-3b50-4395-af07-3e66769bc9f7-rescue.vmdk or device None with type thin {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 843.168725] env[62066]: INFO nova.scheduler.client.report [None req-6075adee-b71b-4aa1-9781-77a864c59e59 tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Deleted allocations for instance 5f645eaa-6edc-4362-82ad-38c4c57b2be2 [ 843.171769] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f1ca172d-6df8-4226-97a1-0d27f306cf5b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.196929] env[62066]: DEBUG oslo_vmware.api [None req-ef8a6667-cf2e-41c6-b3b6-d096c3290f66 tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Waiting for the task: (returnval){ [ 843.196929] env[62066]: value = "task-1340911" [ 843.196929] env[62066]: _type = "Task" [ 843.196929] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.205722] env[62066]: DEBUG oslo_vmware.api [None req-ef8a6667-cf2e-41c6-b3b6-d096c3290f66 tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Task: {'id': task-1340911, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.376764] env[62066]: DEBUG oslo_concurrency.lockutils [None req-63dd5fc5-276d-49c0-99ba-71a261a8efa4 tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 843.411045] env[62066]: DEBUG oslo_vmware.api [None req-e3db032e-002b-4e25-a569-0fed0f11200b tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]521b466c-9983-ea52-4823-2bedef8f7ecf, 'name': SearchDatastore_Task, 'duration_secs': 0.010428} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.412169] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e3db032e-002b-4e25-a569-0fed0f11200b tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Releasing lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 843.412604] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-e3db032e-002b-4e25-a569-0fed0f11200b tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: b08f27b1-429e-4d05-9a95-f45fd0cb2904] Processing image 50ff584c-3b50-4395-af07-3e66769bc9f7 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 843.413322] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e3db032e-002b-4e25-a569-0fed0f11200b tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 843.413322] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e3db032e-002b-4e25-a569-0fed0f11200b tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Acquired lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 843.413322] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-e3db032e-002b-4e25-a569-0fed0f11200b tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 843.414176] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4ecd4e83-0898-46a3-b542-d5dfa7a44fee {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.425102] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-e3db032e-002b-4e25-a569-0fed0f11200b tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 843.425612] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-e3db032e-002b-4e25-a569-0fed0f11200b tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 843.426730] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-06797cfe-e7ce-4ad2-b60b-96bd9dd9da43 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.434511] env[62066]: DEBUG oslo_vmware.api [None req-e3db032e-002b-4e25-a569-0fed0f11200b tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Waiting for the task: (returnval){ [ 843.434511] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]5248bc5d-e9dd-3a8c-ee97-e4498b232100" [ 843.434511] env[62066]: _type = "Task" [ 843.434511] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.447957] env[62066]: DEBUG oslo_vmware.api [None req-e3db032e-002b-4e25-a569-0fed0f11200b tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]5248bc5d-e9dd-3a8c-ee97-e4498b232100, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.483453] env[62066]: DEBUG nova.network.neutron [req-22c69ada-eeaa-44ad-9830-b8e171c9f91e req-595929fe-d3e7-40f1-8d16-e9e986d2911f service nova] [instance: b08f27b1-429e-4d05-9a95-f45fd0cb2904] Updated VIF entry in instance network info cache for port 62d600d9-8ad3-43f8-a2b0-050747916fdb. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 843.483890] env[62066]: DEBUG nova.network.neutron [req-22c69ada-eeaa-44ad-9830-b8e171c9f91e req-595929fe-d3e7-40f1-8d16-e9e986d2911f service nova] [instance: b08f27b1-429e-4d05-9a95-f45fd0cb2904] Updating instance_info_cache with network_info: [{"id": "62d600d9-8ad3-43f8-a2b0-050747916fdb", "address": "fa:16:3e:51:fd:46", "network": {"id": "a3852490-7a1f-499c-813a-46cbf52318da", "bridge": "br-int", "label": "tempest-ImagesTestJSON-735840736-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "75f0352c852947369474c9e05766c584", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "50171613-b419-45e3-9ada-fcb6cd921428", "external-id": "nsx-vlan-transportzone-914", "segmentation_id": 914, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap62d600d9-8a", "ovs_interfaceid": "62d600d9-8ad3-43f8-a2b0-050747916fdb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 843.694258] env[62066]: DEBUG oslo_concurrency.lockutils [None req-6075adee-b71b-4aa1-9781-77a864c59e59 tempest-SecurityGroupsTestJSON-304471880 tempest-SecurityGroupsTestJSON-304471880-project-member] Lock "5f645eaa-6edc-4362-82ad-38c4c57b2be2" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.409s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 843.710807] env[62066]: DEBUG oslo_vmware.api [None req-ef8a6667-cf2e-41c6-b3b6-d096c3290f66 tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Task: {'id': task-1340911, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.946370] env[62066]: DEBUG oslo_vmware.api [None req-e3db032e-002b-4e25-a569-0fed0f11200b tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]5248bc5d-e9dd-3a8c-ee97-e4498b232100, 'name': SearchDatastore_Task, 'duration_secs': 0.018712} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.947839] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a551f308-c1d7-4df2-8016-fcf5cac94a69 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.955415] env[62066]: DEBUG oslo_vmware.api [None req-e3db032e-002b-4e25-a569-0fed0f11200b tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Waiting for the task: (returnval){ [ 843.955415] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]525915e7-0405-ff2e-544f-15d84bf2acbd" [ 843.955415] env[62066]: _type = "Task" [ 843.955415] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.966012] env[62066]: DEBUG oslo_vmware.api [None req-e3db032e-002b-4e25-a569-0fed0f11200b tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]525915e7-0405-ff2e-544f-15d84bf2acbd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.988805] env[62066]: DEBUG oslo_concurrency.lockutils [req-22c69ada-eeaa-44ad-9830-b8e171c9f91e req-595929fe-d3e7-40f1-8d16-e9e986d2911f service nova] Releasing lock "refresh_cache-b08f27b1-429e-4d05-9a95-f45fd0cb2904" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 843.989107] env[62066]: DEBUG nova.compute.manager [req-22c69ada-eeaa-44ad-9830-b8e171c9f91e req-595929fe-d3e7-40f1-8d16-e9e986d2911f service nova] [instance: 86a92b61-f3e9-48f9-8ee2-756669d558ef] Received event network-vif-deleted-fd69e6c7-721d-4947-80e1-4db6e9a230b0 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 844.139530] env[62066]: DEBUG oslo_concurrency.lockutils [None req-68383c37-9698-4391-bbf2-b9ebf656dc03 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Acquiring lock "c8e0d47c-4421-4e00-9183-206fceeabc40" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 844.140122] env[62066]: DEBUG oslo_concurrency.lockutils [None req-68383c37-9698-4391-bbf2-b9ebf656dc03 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Lock "c8e0d47c-4421-4e00-9183-206fceeabc40" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 844.209073] env[62066]: DEBUG oslo_vmware.api [None req-ef8a6667-cf2e-41c6-b3b6-d096c3290f66 tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Task: {'id': task-1340911, 'name': ReconfigVM_Task, 'duration_secs': 0.623731} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.211818] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-ef8a6667-cf2e-41c6-b3b6-d096c3290f66 tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] [instance: 5e96def2-0cbd-4bd9-93f4-6a365a0142b0] Reconfigured VM instance instance-00000049 to attach disk [datastore1] 5e96def2-0cbd-4bd9-93f4-6a365a0142b0/50ff584c-3b50-4395-af07-3e66769bc9f7-rescue.vmdk or device None with type thin {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 844.212964] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-793833fb-599b-4901-9ff8-4ee94f4f23ba {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.242000] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3a9159e1-8a3b-4b03-9447-ead6599474e9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.257892] env[62066]: DEBUG oslo_vmware.api [None req-ef8a6667-cf2e-41c6-b3b6-d096c3290f66 tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Waiting for the task: (returnval){ [ 844.257892] env[62066]: value = "task-1340912" [ 844.257892] env[62066]: _type = "Task" [ 844.257892] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.266958] env[62066]: DEBUG oslo_vmware.api [None req-ef8a6667-cf2e-41c6-b3b6-d096c3290f66 tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Task: {'id': task-1340912, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.384705] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1dd6435e-5a62-4e44-bea1-4ade72096a9d tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Acquiring lock "interface-459f432b-e817-405f-9c3e-dfa5d957788c-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 844.385122] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1dd6435e-5a62-4e44-bea1-4ade72096a9d tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Lock "interface-459f432b-e817-405f-9c3e-dfa5d957788c-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 844.385529] env[62066]: DEBUG nova.objects.instance [None req-1dd6435e-5a62-4e44-bea1-4ade72096a9d tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Lazy-loading 'flavor' on Instance uuid 459f432b-e817-405f-9c3e-dfa5d957788c {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 844.398676] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88a595da-5e1a-4a4e-b233-6aab66cf562d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.407364] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af11ffbf-3f5e-4108-80b4-862d5cb5fc6a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.439945] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5615ca43-5157-4e54-bcc7-b9e684503f45 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Acquiring lock "88daaaa6-f385-4161-bc74-9fffabb1145d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 844.440249] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5615ca43-5157-4e54-bcc7-b9e684503f45 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Lock "88daaaa6-f385-4161-bc74-9fffabb1145d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 844.440465] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5615ca43-5157-4e54-bcc7-b9e684503f45 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Acquiring lock "88daaaa6-f385-4161-bc74-9fffabb1145d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 844.440651] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5615ca43-5157-4e54-bcc7-b9e684503f45 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Lock "88daaaa6-f385-4161-bc74-9fffabb1145d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 844.440827] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5615ca43-5157-4e54-bcc7-b9e684503f45 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Lock "88daaaa6-f385-4161-bc74-9fffabb1145d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 844.443210] env[62066]: INFO nova.compute.manager [None req-5615ca43-5157-4e54-bcc7-b9e684503f45 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 88daaaa6-f385-4161-bc74-9fffabb1145d] Terminating instance [ 844.445463] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e667dae3-50e0-4cf2-b718-626b07e66b8a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.448734] env[62066]: DEBUG nova.compute.manager [None req-5615ca43-5157-4e54-bcc7-b9e684503f45 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 88daaaa6-f385-4161-bc74-9fffabb1145d] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 844.448935] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-5615ca43-5157-4e54-bcc7-b9e684503f45 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 88daaaa6-f385-4161-bc74-9fffabb1145d] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 844.449772] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84c51e31-2510-43c6-9f46-f7148eeb3de1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.463510] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c5386dc-c24f-4f19-a574-9bffed19ab09 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.470867] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-5615ca43-5157-4e54-bcc7-b9e684503f45 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 88daaaa6-f385-4161-bc74-9fffabb1145d] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 844.471465] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-67bb7d3e-6a3f-425f-b472-cbbee0a376f3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.487040] env[62066]: DEBUG oslo_vmware.api [None req-e3db032e-002b-4e25-a569-0fed0f11200b tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]525915e7-0405-ff2e-544f-15d84bf2acbd, 'name': SearchDatastore_Task, 'duration_secs': 0.010392} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.487040] env[62066]: DEBUG nova.compute.provider_tree [None req-f63339cf-eb8a-4d19-b0c5-1ac7744e7482 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 844.487040] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e3db032e-002b-4e25-a569-0fed0f11200b tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Releasing lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 844.487040] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-e3db032e-002b-4e25-a569-0fed0f11200b tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk to [datastore2] b08f27b1-429e-4d05-9a95-f45fd0cb2904/b08f27b1-429e-4d05-9a95-f45fd0cb2904.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 844.487573] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6c4ffad4-0ccf-4b69-bc26-3d7107082309 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.490621] env[62066]: DEBUG oslo_vmware.api [None req-5615ca43-5157-4e54-bcc7-b9e684503f45 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Waiting for the task: (returnval){ [ 844.490621] env[62066]: value = "task-1340913" [ 844.490621] env[62066]: _type = "Task" [ 844.490621] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.496943] env[62066]: DEBUG oslo_vmware.api [None req-e3db032e-002b-4e25-a569-0fed0f11200b tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Waiting for the task: (returnval){ [ 844.496943] env[62066]: value = "task-1340914" [ 844.496943] env[62066]: _type = "Task" [ 844.496943] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.503304] env[62066]: DEBUG oslo_vmware.api [None req-5615ca43-5157-4e54-bcc7-b9e684503f45 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1340913, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.509160] env[62066]: DEBUG oslo_vmware.api [None req-e3db032e-002b-4e25-a569-0fed0f11200b tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': task-1340914, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.645423] env[62066]: DEBUG nova.compute.manager [None req-68383c37-9698-4391-bbf2-b9ebf656dc03 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: c8e0d47c-4421-4e00-9183-206fceeabc40] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 844.769917] env[62066]: DEBUG oslo_vmware.api [None req-ef8a6667-cf2e-41c6-b3b6-d096c3290f66 tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Task: {'id': task-1340912, 'name': ReconfigVM_Task, 'duration_secs': 0.150769} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.770234] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-ef8a6667-cf2e-41c6-b3b6-d096c3290f66 tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] [instance: 5e96def2-0cbd-4bd9-93f4-6a365a0142b0] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 844.770502] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0f23cec4-c69a-4b11-8a2b-87ee8facf4a9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.776728] env[62066]: DEBUG oslo_vmware.api [None req-ef8a6667-cf2e-41c6-b3b6-d096c3290f66 tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Waiting for the task: (returnval){ [ 844.776728] env[62066]: value = "task-1340915" [ 844.776728] env[62066]: _type = "Task" [ 844.776728] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.784604] env[62066]: DEBUG oslo_vmware.api [None req-ef8a6667-cf2e-41c6-b3b6-d096c3290f66 tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Task: {'id': task-1340915, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.960178] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cf3b688f-a077-42d2-b46f-272004213881 tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] Acquiring lock "daffaf51-4c45-44aa-8fc2-4db066a09971" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 844.960526] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cf3b688f-a077-42d2-b46f-272004213881 tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] Lock "daffaf51-4c45-44aa-8fc2-4db066a09971" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 844.960784] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cf3b688f-a077-42d2-b46f-272004213881 tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] Acquiring lock "daffaf51-4c45-44aa-8fc2-4db066a09971-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 844.961654] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cf3b688f-a077-42d2-b46f-272004213881 tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] Lock "daffaf51-4c45-44aa-8fc2-4db066a09971-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 844.961654] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cf3b688f-a077-42d2-b46f-272004213881 tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] Lock "daffaf51-4c45-44aa-8fc2-4db066a09971-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 844.964940] env[62066]: INFO nova.compute.manager [None req-cf3b688f-a077-42d2-b46f-272004213881 tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] Terminating instance [ 844.968514] env[62066]: DEBUG nova.compute.manager [None req-cf3b688f-a077-42d2-b46f-272004213881 tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 844.968514] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e6e2da52-2e25-4f2c-a6a4-6fc5ac6caee8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.978022] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0717d19f-657d-4826-8f15-bff3f094c51a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.988504] env[62066]: DEBUG nova.objects.instance [None req-1dd6435e-5a62-4e44-bea1-4ade72096a9d tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Lazy-loading 'pci_requests' on Instance uuid 459f432b-e817-405f-9c3e-dfa5d957788c {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 844.990321] env[62066]: DEBUG nova.scheduler.client.report [None req-f63339cf-eb8a-4d19-b0c5-1ac7744e7482 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 845.002187] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f2684bd2-e68b-4819-84d6-b5b0e8bf10bc tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Acquiring lock "76306c3f-0674-40fe-9864-d82d11e4bb5e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 845.002588] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f2684bd2-e68b-4819-84d6-b5b0e8bf10bc tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Lock "76306c3f-0674-40fe-9864-d82d11e4bb5e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 845.017102] env[62066]: WARNING nova.virt.vmwareapi.driver [None req-cf3b688f-a077-42d2-b46f-272004213881 tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] Instance does not exists. Proceeding to delete instance properties on datastore: nova.exception.InstanceNotFound: Instance daffaf51-4c45-44aa-8fc2-4db066a09971 could not be found. [ 845.017541] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-cf3b688f-a077-42d2-b46f-272004213881 tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 845.018501] env[62066]: DEBUG nova.compute.manager [None req-f2684bd2-e68b-4819-84d6-b5b0e8bf10bc tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 76306c3f-0674-40fe-9864-d82d11e4bb5e] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 845.021793] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2e406b9f-1a55-423d-82f5-3633c4fb1d4c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.027367] env[62066]: DEBUG oslo_vmware.api [None req-5615ca43-5157-4e54-bcc7-b9e684503f45 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1340913, 'name': PowerOffVM_Task, 'duration_secs': 0.208932} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.028011] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-5615ca43-5157-4e54-bcc7-b9e684503f45 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 88daaaa6-f385-4161-bc74-9fffabb1145d] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 845.028230] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-5615ca43-5157-4e54-bcc7-b9e684503f45 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 88daaaa6-f385-4161-bc74-9fffabb1145d] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 845.031382] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9069ca02-7ae5-4818-bd28-66d2d4e17f83 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.033744] env[62066]: DEBUG oslo_vmware.api [None req-e3db032e-002b-4e25-a569-0fed0f11200b tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': task-1340914, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.037684] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-804517a6-32ed-42bb-96ce-7d3f370eb173 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.067815] env[62066]: WARNING nova.virt.vmwareapi.vmops [None req-cf3b688f-a077-42d2-b46f-272004213881 tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance daffaf51-4c45-44aa-8fc2-4db066a09971 could not be found. [ 845.068008] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-cf3b688f-a077-42d2-b46f-272004213881 tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 845.068223] env[62066]: INFO nova.compute.manager [None req-cf3b688f-a077-42d2-b46f-272004213881 tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] Took 0.10 seconds to destroy the instance on the hypervisor. [ 845.068473] env[62066]: DEBUG oslo.service.loopingcall [None req-cf3b688f-a077-42d2-b46f-272004213881 tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 845.068772] env[62066]: DEBUG nova.compute.manager [-] [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 845.068865] env[62066]: DEBUG nova.network.neutron [-] [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 845.092834] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-5615ca43-5157-4e54-bcc7-b9e684503f45 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 88daaaa6-f385-4161-bc74-9fffabb1145d] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 845.093046] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-5615ca43-5157-4e54-bcc7-b9e684503f45 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 88daaaa6-f385-4161-bc74-9fffabb1145d] Deleting contents of the VM from datastore datastore1 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 845.093246] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-5615ca43-5157-4e54-bcc7-b9e684503f45 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Deleting the datastore file [datastore1] 88daaaa6-f385-4161-bc74-9fffabb1145d {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 845.093507] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-df6b14e9-f882-452e-ae03-9cedf0852ac8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.099658] env[62066]: DEBUG oslo_vmware.api [None req-5615ca43-5157-4e54-bcc7-b9e684503f45 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Waiting for the task: (returnval){ [ 845.099658] env[62066]: value = "task-1340917" [ 845.099658] env[62066]: _type = "Task" [ 845.099658] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.109946] env[62066]: DEBUG oslo_vmware.api [None req-5615ca43-5157-4e54-bcc7-b9e684503f45 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1340917, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.171961] env[62066]: DEBUG oslo_concurrency.lockutils [None req-68383c37-9698-4391-bbf2-b9ebf656dc03 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 845.290331] env[62066]: DEBUG oslo_vmware.api [None req-ef8a6667-cf2e-41c6-b3b6-d096c3290f66 tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Task: {'id': task-1340915, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.495576] env[62066]: DEBUG nova.objects.base [None req-1dd6435e-5a62-4e44-bea1-4ade72096a9d tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Object Instance<459f432b-e817-405f-9c3e-dfa5d957788c> lazy-loaded attributes: flavor,pci_requests {{(pid=62066) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 845.495813] env[62066]: DEBUG nova.network.neutron [None req-1dd6435e-5a62-4e44-bea1-4ade72096a9d tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: 459f432b-e817-405f-9c3e-dfa5d957788c] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 845.501118] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f63339cf-eb8a-4d19-b0c5-1ac7744e7482 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.369s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 845.502279] env[62066]: DEBUG nova.compute.manager [None req-f63339cf-eb8a-4d19-b0c5-1ac7744e7482 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] [instance: d77e3405-d158-4a8c-9f54-d5fbd26fed48] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 845.504713] env[62066]: DEBUG oslo_concurrency.lockutils [None req-22d0b1dd-f08e-49e4-a262-9fed18806112 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.739s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 845.505970] env[62066]: INFO nova.compute.claims [None req-22d0b1dd-f08e-49e4-a262-9fed18806112 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] [instance: 3120c9a5-0558-4e83-b0fb-994a989398fb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 845.531848] env[62066]: DEBUG oslo_vmware.api [None req-e3db032e-002b-4e25-a569-0fed0f11200b tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': task-1340914, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.549625] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f2684bd2-e68b-4819-84d6-b5b0e8bf10bc tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 845.580044] env[62066]: DEBUG nova.policy [None req-1dd6435e-5a62-4e44-bea1-4ade72096a9d tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '747026d272d4467bbf14a0c30589c3b7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0d4e33a99c7741fb8cdd97f4ec5dbbd8', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 845.610521] env[62066]: DEBUG oslo_vmware.api [None req-5615ca43-5157-4e54-bcc7-b9e684503f45 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1340917, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.147133} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.610803] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-5615ca43-5157-4e54-bcc7-b9e684503f45 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 845.610991] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-5615ca43-5157-4e54-bcc7-b9e684503f45 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 88daaaa6-f385-4161-bc74-9fffabb1145d] Deleted contents of the VM from datastore datastore1 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 845.611265] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-5615ca43-5157-4e54-bcc7-b9e684503f45 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 88daaaa6-f385-4161-bc74-9fffabb1145d] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 845.612420] env[62066]: INFO nova.compute.manager [None req-5615ca43-5157-4e54-bcc7-b9e684503f45 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 88daaaa6-f385-4161-bc74-9fffabb1145d] Took 1.16 seconds to destroy the instance on the hypervisor. [ 845.612420] env[62066]: DEBUG oslo.service.loopingcall [None req-5615ca43-5157-4e54-bcc7-b9e684503f45 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 845.612420] env[62066]: DEBUG nova.compute.manager [-] [instance: 88daaaa6-f385-4161-bc74-9fffabb1145d] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 845.612420] env[62066]: DEBUG nova.network.neutron [-] [instance: 88daaaa6-f385-4161-bc74-9fffabb1145d] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 845.710027] env[62066]: DEBUG nova.compute.manager [req-1d8b16ee-edbc-4fd8-89c5-0bdbbd060f3e req-b80d4f11-a956-4387-989e-c82beb190342 service nova] [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] Received event network-vif-deleted-88380b24-b187-4b82-b92e-b93adeeab3eb {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 845.710367] env[62066]: INFO nova.compute.manager [req-1d8b16ee-edbc-4fd8-89c5-0bdbbd060f3e req-b80d4f11-a956-4387-989e-c82beb190342 service nova] [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] Neutron deleted interface 88380b24-b187-4b82-b92e-b93adeeab3eb; detaching it from the instance and deleting it from the info cache [ 845.710461] env[62066]: DEBUG nova.network.neutron [req-1d8b16ee-edbc-4fd8-89c5-0bdbbd060f3e req-b80d4f11-a956-4387-989e-c82beb190342 service nova] [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 845.791511] env[62066]: DEBUG oslo_vmware.api [None req-ef8a6667-cf2e-41c6-b3b6-d096c3290f66 tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Task: {'id': task-1340915, 'name': PowerOnVM_Task} progress is 90%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.013692] env[62066]: DEBUG nova.compute.utils [None req-f63339cf-eb8a-4d19-b0c5-1ac7744e7482 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 846.020579] env[62066]: DEBUG nova.compute.manager [None req-f63339cf-eb8a-4d19-b0c5-1ac7744e7482 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] [instance: d77e3405-d158-4a8c-9f54-d5fbd26fed48] Not allocating networking since 'none' was specified. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 846.028020] env[62066]: DEBUG oslo_vmware.api [None req-e3db032e-002b-4e25-a569-0fed0f11200b tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': task-1340914, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.398131} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.028020] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-e3db032e-002b-4e25-a569-0fed0f11200b tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk to [datastore2] b08f27b1-429e-4d05-9a95-f45fd0cb2904/b08f27b1-429e-4d05-9a95-f45fd0cb2904.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 846.028020] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-e3db032e-002b-4e25-a569-0fed0f11200b tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: b08f27b1-429e-4d05-9a95-f45fd0cb2904] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 846.028355] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b637612a-4558-460f-8a52-aa00d1b7e9ed {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.035382] env[62066]: DEBUG oslo_vmware.api [None req-e3db032e-002b-4e25-a569-0fed0f11200b tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Waiting for the task: (returnval){ [ 846.035382] env[62066]: value = "task-1340918" [ 846.035382] env[62066]: _type = "Task" [ 846.035382] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.051113] env[62066]: DEBUG oslo_vmware.api [None req-e3db032e-002b-4e25-a569-0fed0f11200b tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': task-1340918, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.100097] env[62066]: DEBUG nova.network.neutron [None req-1dd6435e-5a62-4e44-bea1-4ade72096a9d tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: 459f432b-e817-405f-9c3e-dfa5d957788c] Successfully created port: 56efe6b6-2fe3-4c7b-9d59-f8ecf74ac772 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 846.122615] env[62066]: DEBUG nova.network.neutron [-] [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 846.212996] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a8a4445d-1b4f-4def-ab07-4e29fa86775f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.223148] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94847882-eb15-44aa-8f5a-12a569f52902 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.260172] env[62066]: DEBUG nova.compute.manager [req-1d8b16ee-edbc-4fd8-89c5-0bdbbd060f3e req-b80d4f11-a956-4387-989e-c82beb190342 service nova] [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] Detach interface failed, port_id=88380b24-b187-4b82-b92e-b93adeeab3eb, reason: Instance daffaf51-4c45-44aa-8fc2-4db066a09971 could not be found. {{(pid=62066) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 846.290556] env[62066]: DEBUG oslo_vmware.api [None req-ef8a6667-cf2e-41c6-b3b6-d096c3290f66 tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Task: {'id': task-1340915, 'name': PowerOnVM_Task, 'duration_secs': 1.177861} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.291035] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-ef8a6667-cf2e-41c6-b3b6-d096c3290f66 tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] [instance: 5e96def2-0cbd-4bd9-93f4-6a365a0142b0] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 846.295219] env[62066]: DEBUG nova.compute.manager [None req-ef8a6667-cf2e-41c6-b3b6-d096c3290f66 tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] [instance: 5e96def2-0cbd-4bd9-93f4-6a365a0142b0] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 846.296120] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-242b000a-8eff-4a5f-9a0b-90f37eea84b3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.524355] env[62066]: DEBUG nova.compute.manager [None req-f63339cf-eb8a-4d19-b0c5-1ac7744e7482 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] [instance: d77e3405-d158-4a8c-9f54-d5fbd26fed48] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 846.549021] env[62066]: DEBUG oslo_vmware.api [None req-e3db032e-002b-4e25-a569-0fed0f11200b tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': task-1340918, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071048} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.549812] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-e3db032e-002b-4e25-a569-0fed0f11200b tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: b08f27b1-429e-4d05-9a95-f45fd0cb2904] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 846.551030] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48b0fddf-4789-4024-9513-fbccdef221ac {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.580381] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-e3db032e-002b-4e25-a569-0fed0f11200b tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: b08f27b1-429e-4d05-9a95-f45fd0cb2904] Reconfiguring VM instance instance-0000004b to attach disk [datastore2] b08f27b1-429e-4d05-9a95-f45fd0cb2904/b08f27b1-429e-4d05-9a95-f45fd0cb2904.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 846.580858] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2c31bc62-627d-447a-8f66-487a43ed2944 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.596649] env[62066]: DEBUG nova.network.neutron [-] [instance: 88daaaa6-f385-4161-bc74-9fffabb1145d] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 846.603834] env[62066]: DEBUG oslo_vmware.api [None req-e3db032e-002b-4e25-a569-0fed0f11200b tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Waiting for the task: (returnval){ [ 846.603834] env[62066]: value = "task-1340919" [ 846.603834] env[62066]: _type = "Task" [ 846.603834] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.621149] env[62066]: DEBUG oslo_vmware.api [None req-e3db032e-002b-4e25-a569-0fed0f11200b tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': task-1340919, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.625203] env[62066]: INFO nova.compute.manager [-] [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] Took 1.56 seconds to deallocate network for instance. [ 846.857903] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19d91afe-d21b-430d-b149-9334363d83d1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.867232] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b63021f-8ce2-478c-9acb-6214800b6103 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.903585] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1683913e-36f8-466d-821a-bc8d81a80f74 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.911804] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b32fb36-6609-4b05-ac01-bafcc7c17b77 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.925621] env[62066]: DEBUG nova.compute.provider_tree [None req-22d0b1dd-f08e-49e4-a262-9fed18806112 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 847.100974] env[62066]: INFO nova.compute.manager [-] [instance: 88daaaa6-f385-4161-bc74-9fffabb1145d] Took 1.49 seconds to deallocate network for instance. [ 847.117254] env[62066]: DEBUG oslo_vmware.api [None req-e3db032e-002b-4e25-a569-0fed0f11200b tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': task-1340919, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.186930] env[62066]: INFO nova.compute.manager [None req-cf3b688f-a077-42d2-b46f-272004213881 tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] Took 0.56 seconds to detach 1 volumes for instance. [ 847.189372] env[62066]: DEBUG nova.compute.manager [None req-cf3b688f-a077-42d2-b46f-272004213881 tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] Deleting volume: 26ceca45-1cae-48c9-a567-e7cc4f9970c1 {{(pid=62066) _cleanup_volumes /opt/stack/nova/nova/compute/manager.py:3247}} [ 847.430145] env[62066]: DEBUG nova.scheduler.client.report [None req-22d0b1dd-f08e-49e4-a262-9fed18806112 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 847.539146] env[62066]: DEBUG nova.compute.manager [None req-f63339cf-eb8a-4d19-b0c5-1ac7744e7482 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] [instance: d77e3405-d158-4a8c-9f54-d5fbd26fed48] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 847.576022] env[62066]: DEBUG nova.virt.hardware [None req-f63339cf-eb8a-4d19-b0c5-1ac7744e7482 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-23T13:40:41Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-23T13:40:23Z,direct_url=,disk_format='vmdk',id=50ff584c-3b50-4395-af07-3e66769bc9f7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='eb52f7069a374c61ae946f052007c6d9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-23T13:40:24Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 847.576022] env[62066]: DEBUG nova.virt.hardware [None req-f63339cf-eb8a-4d19-b0c5-1ac7744e7482 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 847.576022] env[62066]: DEBUG nova.virt.hardware [None req-f63339cf-eb8a-4d19-b0c5-1ac7744e7482 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 847.576022] env[62066]: DEBUG nova.virt.hardware [None req-f63339cf-eb8a-4d19-b0c5-1ac7744e7482 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 847.576022] env[62066]: DEBUG nova.virt.hardware [None req-f63339cf-eb8a-4d19-b0c5-1ac7744e7482 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 847.576022] env[62066]: DEBUG nova.virt.hardware [None req-f63339cf-eb8a-4d19-b0c5-1ac7744e7482 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 847.576022] env[62066]: DEBUG nova.virt.hardware [None req-f63339cf-eb8a-4d19-b0c5-1ac7744e7482 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 847.576601] env[62066]: DEBUG nova.virt.hardware [None req-f63339cf-eb8a-4d19-b0c5-1ac7744e7482 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 847.576920] env[62066]: DEBUG nova.virt.hardware [None req-f63339cf-eb8a-4d19-b0c5-1ac7744e7482 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 847.577226] env[62066]: DEBUG nova.virt.hardware [None req-f63339cf-eb8a-4d19-b0c5-1ac7744e7482 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 847.577587] env[62066]: DEBUG nova.virt.hardware [None req-f63339cf-eb8a-4d19-b0c5-1ac7744e7482 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 847.578862] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e547a725-e8ae-4b77-9b68-46808e4bfc10 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.588319] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab314daa-249a-44c1-9cee-ab1ee55d3ac3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.610171] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5615ca43-5157-4e54-bcc7-b9e684503f45 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 847.610609] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-f63339cf-eb8a-4d19-b0c5-1ac7744e7482 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] [instance: d77e3405-d158-4a8c-9f54-d5fbd26fed48] Instance VIF info [] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 847.617203] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-f63339cf-eb8a-4d19-b0c5-1ac7744e7482 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Creating folder: Project (fe38d2d8cce14298a29fb40771407cb1). Parent ref: group-v285980. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 847.622112] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-cc1bc543-015d-4a13-82ea-3f8112168b1c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.630049] env[62066]: DEBUG oslo_vmware.api [None req-e3db032e-002b-4e25-a569-0fed0f11200b tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': task-1340919, 'name': ReconfigVM_Task, 'duration_secs': 0.539817} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.630594] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-e3db032e-002b-4e25-a569-0fed0f11200b tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: b08f27b1-429e-4d05-9a95-f45fd0cb2904] Reconfigured VM instance instance-0000004b to attach disk [datastore2] b08f27b1-429e-4d05-9a95-f45fd0cb2904/b08f27b1-429e-4d05-9a95-f45fd0cb2904.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 847.631416] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8a1c0c0f-3023-46d9-8d85-f0f8e472a2af {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.634590] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-f63339cf-eb8a-4d19-b0c5-1ac7744e7482 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Created folder: Project (fe38d2d8cce14298a29fb40771407cb1) in parent group-v285980. [ 847.634916] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-f63339cf-eb8a-4d19-b0c5-1ac7744e7482 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Creating folder: Instances. Parent ref: group-v286064. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 847.635655] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3fe9ffc1-3633-4ac0-b028-a7749d4ddbe7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.639997] env[62066]: DEBUG oslo_vmware.api [None req-e3db032e-002b-4e25-a569-0fed0f11200b tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Waiting for the task: (returnval){ [ 847.639997] env[62066]: value = "task-1340922" [ 847.639997] env[62066]: _type = "Task" [ 847.639997] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.653402] env[62066]: DEBUG oslo_vmware.api [None req-e3db032e-002b-4e25-a569-0fed0f11200b tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': task-1340922, 'name': Rename_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.655935] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-f63339cf-eb8a-4d19-b0c5-1ac7744e7482 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Created folder: Instances in parent group-v286064. [ 847.656377] env[62066]: DEBUG oslo.service.loopingcall [None req-f63339cf-eb8a-4d19-b0c5-1ac7744e7482 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 847.656866] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d77e3405-d158-4a8c-9f54-d5fbd26fed48] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 847.657221] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4f5aef54-bca0-4b16-aa46-3f4c2a4e4334 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.676791] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 847.676791] env[62066]: value = "task-1340924" [ 847.676791] env[62066]: _type = "Task" [ 847.676791] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.684502] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1340924, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.732689] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cf3b688f-a077-42d2-b46f-272004213881 tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 847.790813] env[62066]: DEBUG nova.compute.manager [req-74639cc8-e025-4eef-9afc-801475559e0a req-3607c7c3-b005-4180-9cfd-1605fc80f45a service nova] [instance: 88daaaa6-f385-4161-bc74-9fffabb1145d] Received event network-vif-deleted-ba002f90-5218-46f2-b7e1-cd96ae3018c7 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 847.934220] env[62066]: DEBUG oslo_concurrency.lockutils [None req-22d0b1dd-f08e-49e4-a262-9fed18806112 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.430s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 847.934750] env[62066]: DEBUG nova.compute.manager [None req-22d0b1dd-f08e-49e4-a262-9fed18806112 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] [instance: 3120c9a5-0558-4e83-b0fb-994a989398fb] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 847.938556] env[62066]: DEBUG oslo_concurrency.lockutils [None req-73e4bc46-8266-447d-984c-3ca2b99ffd23 tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.776s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 847.940071] env[62066]: INFO nova.compute.claims [None req-73e4bc46-8266-447d-984c-3ca2b99ffd23 tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] [instance: 0bc31a8b-a5fc-4f6e-a4c0-024bf4260755] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 848.010623] env[62066]: DEBUG oslo_concurrency.lockutils [None req-59107a59-1a77-47be-8dae-1ec8f16f1fda tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Acquiring lock "2634fcc0-96bd-4513-8f92-515f83fe23d7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 848.010951] env[62066]: DEBUG oslo_concurrency.lockutils [None req-59107a59-1a77-47be-8dae-1ec8f16f1fda tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Lock "2634fcc0-96bd-4513-8f92-515f83fe23d7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 848.126913] env[62066]: DEBUG nova.network.neutron [None req-1dd6435e-5a62-4e44-bea1-4ade72096a9d tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: 459f432b-e817-405f-9c3e-dfa5d957788c] Successfully updated port: 56efe6b6-2fe3-4c7b-9d59-f8ecf74ac772 {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 848.151993] env[62066]: DEBUG oslo_vmware.api [None req-e3db032e-002b-4e25-a569-0fed0f11200b tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': task-1340922, 'name': Rename_Task, 'duration_secs': 0.23652} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.153326] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-e3db032e-002b-4e25-a569-0fed0f11200b tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: b08f27b1-429e-4d05-9a95-f45fd0cb2904] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 848.153644] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-074e99d7-b616-476c-9b1d-01cd02435324 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.161354] env[62066]: DEBUG oslo_vmware.api [None req-e3db032e-002b-4e25-a569-0fed0f11200b tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Waiting for the task: (returnval){ [ 848.161354] env[62066]: value = "task-1340925" [ 848.161354] env[62066]: _type = "Task" [ 848.161354] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.170640] env[62066]: DEBUG oslo_vmware.api [None req-e3db032e-002b-4e25-a569-0fed0f11200b tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': task-1340925, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.187121] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1340924, 'name': CreateVM_Task, 'duration_secs': 0.334589} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.187320] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d77e3405-d158-4a8c-9f54-d5fbd26fed48] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 848.187767] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f63339cf-eb8a-4d19-b0c5-1ac7744e7482 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 848.187932] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f63339cf-eb8a-4d19-b0c5-1ac7744e7482 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Acquired lock "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 848.188306] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f63339cf-eb8a-4d19-b0c5-1ac7744e7482 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 848.188564] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1f26148a-0bb6-4e50-8246-5185730cc15c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.196491] env[62066]: DEBUG oslo_vmware.api [None req-f63339cf-eb8a-4d19-b0c5-1ac7744e7482 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Waiting for the task: (returnval){ [ 848.196491] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]52062ff9-4b6c-ec83-781b-86665687a525" [ 848.196491] env[62066]: _type = "Task" [ 848.196491] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.205889] env[62066]: DEBUG oslo_vmware.api [None req-f63339cf-eb8a-4d19-b0c5-1ac7744e7482 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52062ff9-4b6c-ec83-781b-86665687a525, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.444492] env[62066]: DEBUG nova.compute.utils [None req-22d0b1dd-f08e-49e4-a262-9fed18806112 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 848.446081] env[62066]: DEBUG nova.compute.manager [None req-22d0b1dd-f08e-49e4-a262-9fed18806112 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] [instance: 3120c9a5-0558-4e83-b0fb-994a989398fb] Not allocating networking since 'none' was specified. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 848.515990] env[62066]: DEBUG nova.compute.manager [None req-59107a59-1a77-47be-8dae-1ec8f16f1fda tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] [instance: 2634fcc0-96bd-4513-8f92-515f83fe23d7] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 848.629942] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1dd6435e-5a62-4e44-bea1-4ade72096a9d tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Acquiring lock "refresh_cache-459f432b-e817-405f-9c3e-dfa5d957788c" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 848.634023] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1dd6435e-5a62-4e44-bea1-4ade72096a9d tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Acquired lock "refresh_cache-459f432b-e817-405f-9c3e-dfa5d957788c" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 848.634023] env[62066]: DEBUG nova.network.neutron [None req-1dd6435e-5a62-4e44-bea1-4ade72096a9d tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: 459f432b-e817-405f-9c3e-dfa5d957788c] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 848.678508] env[62066]: DEBUG oslo_vmware.api [None req-e3db032e-002b-4e25-a569-0fed0f11200b tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': task-1340925, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.715328] env[62066]: DEBUG oslo_vmware.api [None req-f63339cf-eb8a-4d19-b0c5-1ac7744e7482 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52062ff9-4b6c-ec83-781b-86665687a525, 'name': SearchDatastore_Task, 'duration_secs': 0.011175} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.715855] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f63339cf-eb8a-4d19-b0c5-1ac7744e7482 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Releasing lock "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 848.716259] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-f63339cf-eb8a-4d19-b0c5-1ac7744e7482 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] [instance: d77e3405-d158-4a8c-9f54-d5fbd26fed48] Processing image 50ff584c-3b50-4395-af07-3e66769bc9f7 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 848.716608] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f63339cf-eb8a-4d19-b0c5-1ac7744e7482 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 848.716771] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f63339cf-eb8a-4d19-b0c5-1ac7744e7482 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Acquired lock "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 848.717043] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-f63339cf-eb8a-4d19-b0c5-1ac7744e7482 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 848.717340] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-129271ef-a5dd-4aa9-8738-fd8642ce0f0c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.741013] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-f63339cf-eb8a-4d19-b0c5-1ac7744e7482 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 848.741013] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-f63339cf-eb8a-4d19-b0c5-1ac7744e7482 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 848.741013] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a1d67751-fd7f-4eb7-9c92-43ca4e07270f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.750158] env[62066]: DEBUG oslo_vmware.api [None req-f63339cf-eb8a-4d19-b0c5-1ac7744e7482 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Waiting for the task: (returnval){ [ 848.750158] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]522cfbff-c3bc-1b42-9609-5ddb9ac3e6b1" [ 848.750158] env[62066]: _type = "Task" [ 848.750158] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.759347] env[62066]: DEBUG oslo_vmware.api [None req-f63339cf-eb8a-4d19-b0c5-1ac7744e7482 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]522cfbff-c3bc-1b42-9609-5ddb9ac3e6b1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.949173] env[62066]: DEBUG nova.compute.manager [None req-22d0b1dd-f08e-49e4-a262-9fed18806112 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] [instance: 3120c9a5-0558-4e83-b0fb-994a989398fb] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 849.046262] env[62066]: DEBUG oslo_concurrency.lockutils [None req-59107a59-1a77-47be-8dae-1ec8f16f1fda tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 849.175711] env[62066]: DEBUG oslo_vmware.api [None req-e3db032e-002b-4e25-a569-0fed0f11200b tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': task-1340925, 'name': PowerOnVM_Task, 'duration_secs': 0.61625} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.176440] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-e3db032e-002b-4e25-a569-0fed0f11200b tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: b08f27b1-429e-4d05-9a95-f45fd0cb2904] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 849.177074] env[62066]: INFO nova.compute.manager [None req-e3db032e-002b-4e25-a569-0fed0f11200b tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: b08f27b1-429e-4d05-9a95-f45fd0cb2904] Took 9.98 seconds to spawn the instance on the hypervisor. [ 849.178897] env[62066]: DEBUG nova.compute.manager [None req-e3db032e-002b-4e25-a569-0fed0f11200b tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: b08f27b1-429e-4d05-9a95-f45fd0cb2904] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 849.178897] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89d6f1a7-d91e-4bd9-af35-0cdaed930550 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.190541] env[62066]: WARNING nova.network.neutron [None req-1dd6435e-5a62-4e44-bea1-4ade72096a9d tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: 459f432b-e817-405f-9c3e-dfa5d957788c] 449dfe4e-bb49-43b6-9ba7-b57af74ebfb3 already exists in list: networks containing: ['449dfe4e-bb49-43b6-9ba7-b57af74ebfb3']. ignoring it [ 849.269341] env[62066]: DEBUG oslo_vmware.api [None req-f63339cf-eb8a-4d19-b0c5-1ac7744e7482 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]522cfbff-c3bc-1b42-9609-5ddb9ac3e6b1, 'name': SearchDatastore_Task, 'duration_secs': 0.019578} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.269868] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3f3454d4-77da-4a04-a564-9a13947205d9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.278335] env[62066]: DEBUG oslo_vmware.api [None req-f63339cf-eb8a-4d19-b0c5-1ac7744e7482 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Waiting for the task: (returnval){ [ 849.278335] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]5258e8b0-7f36-32ab-fbca-c17d914ef179" [ 849.278335] env[62066]: _type = "Task" [ 849.278335] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.288663] env[62066]: DEBUG oslo_vmware.api [None req-f63339cf-eb8a-4d19-b0c5-1ac7744e7482 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]5258e8b0-7f36-32ab-fbca-c17d914ef179, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.294321] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a09fec9a-37ae-4073-aa64-59c6e68199c0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.307047] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-827b9c17-efd5-43f7-af55-c2653504e53c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.313245] env[62066]: DEBUG oslo_vmware.rw_handles [None req-838d4b27-f790-4e25-a20b-eab8ecc5b3e3 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52e7a16c-6dba-69f6-aeb6-0f9e2db1f0fa/disk-0.vmdk. {{(pid=62066) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 849.314160] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4be098e8-41a8-4488-a1d3-43f25de29695 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.342130] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e87fcd9-0a04-4b82-aacc-e856a2e4640c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.346303] env[62066]: DEBUG oslo_vmware.rw_handles [None req-838d4b27-f790-4e25-a20b-eab8ecc5b3e3 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52e7a16c-6dba-69f6-aeb6-0f9e2db1f0fa/disk-0.vmdk is in state: ready. {{(pid=62066) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 849.346495] env[62066]: ERROR oslo_vmware.rw_handles [None req-838d4b27-f790-4e25-a20b-eab8ecc5b3e3 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52e7a16c-6dba-69f6-aeb6-0f9e2db1f0fa/disk-0.vmdk due to incomplete transfer. [ 849.347247] env[62066]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-81831121-f207-495b-878c-559ab3359676 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.354438] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a86f7e84-397e-4fec-85db-767ea503c493 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.361295] env[62066]: DEBUG oslo_vmware.rw_handles [None req-838d4b27-f790-4e25-a20b-eab8ecc5b3e3 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52e7a16c-6dba-69f6-aeb6-0f9e2db1f0fa/disk-0.vmdk. {{(pid=62066) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 849.361502] env[62066]: DEBUG nova.virt.vmwareapi.images [None req-838d4b27-f790-4e25-a20b-eab8ecc5b3e3 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] [instance: 1a4b0637-1a56-41ef-b89b-6b56d24ed206] Uploaded image 70769f31-9c4a-4989-973b-7d69d5ea0e47 to the Glance image server {{(pid=62066) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 849.363790] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-838d4b27-f790-4e25-a20b-eab8ecc5b3e3 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] [instance: 1a4b0637-1a56-41ef-b89b-6b56d24ed206] Destroying the VM {{(pid=62066) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 849.364375] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-222536ea-d719-497d-8a6f-5ed44cbade62 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.373831] env[62066]: DEBUG nova.compute.provider_tree [None req-73e4bc46-8266-447d-984c-3ca2b99ffd23 tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 849.378747] env[62066]: DEBUG oslo_vmware.api [None req-838d4b27-f790-4e25-a20b-eab8ecc5b3e3 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Waiting for the task: (returnval){ [ 849.378747] env[62066]: value = "task-1340926" [ 849.378747] env[62066]: _type = "Task" [ 849.378747] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.390100] env[62066]: DEBUG oslo_vmware.api [None req-838d4b27-f790-4e25-a20b-eab8ecc5b3e3 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Task: {'id': task-1340926, 'name': Destroy_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.693937] env[62066]: DEBUG nova.network.neutron [None req-1dd6435e-5a62-4e44-bea1-4ade72096a9d tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: 459f432b-e817-405f-9c3e-dfa5d957788c] Updating instance_info_cache with network_info: [{"id": "cb74e58f-b481-4a0f-861b-96381e2d52dc", "address": "fa:16:3e:87:b9:2f", "network": {"id": "449dfe4e-bb49-43b6-9ba7-b57af74ebfb3", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-282029676-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.159", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0d4e33a99c7741fb8cdd97f4ec5dbbd8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db68bd64-5b56-49af-a075-13dcf85cb2e0", "external-id": "nsx-vlan-transportzone-590", "segmentation_id": 590, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcb74e58f-b4", "ovs_interfaceid": "cb74e58f-b481-4a0f-861b-96381e2d52dc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "56efe6b6-2fe3-4c7b-9d59-f8ecf74ac772", "address": "fa:16:3e:c3:7b:ac", "network": {"id": "449dfe4e-bb49-43b6-9ba7-b57af74ebfb3", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-282029676-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0d4e33a99c7741fb8cdd97f4ec5dbbd8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db68bd64-5b56-49af-a075-13dcf85cb2e0", "external-id": "nsx-vlan-transportzone-590", "segmentation_id": 590, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap56efe6b6-2f", "ovs_interfaceid": "56efe6b6-2fe3-4c7b-9d59-f8ecf74ac772", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 849.704707] env[62066]: INFO nova.compute.manager [None req-e3db032e-002b-4e25-a569-0fed0f11200b tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: b08f27b1-429e-4d05-9a95-f45fd0cb2904] Took 16.30 seconds to build instance. [ 849.789197] env[62066]: DEBUG oslo_vmware.api [None req-f63339cf-eb8a-4d19-b0c5-1ac7744e7482 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]5258e8b0-7f36-32ab-fbca-c17d914ef179, 'name': SearchDatastore_Task, 'duration_secs': 0.031209} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.789483] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f63339cf-eb8a-4d19-b0c5-1ac7744e7482 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Releasing lock "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 849.789787] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-f63339cf-eb8a-4d19-b0c5-1ac7744e7482 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk to [datastore1] d77e3405-d158-4a8c-9f54-d5fbd26fed48/d77e3405-d158-4a8c-9f54-d5fbd26fed48.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 849.790390] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-40dfcd21-c4b1-4a35-b65c-72856090d9cf {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.796831] env[62066]: DEBUG oslo_vmware.api [None req-f63339cf-eb8a-4d19-b0c5-1ac7744e7482 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Waiting for the task: (returnval){ [ 849.796831] env[62066]: value = "task-1340927" [ 849.796831] env[62066]: _type = "Task" [ 849.796831] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.805468] env[62066]: DEBUG oslo_vmware.api [None req-f63339cf-eb8a-4d19-b0c5-1ac7744e7482 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Task: {'id': task-1340927, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.880606] env[62066]: DEBUG nova.scheduler.client.report [None req-73e4bc46-8266-447d-984c-3ca2b99ffd23 tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 849.895912] env[62066]: DEBUG oslo_vmware.api [None req-838d4b27-f790-4e25-a20b-eab8ecc5b3e3 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Task: {'id': task-1340926, 'name': Destroy_Task, 'duration_secs': 0.354343} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.896406] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-838d4b27-f790-4e25-a20b-eab8ecc5b3e3 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] [instance: 1a4b0637-1a56-41ef-b89b-6b56d24ed206] Destroyed the VM [ 849.896663] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-838d4b27-f790-4e25-a20b-eab8ecc5b3e3 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] [instance: 1a4b0637-1a56-41ef-b89b-6b56d24ed206] Deleting Snapshot of the VM instance {{(pid=62066) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 849.896931] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-61daefc0-5828-4cbe-af00-793fe567a7ab {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.904973] env[62066]: DEBUG oslo_vmware.api [None req-838d4b27-f790-4e25-a20b-eab8ecc5b3e3 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Waiting for the task: (returnval){ [ 849.904973] env[62066]: value = "task-1340928" [ 849.904973] env[62066]: _type = "Task" [ 849.904973] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.914152] env[62066]: DEBUG oslo_vmware.api [None req-838d4b27-f790-4e25-a20b-eab8ecc5b3e3 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Task: {'id': task-1340928, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.948674] env[62066]: DEBUG nova.compute.manager [req-2bb5ad3e-480e-4f5b-9970-62a072b24b02 req-e006da65-1a7a-4555-a68d-3fc47f70187f service nova] [instance: 459f432b-e817-405f-9c3e-dfa5d957788c] Received event network-vif-plugged-56efe6b6-2fe3-4c7b-9d59-f8ecf74ac772 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 849.948674] env[62066]: DEBUG oslo_concurrency.lockutils [req-2bb5ad3e-480e-4f5b-9970-62a072b24b02 req-e006da65-1a7a-4555-a68d-3fc47f70187f service nova] Acquiring lock "459f432b-e817-405f-9c3e-dfa5d957788c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 849.949114] env[62066]: DEBUG oslo_concurrency.lockutils [req-2bb5ad3e-480e-4f5b-9970-62a072b24b02 req-e006da65-1a7a-4555-a68d-3fc47f70187f service nova] Lock "459f432b-e817-405f-9c3e-dfa5d957788c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 849.949114] env[62066]: DEBUG oslo_concurrency.lockutils [req-2bb5ad3e-480e-4f5b-9970-62a072b24b02 req-e006da65-1a7a-4555-a68d-3fc47f70187f service nova] Lock "459f432b-e817-405f-9c3e-dfa5d957788c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 849.949362] env[62066]: DEBUG nova.compute.manager [req-2bb5ad3e-480e-4f5b-9970-62a072b24b02 req-e006da65-1a7a-4555-a68d-3fc47f70187f service nova] [instance: 459f432b-e817-405f-9c3e-dfa5d957788c] No waiting events found dispatching network-vif-plugged-56efe6b6-2fe3-4c7b-9d59-f8ecf74ac772 {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 849.949592] env[62066]: WARNING nova.compute.manager [req-2bb5ad3e-480e-4f5b-9970-62a072b24b02 req-e006da65-1a7a-4555-a68d-3fc47f70187f service nova] [instance: 459f432b-e817-405f-9c3e-dfa5d957788c] Received unexpected event network-vif-plugged-56efe6b6-2fe3-4c7b-9d59-f8ecf74ac772 for instance with vm_state active and task_state None. [ 849.949801] env[62066]: DEBUG nova.compute.manager [req-2bb5ad3e-480e-4f5b-9970-62a072b24b02 req-e006da65-1a7a-4555-a68d-3fc47f70187f service nova] [instance: 459f432b-e817-405f-9c3e-dfa5d957788c] Received event network-changed-56efe6b6-2fe3-4c7b-9d59-f8ecf74ac772 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 849.949997] env[62066]: DEBUG nova.compute.manager [req-2bb5ad3e-480e-4f5b-9970-62a072b24b02 req-e006da65-1a7a-4555-a68d-3fc47f70187f service nova] [instance: 459f432b-e817-405f-9c3e-dfa5d957788c] Refreshing instance network info cache due to event network-changed-56efe6b6-2fe3-4c7b-9d59-f8ecf74ac772. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 849.950248] env[62066]: DEBUG oslo_concurrency.lockutils [req-2bb5ad3e-480e-4f5b-9970-62a072b24b02 req-e006da65-1a7a-4555-a68d-3fc47f70187f service nova] Acquiring lock "refresh_cache-459f432b-e817-405f-9c3e-dfa5d957788c" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 849.963358] env[62066]: DEBUG nova.compute.manager [None req-22d0b1dd-f08e-49e4-a262-9fed18806112 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] [instance: 3120c9a5-0558-4e83-b0fb-994a989398fb] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 849.991748] env[62066]: DEBUG nova.virt.hardware [None req-22d0b1dd-f08e-49e4-a262-9fed18806112 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-23T13:40:41Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-23T13:40:23Z,direct_url=,disk_format='vmdk',id=50ff584c-3b50-4395-af07-3e66769bc9f7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='eb52f7069a374c61ae946f052007c6d9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-23T13:40:24Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 849.992178] env[62066]: DEBUG nova.virt.hardware [None req-22d0b1dd-f08e-49e4-a262-9fed18806112 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 849.992481] env[62066]: DEBUG nova.virt.hardware [None req-22d0b1dd-f08e-49e4-a262-9fed18806112 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 849.992814] env[62066]: DEBUG nova.virt.hardware [None req-22d0b1dd-f08e-49e4-a262-9fed18806112 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 849.993101] env[62066]: DEBUG nova.virt.hardware [None req-22d0b1dd-f08e-49e4-a262-9fed18806112 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 849.993385] env[62066]: DEBUG nova.virt.hardware [None req-22d0b1dd-f08e-49e4-a262-9fed18806112 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 849.993748] env[62066]: DEBUG nova.virt.hardware [None req-22d0b1dd-f08e-49e4-a262-9fed18806112 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 849.994068] env[62066]: DEBUG nova.virt.hardware [None req-22d0b1dd-f08e-49e4-a262-9fed18806112 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 849.994388] env[62066]: DEBUG nova.virt.hardware [None req-22d0b1dd-f08e-49e4-a262-9fed18806112 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 849.994691] env[62066]: DEBUG nova.virt.hardware [None req-22d0b1dd-f08e-49e4-a262-9fed18806112 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 849.995057] env[62066]: DEBUG nova.virt.hardware [None req-22d0b1dd-f08e-49e4-a262-9fed18806112 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 849.996424] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-762432d7-bebb-4b00-bbf5-a1ce04e66e0c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.008159] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ee56714-b1e5-4dc7-8108-b81230ebe036 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.030079] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-22d0b1dd-f08e-49e4-a262-9fed18806112 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] [instance: 3120c9a5-0558-4e83-b0fb-994a989398fb] Instance VIF info [] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 850.037998] env[62066]: DEBUG oslo.service.loopingcall [None req-22d0b1dd-f08e-49e4-a262-9fed18806112 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 850.038598] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3120c9a5-0558-4e83-b0fb-994a989398fb] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 850.038838] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-738b1956-7243-4593-8589-7e49d083a1f1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.055959] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 850.055959] env[62066]: value = "task-1340929" [ 850.055959] env[62066]: _type = "Task" [ 850.055959] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.063924] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1340929, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.198027] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1dd6435e-5a62-4e44-bea1-4ade72096a9d tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Releasing lock "refresh_cache-459f432b-e817-405f-9c3e-dfa5d957788c" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 850.198726] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1dd6435e-5a62-4e44-bea1-4ade72096a9d tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Acquiring lock "459f432b-e817-405f-9c3e-dfa5d957788c" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 850.199264] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1dd6435e-5a62-4e44-bea1-4ade72096a9d tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Acquired lock "459f432b-e817-405f-9c3e-dfa5d957788c" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 850.199264] env[62066]: DEBUG oslo_concurrency.lockutils [req-2bb5ad3e-480e-4f5b-9970-62a072b24b02 req-e006da65-1a7a-4555-a68d-3fc47f70187f service nova] Acquired lock "refresh_cache-459f432b-e817-405f-9c3e-dfa5d957788c" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 850.199495] env[62066]: DEBUG nova.network.neutron [req-2bb5ad3e-480e-4f5b-9970-62a072b24b02 req-e006da65-1a7a-4555-a68d-3fc47f70187f service nova] [instance: 459f432b-e817-405f-9c3e-dfa5d957788c] Refreshing network info cache for port 56efe6b6-2fe3-4c7b-9d59-f8ecf74ac772 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 850.201363] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf5db6a7-ce22-4336-8e6f-f4c107f4f1cc {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.205982] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e3db032e-002b-4e25-a569-0fed0f11200b tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Lock "b08f27b1-429e-4d05-9a95-f45fd0cb2904" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.814s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 850.223431] env[62066]: DEBUG nova.virt.hardware [None req-1dd6435e-5a62-4e44-bea1-4ade72096a9d tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-23T13:40:41Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 850.223665] env[62066]: DEBUG nova.virt.hardware [None req-1dd6435e-5a62-4e44-bea1-4ade72096a9d tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 850.223830] env[62066]: DEBUG nova.virt.hardware [None req-1dd6435e-5a62-4e44-bea1-4ade72096a9d tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 850.224043] env[62066]: DEBUG nova.virt.hardware [None req-1dd6435e-5a62-4e44-bea1-4ade72096a9d tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 850.224202] env[62066]: DEBUG nova.virt.hardware [None req-1dd6435e-5a62-4e44-bea1-4ade72096a9d tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 850.224353] env[62066]: DEBUG nova.virt.hardware [None req-1dd6435e-5a62-4e44-bea1-4ade72096a9d tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 850.224563] env[62066]: DEBUG nova.virt.hardware [None req-1dd6435e-5a62-4e44-bea1-4ade72096a9d tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 850.224726] env[62066]: DEBUG nova.virt.hardware [None req-1dd6435e-5a62-4e44-bea1-4ade72096a9d tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 850.224897] env[62066]: DEBUG nova.virt.hardware [None req-1dd6435e-5a62-4e44-bea1-4ade72096a9d tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 850.225074] env[62066]: DEBUG nova.virt.hardware [None req-1dd6435e-5a62-4e44-bea1-4ade72096a9d tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 850.225254] env[62066]: DEBUG nova.virt.hardware [None req-1dd6435e-5a62-4e44-bea1-4ade72096a9d tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 850.231607] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-1dd6435e-5a62-4e44-bea1-4ade72096a9d tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: 459f432b-e817-405f-9c3e-dfa5d957788c] Reconfiguring VM to attach interface {{(pid=62066) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 850.232582] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7dc4c51e-d372-442c-b05c-a1b0ce48c8a8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.249908] env[62066]: DEBUG oslo_vmware.api [None req-1dd6435e-5a62-4e44-bea1-4ade72096a9d tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Waiting for the task: (returnval){ [ 850.249908] env[62066]: value = "task-1340930" [ 850.249908] env[62066]: _type = "Task" [ 850.249908] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.262021] env[62066]: DEBUG oslo_vmware.api [None req-1dd6435e-5a62-4e44-bea1-4ade72096a9d tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': task-1340930, 'name': ReconfigVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.308569] env[62066]: DEBUG oslo_vmware.api [None req-f63339cf-eb8a-4d19-b0c5-1ac7744e7482 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Task: {'id': task-1340927, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.391193] env[62066]: DEBUG oslo_concurrency.lockutils [None req-73e4bc46-8266-447d-984c-3ca2b99ffd23 tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.452s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 850.391831] env[62066]: DEBUG nova.compute.manager [None req-73e4bc46-8266-447d-984c-3ca2b99ffd23 tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] [instance: 0bc31a8b-a5fc-4f6e-a4c0-024bf4260755] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 850.394740] env[62066]: DEBUG oslo_concurrency.lockutils [None req-28a644a7-dc09-4460-bd7c-0b89c8effe40 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 7.857s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 850.394982] env[62066]: DEBUG nova.objects.instance [None req-28a644a7-dc09-4460-bd7c-0b89c8effe40 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 88daaaa6-f385-4161-bc74-9fffabb1145d] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62066) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 850.420092] env[62066]: DEBUG oslo_vmware.api [None req-838d4b27-f790-4e25-a20b-eab8ecc5b3e3 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Task: {'id': task-1340928, 'name': RemoveSnapshot_Task} progress is 16%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.571547] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1340929, 'name': CreateVM_Task} progress is 25%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.762334] env[62066]: DEBUG oslo_vmware.api [None req-1dd6435e-5a62-4e44-bea1-4ade72096a9d tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': task-1340930, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.811842] env[62066]: DEBUG oslo_vmware.api [None req-f63339cf-eb8a-4d19-b0c5-1ac7744e7482 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Task: {'id': task-1340927, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.840841} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.812255] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-f63339cf-eb8a-4d19-b0c5-1ac7744e7482 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk to [datastore1] d77e3405-d158-4a8c-9f54-d5fbd26fed48/d77e3405-d158-4a8c-9f54-d5fbd26fed48.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 850.812704] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-f63339cf-eb8a-4d19-b0c5-1ac7744e7482 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] [instance: d77e3405-d158-4a8c-9f54-d5fbd26fed48] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 850.812984] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b4b5f283-5a4e-46bf-a2d0-070bfd3723bb {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.820146] env[62066]: DEBUG oslo_vmware.api [None req-f63339cf-eb8a-4d19-b0c5-1ac7744e7482 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Waiting for the task: (returnval){ [ 850.820146] env[62066]: value = "task-1340931" [ 850.820146] env[62066]: _type = "Task" [ 850.820146] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.829577] env[62066]: DEBUG oslo_vmware.api [None req-f63339cf-eb8a-4d19-b0c5-1ac7744e7482 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Task: {'id': task-1340931, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.899355] env[62066]: DEBUG nova.compute.utils [None req-73e4bc46-8266-447d-984c-3ca2b99ffd23 tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 850.900883] env[62066]: DEBUG nova.compute.manager [None req-73e4bc46-8266-447d-984c-3ca2b99ffd23 tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] [instance: 0bc31a8b-a5fc-4f6e-a4c0-024bf4260755] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 850.901107] env[62066]: DEBUG nova.network.neutron [None req-73e4bc46-8266-447d-984c-3ca2b99ffd23 tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] [instance: 0bc31a8b-a5fc-4f6e-a4c0-024bf4260755] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 850.917615] env[62066]: DEBUG oslo_vmware.api [None req-838d4b27-f790-4e25-a20b-eab8ecc5b3e3 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Task: {'id': task-1340928, 'name': RemoveSnapshot_Task, 'duration_secs': 0.712201} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.918878] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-838d4b27-f790-4e25-a20b-eab8ecc5b3e3 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] [instance: 1a4b0637-1a56-41ef-b89b-6b56d24ed206] Deleted Snapshot of the VM instance {{(pid=62066) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 850.919457] env[62066]: INFO nova.compute.manager [None req-838d4b27-f790-4e25-a20b-eab8ecc5b3e3 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] [instance: 1a4b0637-1a56-41ef-b89b-6b56d24ed206] Took 14.13 seconds to snapshot the instance on the hypervisor. [ 850.991558] env[62066]: DEBUG nova.policy [None req-73e4bc46-8266-447d-984c-3ca2b99ffd23 tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'cc778cb332b145579be618e622de898d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5281a37e471347b79e5dd6e70e03aa0f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 851.067457] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1340929, 'name': CreateVM_Task} progress is 99%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.102271] env[62066]: DEBUG nova.network.neutron [req-2bb5ad3e-480e-4f5b-9970-62a072b24b02 req-e006da65-1a7a-4555-a68d-3fc47f70187f service nova] [instance: 459f432b-e817-405f-9c3e-dfa5d957788c] Updated VIF entry in instance network info cache for port 56efe6b6-2fe3-4c7b-9d59-f8ecf74ac772. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 851.102719] env[62066]: DEBUG nova.network.neutron [req-2bb5ad3e-480e-4f5b-9970-62a072b24b02 req-e006da65-1a7a-4555-a68d-3fc47f70187f service nova] [instance: 459f432b-e817-405f-9c3e-dfa5d957788c] Updating instance_info_cache with network_info: [{"id": "cb74e58f-b481-4a0f-861b-96381e2d52dc", "address": "fa:16:3e:87:b9:2f", "network": {"id": "449dfe4e-bb49-43b6-9ba7-b57af74ebfb3", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-282029676-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.159", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0d4e33a99c7741fb8cdd97f4ec5dbbd8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db68bd64-5b56-49af-a075-13dcf85cb2e0", "external-id": "nsx-vlan-transportzone-590", "segmentation_id": 590, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcb74e58f-b4", "ovs_interfaceid": "cb74e58f-b481-4a0f-861b-96381e2d52dc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "56efe6b6-2fe3-4c7b-9d59-f8ecf74ac772", "address": "fa:16:3e:c3:7b:ac", "network": {"id": "449dfe4e-bb49-43b6-9ba7-b57af74ebfb3", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-282029676-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0d4e33a99c7741fb8cdd97f4ec5dbbd8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db68bd64-5b56-49af-a075-13dcf85cb2e0", "external-id": "nsx-vlan-transportzone-590", "segmentation_id": 590, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap56efe6b6-2f", "ovs_interfaceid": "56efe6b6-2fe3-4c7b-9d59-f8ecf74ac772", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 851.267080] env[62066]: DEBUG oslo_vmware.api [None req-1dd6435e-5a62-4e44-bea1-4ade72096a9d tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': task-1340930, 'name': ReconfigVM_Task, 'duration_secs': 0.887366} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.267784] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1dd6435e-5a62-4e44-bea1-4ade72096a9d tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Releasing lock "459f432b-e817-405f-9c3e-dfa5d957788c" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 851.268103] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-1dd6435e-5a62-4e44-bea1-4ade72096a9d tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: 459f432b-e817-405f-9c3e-dfa5d957788c] Reconfigured VM to attach interface {{(pid=62066) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 851.304156] env[62066]: DEBUG nova.network.neutron [None req-73e4bc46-8266-447d-984c-3ca2b99ffd23 tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] [instance: 0bc31a8b-a5fc-4f6e-a4c0-024bf4260755] Successfully created port: 8f34a054-f99b-447b-89d6-ab1ef9bc3c76 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 851.329794] env[62066]: DEBUG oslo_vmware.api [None req-f63339cf-eb8a-4d19-b0c5-1ac7744e7482 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Task: {'id': task-1340931, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066802} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.330095] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-f63339cf-eb8a-4d19-b0c5-1ac7744e7482 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] [instance: d77e3405-d158-4a8c-9f54-d5fbd26fed48] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 851.330979] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ba54d24-b830-41c9-81a4-7bea54f4b483 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.351542] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-f63339cf-eb8a-4d19-b0c5-1ac7744e7482 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] [instance: d77e3405-d158-4a8c-9f54-d5fbd26fed48] Reconfiguring VM instance instance-0000004c to attach disk [datastore1] d77e3405-d158-4a8c-9f54-d5fbd26fed48/d77e3405-d158-4a8c-9f54-d5fbd26fed48.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 851.351887] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0174297d-ba20-4e2d-8699-1bf58c3fc968 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.372804] env[62066]: DEBUG oslo_vmware.api [None req-f63339cf-eb8a-4d19-b0c5-1ac7744e7482 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Waiting for the task: (returnval){ [ 851.372804] env[62066]: value = "task-1340932" [ 851.372804] env[62066]: _type = "Task" [ 851.372804] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.381281] env[62066]: DEBUG oslo_vmware.api [None req-f63339cf-eb8a-4d19-b0c5-1ac7744e7482 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Task: {'id': task-1340932, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.407157] env[62066]: DEBUG nova.compute.manager [None req-73e4bc46-8266-447d-984c-3ca2b99ffd23 tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] [instance: 0bc31a8b-a5fc-4f6e-a4c0-024bf4260755] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 851.410907] env[62066]: DEBUG oslo_concurrency.lockutils [None req-28a644a7-dc09-4460-bd7c-0b89c8effe40 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.016s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 851.412618] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e0d7e39c-c972-4829-872c-c4830b26c06b tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 8.819s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 851.412896] env[62066]: DEBUG nova.objects.instance [None req-e0d7e39c-c972-4829-872c-c4830b26c06b tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Lazy-loading 'resources' on Instance uuid 86a92b61-f3e9-48f9-8ee2-756669d558ef {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 851.568402] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1340929, 'name': CreateVM_Task, 'duration_secs': 1.07682} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.568769] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3120c9a5-0558-4e83-b0fb-994a989398fb] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 851.569343] env[62066]: DEBUG oslo_concurrency.lockutils [None req-22d0b1dd-f08e-49e4-a262-9fed18806112 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 851.569501] env[62066]: DEBUG oslo_concurrency.lockutils [None req-22d0b1dd-f08e-49e4-a262-9fed18806112 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Acquired lock "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 851.570514] env[62066]: DEBUG oslo_concurrency.lockutils [None req-22d0b1dd-f08e-49e4-a262-9fed18806112 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 851.570514] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f5b86aac-01f8-4fc2-8bc4-03716d563b20 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.577117] env[62066]: DEBUG oslo_vmware.api [None req-22d0b1dd-f08e-49e4-a262-9fed18806112 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Waiting for the task: (returnval){ [ 851.577117] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]52e7f5cb-65c5-5fe8-2f66-d0f963189a14" [ 851.577117] env[62066]: _type = "Task" [ 851.577117] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.583955] env[62066]: DEBUG oslo_vmware.api [None req-22d0b1dd-f08e-49e4-a262-9fed18806112 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52e7f5cb-65c5-5fe8-2f66-d0f963189a14, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.605356] env[62066]: DEBUG oslo_concurrency.lockutils [req-2bb5ad3e-480e-4f5b-9970-62a072b24b02 req-e006da65-1a7a-4555-a68d-3fc47f70187f service nova] Releasing lock "refresh_cache-459f432b-e817-405f-9c3e-dfa5d957788c" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 851.660838] env[62066]: DEBUG nova.compute.manager [None req-5b85ae79-b544-4bc9-a850-092bba375a6b tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: b08f27b1-429e-4d05-9a95-f45fd0cb2904] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 851.662030] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bb45a6c-fc00-4fa6-8aba-1dd9274df624 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.775306] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1dd6435e-5a62-4e44-bea1-4ade72096a9d tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Lock "interface-459f432b-e817-405f-9c3e-dfa5d957788c-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 7.390s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 851.885189] env[62066]: DEBUG oslo_vmware.api [None req-f63339cf-eb8a-4d19-b0c5-1ac7744e7482 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Task: {'id': task-1340932, 'name': ReconfigVM_Task, 'duration_secs': 0.268278} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.885600] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-f63339cf-eb8a-4d19-b0c5-1ac7744e7482 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] [instance: d77e3405-d158-4a8c-9f54-d5fbd26fed48] Reconfigured VM instance instance-0000004c to attach disk [datastore1] d77e3405-d158-4a8c-9f54-d5fbd26fed48/d77e3405-d158-4a8c-9f54-d5fbd26fed48.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 851.886659] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-961c9560-6ab6-4e82-9d41-d93d5607cd94 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.893245] env[62066]: DEBUG oslo_vmware.api [None req-f63339cf-eb8a-4d19-b0c5-1ac7744e7482 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Waiting for the task: (returnval){ [ 851.893245] env[62066]: value = "task-1340933" [ 851.893245] env[62066]: _type = "Task" [ 851.893245] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.901354] env[62066]: DEBUG oslo_vmware.api [None req-f63339cf-eb8a-4d19-b0c5-1ac7744e7482 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Task: {'id': task-1340933, 'name': Rename_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.085835] env[62066]: DEBUG oslo_vmware.api [None req-22d0b1dd-f08e-49e4-a262-9fed18806112 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52e7f5cb-65c5-5fe8-2f66-d0f963189a14, 'name': SearchDatastore_Task, 'duration_secs': 0.011683} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.088582] env[62066]: DEBUG oslo_concurrency.lockutils [None req-22d0b1dd-f08e-49e4-a262-9fed18806112 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Releasing lock "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 852.088672] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-22d0b1dd-f08e-49e4-a262-9fed18806112 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] [instance: 3120c9a5-0558-4e83-b0fb-994a989398fb] Processing image 50ff584c-3b50-4395-af07-3e66769bc9f7 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 852.089239] env[62066]: DEBUG oslo_concurrency.lockutils [None req-22d0b1dd-f08e-49e4-a262-9fed18806112 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 852.089239] env[62066]: DEBUG oslo_concurrency.lockutils [None req-22d0b1dd-f08e-49e4-a262-9fed18806112 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Acquired lock "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 852.089239] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-22d0b1dd-f08e-49e4-a262-9fed18806112 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 852.089728] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-662a4d2a-7e4a-4fc8-81ff-9c126d4aa18b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.098962] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-22d0b1dd-f08e-49e4-a262-9fed18806112 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 852.099759] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-22d0b1dd-f08e-49e4-a262-9fed18806112 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 852.099980] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-59ce6cc4-0928-48a2-972b-aa64b0a73222 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.109052] env[62066]: DEBUG oslo_vmware.api [None req-22d0b1dd-f08e-49e4-a262-9fed18806112 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Waiting for the task: (returnval){ [ 852.109052] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]5212035c-2872-1e38-dfad-6ad8c66edd02" [ 852.109052] env[62066]: _type = "Task" [ 852.109052] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.117812] env[62066]: DEBUG oslo_vmware.api [None req-22d0b1dd-f08e-49e4-a262-9fed18806112 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]5212035c-2872-1e38-dfad-6ad8c66edd02, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.172464] env[62066]: INFO nova.compute.manager [None req-5b85ae79-b544-4bc9-a850-092bba375a6b tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: b08f27b1-429e-4d05-9a95-f45fd0cb2904] instance snapshotting [ 852.175427] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fab5c3bd-ab5a-463f-bafc-54498150b3b2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.196715] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-757a67e4-6989-4bd2-b69d-349758d5c553 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.210977] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d11f9c2-4b04-4604-a77f-168d6333122f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.218986] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aed7b00d-dc4b-448d-8d65-a2b1b7546fbd {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.256292] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ca39414-cf35-49bb-8176-093bed62a9e2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.267531] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6178cbea-1ec4-4300-ab5e-4399c616e283 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.282900] env[62066]: DEBUG nova.compute.provider_tree [None req-e0d7e39c-c972-4829-872c-c4830b26c06b tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 852.403548] env[62066]: DEBUG oslo_vmware.api [None req-f63339cf-eb8a-4d19-b0c5-1ac7744e7482 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Task: {'id': task-1340933, 'name': Rename_Task, 'duration_secs': 0.300193} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.403834] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-f63339cf-eb8a-4d19-b0c5-1ac7744e7482 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] [instance: d77e3405-d158-4a8c-9f54-d5fbd26fed48] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 852.404081] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8d9a0d66-7322-46b0-b6f4-c569f8a783a5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.410425] env[62066]: DEBUG oslo_vmware.api [None req-f63339cf-eb8a-4d19-b0c5-1ac7744e7482 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Waiting for the task: (returnval){ [ 852.410425] env[62066]: value = "task-1340934" [ 852.410425] env[62066]: _type = "Task" [ 852.410425] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.418849] env[62066]: DEBUG oslo_vmware.api [None req-f63339cf-eb8a-4d19-b0c5-1ac7744e7482 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Task: {'id': task-1340934, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.423794] env[62066]: DEBUG nova.compute.manager [None req-73e4bc46-8266-447d-984c-3ca2b99ffd23 tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] [instance: 0bc31a8b-a5fc-4f6e-a4c0-024bf4260755] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 852.448951] env[62066]: DEBUG nova.virt.hardware [None req-73e4bc46-8266-447d-984c-3ca2b99ffd23 tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-23T13:40:41Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-23T13:40:23Z,direct_url=,disk_format='vmdk',id=50ff584c-3b50-4395-af07-3e66769bc9f7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='eb52f7069a374c61ae946f052007c6d9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-23T13:40:24Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 852.449271] env[62066]: DEBUG nova.virt.hardware [None req-73e4bc46-8266-447d-984c-3ca2b99ffd23 tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 852.449476] env[62066]: DEBUG nova.virt.hardware [None req-73e4bc46-8266-447d-984c-3ca2b99ffd23 tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 852.449786] env[62066]: DEBUG nova.virt.hardware [None req-73e4bc46-8266-447d-984c-3ca2b99ffd23 tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 852.449980] env[62066]: DEBUG nova.virt.hardware [None req-73e4bc46-8266-447d-984c-3ca2b99ffd23 tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 852.450191] env[62066]: DEBUG nova.virt.hardware [None req-73e4bc46-8266-447d-984c-3ca2b99ffd23 tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 852.450446] env[62066]: DEBUG nova.virt.hardware [None req-73e4bc46-8266-447d-984c-3ca2b99ffd23 tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 852.450751] env[62066]: DEBUG nova.virt.hardware [None req-73e4bc46-8266-447d-984c-3ca2b99ffd23 tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 852.450973] env[62066]: DEBUG nova.virt.hardware [None req-73e4bc46-8266-447d-984c-3ca2b99ffd23 tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 852.451197] env[62066]: DEBUG nova.virt.hardware [None req-73e4bc46-8266-447d-984c-3ca2b99ffd23 tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 852.451415] env[62066]: DEBUG nova.virt.hardware [None req-73e4bc46-8266-447d-984c-3ca2b99ffd23 tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 852.452347] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11ecaf2e-2e7a-462d-9a6f-0df4516a7013 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.460994] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61da303c-7a42-4947-b0cd-e4b2dbbe68b0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.621980] env[62066]: DEBUG oslo_vmware.api [None req-22d0b1dd-f08e-49e4-a262-9fed18806112 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]5212035c-2872-1e38-dfad-6ad8c66edd02, 'name': SearchDatastore_Task, 'duration_secs': 0.054794} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.622983] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-94158382-3b32-492e-8af9-f1d16645360c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.634021] env[62066]: DEBUG oslo_vmware.api [None req-22d0b1dd-f08e-49e4-a262-9fed18806112 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Waiting for the task: (returnval){ [ 852.634021] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]5236b3fd-b4ab-cc23-665f-b15d0833657e" [ 852.634021] env[62066]: _type = "Task" [ 852.634021] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.640482] env[62066]: DEBUG oslo_vmware.api [None req-22d0b1dd-f08e-49e4-a262-9fed18806112 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]5236b3fd-b4ab-cc23-665f-b15d0833657e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.710244] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-5b85ae79-b544-4bc9-a850-092bba375a6b tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: b08f27b1-429e-4d05-9a95-f45fd0cb2904] Creating Snapshot of the VM instance {{(pid=62066) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 852.710244] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-eb973a9a-bf84-4cc1-afa5-c03879d29b0b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.717537] env[62066]: DEBUG oslo_vmware.api [None req-5b85ae79-b544-4bc9-a850-092bba375a6b tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Waiting for the task: (returnval){ [ 852.717537] env[62066]: value = "task-1340935" [ 852.717537] env[62066]: _type = "Task" [ 852.717537] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.730038] env[62066]: DEBUG oslo_vmware.api [None req-5b85ae79-b544-4bc9-a850-092bba375a6b tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': task-1340935, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.786352] env[62066]: DEBUG nova.scheduler.client.report [None req-e0d7e39c-c972-4829-872c-c4830b26c06b tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 852.806673] env[62066]: DEBUG nova.compute.manager [req-7616cf97-70bd-4f97-8f29-bf11928938a8 req-4a42ae19-c472-45f4-8aa1-5fb625bc993a service nova] [instance: 0bc31a8b-a5fc-4f6e-a4c0-024bf4260755] Received event network-vif-plugged-8f34a054-f99b-447b-89d6-ab1ef9bc3c76 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 852.806673] env[62066]: DEBUG oslo_concurrency.lockutils [req-7616cf97-70bd-4f97-8f29-bf11928938a8 req-4a42ae19-c472-45f4-8aa1-5fb625bc993a service nova] Acquiring lock "0bc31a8b-a5fc-4f6e-a4c0-024bf4260755-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 852.806673] env[62066]: DEBUG oslo_concurrency.lockutils [req-7616cf97-70bd-4f97-8f29-bf11928938a8 req-4a42ae19-c472-45f4-8aa1-5fb625bc993a service nova] Lock "0bc31a8b-a5fc-4f6e-a4c0-024bf4260755-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 852.806673] env[62066]: DEBUG oslo_concurrency.lockutils [req-7616cf97-70bd-4f97-8f29-bf11928938a8 req-4a42ae19-c472-45f4-8aa1-5fb625bc993a service nova] Lock "0bc31a8b-a5fc-4f6e-a4c0-024bf4260755-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 852.806673] env[62066]: DEBUG nova.compute.manager [req-7616cf97-70bd-4f97-8f29-bf11928938a8 req-4a42ae19-c472-45f4-8aa1-5fb625bc993a service nova] [instance: 0bc31a8b-a5fc-4f6e-a4c0-024bf4260755] No waiting events found dispatching network-vif-plugged-8f34a054-f99b-447b-89d6-ab1ef9bc3c76 {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 852.806673] env[62066]: WARNING nova.compute.manager [req-7616cf97-70bd-4f97-8f29-bf11928938a8 req-4a42ae19-c472-45f4-8aa1-5fb625bc993a service nova] [instance: 0bc31a8b-a5fc-4f6e-a4c0-024bf4260755] Received unexpected event network-vif-plugged-8f34a054-f99b-447b-89d6-ab1ef9bc3c76 for instance with vm_state building and task_state spawning. [ 852.911990] env[62066]: DEBUG nova.network.neutron [None req-73e4bc46-8266-447d-984c-3ca2b99ffd23 tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] [instance: 0bc31a8b-a5fc-4f6e-a4c0-024bf4260755] Successfully updated port: 8f34a054-f99b-447b-89d6-ab1ef9bc3c76 {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 852.930128] env[62066]: DEBUG oslo_vmware.api [None req-f63339cf-eb8a-4d19-b0c5-1ac7744e7482 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Task: {'id': task-1340934, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.145342] env[62066]: DEBUG oslo_vmware.api [None req-22d0b1dd-f08e-49e4-a262-9fed18806112 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]5236b3fd-b4ab-cc23-665f-b15d0833657e, 'name': SearchDatastore_Task, 'duration_secs': 0.030595} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.145342] env[62066]: DEBUG oslo_concurrency.lockutils [None req-22d0b1dd-f08e-49e4-a262-9fed18806112 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Releasing lock "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 853.145583] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-22d0b1dd-f08e-49e4-a262-9fed18806112 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk to [datastore1] 3120c9a5-0558-4e83-b0fb-994a989398fb/3120c9a5-0558-4e83-b0fb-994a989398fb.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 853.145862] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-db30c8df-62f4-4a87-a1b8-abbe9badca0c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.157302] env[62066]: DEBUG oslo_vmware.api [None req-22d0b1dd-f08e-49e4-a262-9fed18806112 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Waiting for the task: (returnval){ [ 853.157302] env[62066]: value = "task-1340936" [ 853.157302] env[62066]: _type = "Task" [ 853.157302] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.169844] env[62066]: DEBUG oslo_vmware.api [None req-22d0b1dd-f08e-49e4-a262-9fed18806112 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Task: {'id': task-1340936, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.228795] env[62066]: DEBUG oslo_vmware.api [None req-5b85ae79-b544-4bc9-a850-092bba375a6b tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': task-1340935, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.295530] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e0d7e39c-c972-4829-872c-c4830b26c06b tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.883s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 853.298031] env[62066]: DEBUG oslo_concurrency.lockutils [None req-63dd5fc5-276d-49c0-99ba-71a261a8efa4 tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 9.921s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 853.323569] env[62066]: INFO nova.scheduler.client.report [None req-e0d7e39c-c972-4829-872c-c4830b26c06b tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Deleted allocations for instance 86a92b61-f3e9-48f9-8ee2-756669d558ef [ 853.425873] env[62066]: DEBUG oslo_concurrency.lockutils [None req-73e4bc46-8266-447d-984c-3ca2b99ffd23 tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] Acquiring lock "refresh_cache-0bc31a8b-a5fc-4f6e-a4c0-024bf4260755" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 853.425873] env[62066]: DEBUG oslo_concurrency.lockutils [None req-73e4bc46-8266-447d-984c-3ca2b99ffd23 tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] Acquired lock "refresh_cache-0bc31a8b-a5fc-4f6e-a4c0-024bf4260755" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 853.425873] env[62066]: DEBUG nova.network.neutron [None req-73e4bc46-8266-447d-984c-3ca2b99ffd23 tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] [instance: 0bc31a8b-a5fc-4f6e-a4c0-024bf4260755] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 853.433175] env[62066]: DEBUG oslo_vmware.api [None req-f63339cf-eb8a-4d19-b0c5-1ac7744e7482 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Task: {'id': task-1340934, 'name': PowerOnVM_Task, 'duration_secs': 0.518295} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.433524] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-f63339cf-eb8a-4d19-b0c5-1ac7744e7482 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] [instance: d77e3405-d158-4a8c-9f54-d5fbd26fed48] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 853.433778] env[62066]: INFO nova.compute.manager [None req-f63339cf-eb8a-4d19-b0c5-1ac7744e7482 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] [instance: d77e3405-d158-4a8c-9f54-d5fbd26fed48] Took 5.90 seconds to spawn the instance on the hypervisor. [ 853.434033] env[62066]: DEBUG nova.compute.manager [None req-f63339cf-eb8a-4d19-b0c5-1ac7744e7482 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] [instance: d77e3405-d158-4a8c-9f54-d5fbd26fed48] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 853.434958] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6cadefe-c41b-4a4f-9fce-58794b72c1be {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.646948] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-610a971a-df16-4b6f-ab50-8865951c3cc6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.662207] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-170b746d-3480-4f0a-ad30-41eb753253bb {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.671234] env[62066]: DEBUG oslo_vmware.api [None req-22d0b1dd-f08e-49e4-a262-9fed18806112 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Task: {'id': task-1340936, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.700106] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed73248b-b838-4427-8f42-52b211c657d9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.709157] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-508e2afc-eb52-41f6-ad93-4e71cff0c45e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.724837] env[62066]: DEBUG nova.compute.provider_tree [None req-63dd5fc5-276d-49c0-99ba-71a261a8efa4 tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 853.734818] env[62066]: DEBUG oslo_vmware.api [None req-5b85ae79-b544-4bc9-a850-092bba375a6b tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': task-1340935, 'name': CreateSnapshot_Task, 'duration_secs': 0.883299} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.735819] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-5b85ae79-b544-4bc9-a850-092bba375a6b tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: b08f27b1-429e-4d05-9a95-f45fd0cb2904] Created Snapshot of the VM instance {{(pid=62066) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 853.736697] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7f4e805-ffd5-4ec3-8159-41b62db53902 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.786251] env[62066]: DEBUG oslo_concurrency.lockutils [None req-bace0778-9dfe-4de1-8747-93c45a08d78a tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Acquiring lock "interface-459f432b-e817-405f-9c3e-dfa5d957788c-56efe6b6-2fe3-4c7b-9d59-f8ecf74ac772" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 853.786525] env[62066]: DEBUG oslo_concurrency.lockutils [None req-bace0778-9dfe-4de1-8747-93c45a08d78a tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Lock "interface-459f432b-e817-405f-9c3e-dfa5d957788c-56efe6b6-2fe3-4c7b-9d59-f8ecf74ac772" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 853.834153] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e0d7e39c-c972-4829-872c-c4830b26c06b tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Lock "86a92b61-f3e9-48f9-8ee2-756669d558ef" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 14.262s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 853.960663] env[62066]: INFO nova.compute.manager [None req-f63339cf-eb8a-4d19-b0c5-1ac7744e7482 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] [instance: d77e3405-d158-4a8c-9f54-d5fbd26fed48] Took 15.48 seconds to build instance. [ 853.968201] env[62066]: DEBUG nova.network.neutron [None req-73e4bc46-8266-447d-984c-3ca2b99ffd23 tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] [instance: 0bc31a8b-a5fc-4f6e-a4c0-024bf4260755] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 854.047795] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5327c80f-b6fd-46f0-ac98-3857740d94e2 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Acquiring lock "1e1a2ede-4ad8-4600-851e-6f2046b3f919" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 854.048096] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5327c80f-b6fd-46f0-ac98-3857740d94e2 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Lock "1e1a2ede-4ad8-4600-851e-6f2046b3f919" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 854.048311] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5327c80f-b6fd-46f0-ac98-3857740d94e2 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Acquiring lock "1e1a2ede-4ad8-4600-851e-6f2046b3f919-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 854.048500] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5327c80f-b6fd-46f0-ac98-3857740d94e2 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Lock "1e1a2ede-4ad8-4600-851e-6f2046b3f919-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 854.048669] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5327c80f-b6fd-46f0-ac98-3857740d94e2 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Lock "1e1a2ede-4ad8-4600-851e-6f2046b3f919-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 854.051105] env[62066]: INFO nova.compute.manager [None req-5327c80f-b6fd-46f0-ac98-3857740d94e2 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] [instance: 1e1a2ede-4ad8-4600-851e-6f2046b3f919] Terminating instance [ 854.054184] env[62066]: DEBUG nova.compute.manager [None req-5327c80f-b6fd-46f0-ac98-3857740d94e2 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] [instance: 1e1a2ede-4ad8-4600-851e-6f2046b3f919] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 854.054184] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-5327c80f-b6fd-46f0-ac98-3857740d94e2 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] [instance: 1e1a2ede-4ad8-4600-851e-6f2046b3f919] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 854.054184] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43b5ac2f-a6bb-43ba-8399-a40f2dd538e1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.062034] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-5327c80f-b6fd-46f0-ac98-3857740d94e2 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] [instance: 1e1a2ede-4ad8-4600-851e-6f2046b3f919] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 854.062305] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fb15f5b4-1872-4c4a-80d2-8bda755ccbac {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.068897] env[62066]: DEBUG oslo_vmware.api [None req-5327c80f-b6fd-46f0-ac98-3857740d94e2 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Waiting for the task: (returnval){ [ 854.068897] env[62066]: value = "task-1340937" [ 854.068897] env[62066]: _type = "Task" [ 854.068897] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.080020] env[62066]: DEBUG oslo_vmware.api [None req-5327c80f-b6fd-46f0-ac98-3857740d94e2 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Task: {'id': task-1340937, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.156200] env[62066]: DEBUG nova.network.neutron [None req-73e4bc46-8266-447d-984c-3ca2b99ffd23 tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] [instance: 0bc31a8b-a5fc-4f6e-a4c0-024bf4260755] Updating instance_info_cache with network_info: [{"id": "8f34a054-f99b-447b-89d6-ab1ef9bc3c76", "address": "fa:16:3e:06:43:d8", "network": {"id": "b92d2e33-767e-4958-b805-12e11b75c266", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-583710540-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "5281a37e471347b79e5dd6e70e03aa0f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "13e71dbb-4279-427c-b39d-ba5df9895e58", "external-id": "nsx-vlan-transportzone-417", "segmentation_id": 417, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8f34a054-f9", "ovs_interfaceid": "8f34a054-f99b-447b-89d6-ab1ef9bc3c76", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 854.168080] env[62066]: DEBUG oslo_vmware.api [None req-22d0b1dd-f08e-49e4-a262-9fed18806112 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Task: {'id': task-1340936, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.642302} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.168911] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-22d0b1dd-f08e-49e4-a262-9fed18806112 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk to [datastore1] 3120c9a5-0558-4e83-b0fb-994a989398fb/3120c9a5-0558-4e83-b0fb-994a989398fb.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 854.169164] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-22d0b1dd-f08e-49e4-a262-9fed18806112 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] [instance: 3120c9a5-0558-4e83-b0fb-994a989398fb] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 854.169425] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-94b689be-d0e1-474c-b9a5-7e9eabb98d97 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.177487] env[62066]: DEBUG oslo_vmware.api [None req-22d0b1dd-f08e-49e4-a262-9fed18806112 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Waiting for the task: (returnval){ [ 854.177487] env[62066]: value = "task-1340938" [ 854.177487] env[62066]: _type = "Task" [ 854.177487] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.191260] env[62066]: DEBUG oslo_vmware.api [None req-22d0b1dd-f08e-49e4-a262-9fed18806112 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Task: {'id': task-1340938, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.233780] env[62066]: DEBUG nova.scheduler.client.report [None req-63dd5fc5-276d-49c0-99ba-71a261a8efa4 tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 854.257454] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-5b85ae79-b544-4bc9-a850-092bba375a6b tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: b08f27b1-429e-4d05-9a95-f45fd0cb2904] Creating linked-clone VM from snapshot {{(pid=62066) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 854.257454] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-3c2a80f4-3d01-4637-baa2-f8de9339e776 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.268624] env[62066]: DEBUG oslo_vmware.api [None req-5b85ae79-b544-4bc9-a850-092bba375a6b tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Waiting for the task: (returnval){ [ 854.268624] env[62066]: value = "task-1340939" [ 854.268624] env[62066]: _type = "Task" [ 854.268624] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.277613] env[62066]: DEBUG oslo_vmware.api [None req-5b85ae79-b544-4bc9-a850-092bba375a6b tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': task-1340939, 'name': CloneVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.289235] env[62066]: DEBUG oslo_concurrency.lockutils [None req-bace0778-9dfe-4de1-8747-93c45a08d78a tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Acquiring lock "459f432b-e817-405f-9c3e-dfa5d957788c" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 854.289433] env[62066]: DEBUG oslo_concurrency.lockutils [None req-bace0778-9dfe-4de1-8747-93c45a08d78a tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Acquired lock "459f432b-e817-405f-9c3e-dfa5d957788c" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 854.290350] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e2e19ed-3055-4ac1-be62-d1dacaf61f3e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.310585] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66d726fd-32bf-4878-a8cf-ad0340cf4c13 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.337155] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-bace0778-9dfe-4de1-8747-93c45a08d78a tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: 459f432b-e817-405f-9c3e-dfa5d957788c] Reconfiguring VM to detach interface {{(pid=62066) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 854.337757] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-14d06a13-69ed-444a-8342-3375c13699b5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.357227] env[62066]: DEBUG oslo_vmware.api [None req-bace0778-9dfe-4de1-8747-93c45a08d78a tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Waiting for the task: (returnval){ [ 854.357227] env[62066]: value = "task-1340940" [ 854.357227] env[62066]: _type = "Task" [ 854.357227] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.363961] env[62066]: DEBUG oslo_vmware.api [None req-bace0778-9dfe-4de1-8747-93c45a08d78a tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': task-1340940, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.462522] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f63339cf-eb8a-4d19-b0c5-1ac7744e7482 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Lock "d77e3405-d158-4a8c-9f54-d5fbd26fed48" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.995s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 854.578745] env[62066]: DEBUG oslo_vmware.api [None req-5327c80f-b6fd-46f0-ac98-3857740d94e2 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Task: {'id': task-1340937, 'name': PowerOffVM_Task, 'duration_secs': 0.448146} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.579074] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-5327c80f-b6fd-46f0-ac98-3857740d94e2 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] [instance: 1e1a2ede-4ad8-4600-851e-6f2046b3f919] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 854.579225] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-5327c80f-b6fd-46f0-ac98-3857740d94e2 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] [instance: 1e1a2ede-4ad8-4600-851e-6f2046b3f919] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 854.579479] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-10e72fbf-3807-4a0e-9fb6-10cc63e23f06 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.660607] env[62066]: DEBUG oslo_concurrency.lockutils [None req-73e4bc46-8266-447d-984c-3ca2b99ffd23 tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] Releasing lock "refresh_cache-0bc31a8b-a5fc-4f6e-a4c0-024bf4260755" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 854.660607] env[62066]: DEBUG nova.compute.manager [None req-73e4bc46-8266-447d-984c-3ca2b99ffd23 tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] [instance: 0bc31a8b-a5fc-4f6e-a4c0-024bf4260755] Instance network_info: |[{"id": "8f34a054-f99b-447b-89d6-ab1ef9bc3c76", "address": "fa:16:3e:06:43:d8", "network": {"id": "b92d2e33-767e-4958-b805-12e11b75c266", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-583710540-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "5281a37e471347b79e5dd6e70e03aa0f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "13e71dbb-4279-427c-b39d-ba5df9895e58", "external-id": "nsx-vlan-transportzone-417", "segmentation_id": 417, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8f34a054-f9", "ovs_interfaceid": "8f34a054-f99b-447b-89d6-ab1ef9bc3c76", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 854.661267] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-73e4bc46-8266-447d-984c-3ca2b99ffd23 tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] [instance: 0bc31a8b-a5fc-4f6e-a4c0-024bf4260755] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:06:43:d8', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '13e71dbb-4279-427c-b39d-ba5df9895e58', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8f34a054-f99b-447b-89d6-ab1ef9bc3c76', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 854.671766] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-73e4bc46-8266-447d-984c-3ca2b99ffd23 tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] Creating folder: Project (5281a37e471347b79e5dd6e70e03aa0f). Parent ref: group-v285980. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 854.671766] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7739ca12-df44-419a-a429-650d7aeafec6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.683202] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-73e4bc46-8266-447d-984c-3ca2b99ffd23 tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] Created folder: Project (5281a37e471347b79e5dd6e70e03aa0f) in parent group-v285980. [ 854.683417] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-73e4bc46-8266-447d-984c-3ca2b99ffd23 tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] Creating folder: Instances. Parent ref: group-v286070. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 854.684044] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ad23a6e4-7fbb-4932-a108-3e38da2f0251 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.688959] env[62066]: DEBUG oslo_vmware.api [None req-22d0b1dd-f08e-49e4-a262-9fed18806112 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Task: {'id': task-1340938, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072038} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.689662] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-22d0b1dd-f08e-49e4-a262-9fed18806112 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] [instance: 3120c9a5-0558-4e83-b0fb-994a989398fb] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 854.690441] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5172fd0d-fa14-426b-96be-a3f6eaa4baae {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.696162] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-73e4bc46-8266-447d-984c-3ca2b99ffd23 tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] Created folder: Instances in parent group-v286070. [ 854.696452] env[62066]: DEBUG oslo.service.loopingcall [None req-73e4bc46-8266-447d-984c-3ca2b99ffd23 tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 854.704419] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0bc31a8b-a5fc-4f6e-a4c0-024bf4260755] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 854.714319] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-22d0b1dd-f08e-49e4-a262-9fed18806112 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] [instance: 3120c9a5-0558-4e83-b0fb-994a989398fb] Reconfiguring VM instance instance-0000004d to attach disk [datastore1] 3120c9a5-0558-4e83-b0fb-994a989398fb/3120c9a5-0558-4e83-b0fb-994a989398fb.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 854.715739] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0ba000fa-13f9-48a5-8439-ae7c04883f71 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.730647] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c4f1ed10-0809-41da-b02c-cda9f6c25368 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.749708] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-5327c80f-b6fd-46f0-ac98-3857740d94e2 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] [instance: 1e1a2ede-4ad8-4600-851e-6f2046b3f919] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 854.749951] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-5327c80f-b6fd-46f0-ac98-3857740d94e2 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] [instance: 1e1a2ede-4ad8-4600-851e-6f2046b3f919] Deleting contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 854.750179] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-5327c80f-b6fd-46f0-ac98-3857740d94e2 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Deleting the datastore file [datastore2] 1e1a2ede-4ad8-4600-851e-6f2046b3f919 {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 854.751988] env[62066]: DEBUG oslo_concurrency.lockutils [None req-63dd5fc5-276d-49c0-99ba-71a261a8efa4 tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.453s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 854.751988] env[62066]: INFO nova.compute.manager [None req-63dd5fc5-276d-49c0-99ba-71a261a8efa4 tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] Successfully reverted task state from rebuilding on failure for instance. [ 854.758029] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-bd1849ea-6236-4bb1-85ba-a21db911c513 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.760321] env[62066]: DEBUG oslo_concurrency.lockutils [None req-68383c37-9698-4391-bbf2-b9ebf656dc03 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.589s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 854.765018] env[62066]: INFO nova.compute.claims [None req-68383c37-9698-4391-bbf2-b9ebf656dc03 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: c8e0d47c-4421-4e00-9183-206fceeabc40] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 854.766798] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 854.766798] env[62066]: value = "task-1340944" [ 854.766798] env[62066]: _type = "Task" [ 854.766798] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.772915] env[62066]: DEBUG oslo_vmware.api [None req-22d0b1dd-f08e-49e4-a262-9fed18806112 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Waiting for the task: (returnval){ [ 854.772915] env[62066]: value = "task-1340945" [ 854.772915] env[62066]: _type = "Task" [ 854.772915] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.773201] env[62066]: DEBUG oslo_vmware.api [None req-5327c80f-b6fd-46f0-ac98-3857740d94e2 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Waiting for the task: (returnval){ [ 854.773201] env[62066]: value = "task-1340946" [ 854.773201] env[62066]: _type = "Task" [ 854.773201] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.786577] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1340944, 'name': CreateVM_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.791241] env[62066]: DEBUG oslo_vmware.api [None req-5b85ae79-b544-4bc9-a850-092bba375a6b tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': task-1340939, 'name': CloneVM_Task} progress is 94%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.801224] env[62066]: DEBUG oslo_vmware.api [None req-5327c80f-b6fd-46f0-ac98-3857740d94e2 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Task: {'id': task-1340946, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.801279] env[62066]: DEBUG oslo_vmware.api [None req-22d0b1dd-f08e-49e4-a262-9fed18806112 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Task: {'id': task-1340945, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.872037] env[62066]: DEBUG oslo_vmware.api [None req-bace0778-9dfe-4de1-8747-93c45a08d78a tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': task-1340940, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.027778] env[62066]: DEBUG nova.compute.manager [req-b2ea8705-2cdd-43a8-bc17-8ef80d2acafa req-450fc733-5021-4208-a8f3-a25d0a66183e service nova] [instance: 0bc31a8b-a5fc-4f6e-a4c0-024bf4260755] Received event network-changed-8f34a054-f99b-447b-89d6-ab1ef9bc3c76 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 855.028015] env[62066]: DEBUG nova.compute.manager [req-b2ea8705-2cdd-43a8-bc17-8ef80d2acafa req-450fc733-5021-4208-a8f3-a25d0a66183e service nova] [instance: 0bc31a8b-a5fc-4f6e-a4c0-024bf4260755] Refreshing instance network info cache due to event network-changed-8f34a054-f99b-447b-89d6-ab1ef9bc3c76. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 855.028248] env[62066]: DEBUG oslo_concurrency.lockutils [req-b2ea8705-2cdd-43a8-bc17-8ef80d2acafa req-450fc733-5021-4208-a8f3-a25d0a66183e service nova] Acquiring lock "refresh_cache-0bc31a8b-a5fc-4f6e-a4c0-024bf4260755" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 855.028397] env[62066]: DEBUG oslo_concurrency.lockutils [req-b2ea8705-2cdd-43a8-bc17-8ef80d2acafa req-450fc733-5021-4208-a8f3-a25d0a66183e service nova] Acquired lock "refresh_cache-0bc31a8b-a5fc-4f6e-a4c0-024bf4260755" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 855.028562] env[62066]: DEBUG nova.network.neutron [req-b2ea8705-2cdd-43a8-bc17-8ef80d2acafa req-450fc733-5021-4208-a8f3-a25d0a66183e service nova] [instance: 0bc31a8b-a5fc-4f6e-a4c0-024bf4260755] Refreshing network info cache for port 8f34a054-f99b-447b-89d6-ab1ef9bc3c76 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 855.279909] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1340944, 'name': CreateVM_Task, 'duration_secs': 0.363307} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.286274] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0bc31a8b-a5fc-4f6e-a4c0-024bf4260755] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 855.291243] env[62066]: DEBUG oslo_concurrency.lockutils [None req-73e4bc46-8266-447d-984c-3ca2b99ffd23 tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 855.291428] env[62066]: DEBUG oslo_concurrency.lockutils [None req-73e4bc46-8266-447d-984c-3ca2b99ffd23 tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] Acquired lock "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 855.291809] env[62066]: DEBUG oslo_concurrency.lockutils [None req-73e4bc46-8266-447d-984c-3ca2b99ffd23 tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 855.294794] env[62066]: DEBUG oslo_vmware.api [None req-5b85ae79-b544-4bc9-a850-092bba375a6b tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': task-1340939, 'name': CloneVM_Task} progress is 94%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.295644] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0fdae1a8-339a-4c57-84d6-30ecda0a84cc {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.297856] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4c61eb51-497c-4460-951e-d696b73eff28 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Acquiring lock "170e184a-f009-42cf-82dc-eb9696a00a24" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 855.298092] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4c61eb51-497c-4460-951e-d696b73eff28 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Lock "170e184a-f009-42cf-82dc-eb9696a00a24" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 855.306047] env[62066]: DEBUG oslo_vmware.api [None req-5327c80f-b6fd-46f0-ac98-3857740d94e2 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Task: {'id': task-1340946, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.287179} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.306259] env[62066]: DEBUG oslo_vmware.api [None req-22d0b1dd-f08e-49e4-a262-9fed18806112 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Task: {'id': task-1340945, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.309656] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-5327c80f-b6fd-46f0-ac98-3857740d94e2 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 855.309876] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-5327c80f-b6fd-46f0-ac98-3857740d94e2 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] [instance: 1e1a2ede-4ad8-4600-851e-6f2046b3f919] Deleted contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 855.310113] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-5327c80f-b6fd-46f0-ac98-3857740d94e2 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] [instance: 1e1a2ede-4ad8-4600-851e-6f2046b3f919] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 855.310294] env[62066]: INFO nova.compute.manager [None req-5327c80f-b6fd-46f0-ac98-3857740d94e2 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] [instance: 1e1a2ede-4ad8-4600-851e-6f2046b3f919] Took 1.26 seconds to destroy the instance on the hypervisor. [ 855.310550] env[62066]: DEBUG oslo.service.loopingcall [None req-5327c80f-b6fd-46f0-ac98-3857740d94e2 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 855.311526] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4ccd1f86-3a5c-4e35-bdaa-b8e507c9718e tempest-ServersTestFqdnHostnames-1367628746 tempest-ServersTestFqdnHostnames-1367628746-project-member] Acquiring lock "afaca059-fc4e-4de2-8fa0-d2f226ea7051" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 855.311743] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4ccd1f86-3a5c-4e35-bdaa-b8e507c9718e tempest-ServersTestFqdnHostnames-1367628746 tempest-ServersTestFqdnHostnames-1367628746-project-member] Lock "afaca059-fc4e-4de2-8fa0-d2f226ea7051" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 855.312755] env[62066]: DEBUG nova.compute.manager [-] [instance: 1e1a2ede-4ad8-4600-851e-6f2046b3f919] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 855.312831] env[62066]: DEBUG nova.network.neutron [-] [instance: 1e1a2ede-4ad8-4600-851e-6f2046b3f919] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 855.316269] env[62066]: DEBUG oslo_vmware.api [None req-73e4bc46-8266-447d-984c-3ca2b99ffd23 tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] Waiting for the task: (returnval){ [ 855.316269] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]520e6b9b-a5fd-2179-741a-91b7cb08e74d" [ 855.316269] env[62066]: _type = "Task" [ 855.316269] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.327844] env[62066]: DEBUG oslo_vmware.api [None req-73e4bc46-8266-447d-984c-3ca2b99ffd23 tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]520e6b9b-a5fd-2179-741a-91b7cb08e74d, 'name': SearchDatastore_Task, 'duration_secs': 0.011737} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.328208] env[62066]: DEBUG oslo_concurrency.lockutils [None req-73e4bc46-8266-447d-984c-3ca2b99ffd23 tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] Releasing lock "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 855.328473] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-73e4bc46-8266-447d-984c-3ca2b99ffd23 tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] [instance: 0bc31a8b-a5fc-4f6e-a4c0-024bf4260755] Processing image 50ff584c-3b50-4395-af07-3e66769bc9f7 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 855.328746] env[62066]: DEBUG oslo_concurrency.lockutils [None req-73e4bc46-8266-447d-984c-3ca2b99ffd23 tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 855.328901] env[62066]: DEBUG oslo_concurrency.lockutils [None req-73e4bc46-8266-447d-984c-3ca2b99ffd23 tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] Acquired lock "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 855.329113] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-73e4bc46-8266-447d-984c-3ca2b99ffd23 tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 855.329870] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-953cb41d-af39-46bd-a3e4-5277ce26ee9a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.337865] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-73e4bc46-8266-447d-984c-3ca2b99ffd23 tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 855.338063] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-73e4bc46-8266-447d-984c-3ca2b99ffd23 tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 855.338780] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-06e9c2a4-67ad-4c26-a37a-1b4c35f4ee3a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.343715] env[62066]: DEBUG oslo_vmware.api [None req-73e4bc46-8266-447d-984c-3ca2b99ffd23 tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] Waiting for the task: (returnval){ [ 855.343715] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]521025d3-c754-d09a-74f6-9820b52f3c1c" [ 855.343715] env[62066]: _type = "Task" [ 855.343715] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.354510] env[62066]: DEBUG oslo_vmware.api [None req-73e4bc46-8266-447d-984c-3ca2b99ffd23 tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]521025d3-c754-d09a-74f6-9820b52f3c1c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.366389] env[62066]: DEBUG oslo_vmware.api [None req-bace0778-9dfe-4de1-8747-93c45a08d78a tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': task-1340940, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.770973] env[62066]: DEBUG nova.network.neutron [req-b2ea8705-2cdd-43a8-bc17-8ef80d2acafa req-450fc733-5021-4208-a8f3-a25d0a66183e service nova] [instance: 0bc31a8b-a5fc-4f6e-a4c0-024bf4260755] Updated VIF entry in instance network info cache for port 8f34a054-f99b-447b-89d6-ab1ef9bc3c76. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 855.771357] env[62066]: DEBUG nova.network.neutron [req-b2ea8705-2cdd-43a8-bc17-8ef80d2acafa req-450fc733-5021-4208-a8f3-a25d0a66183e service nova] [instance: 0bc31a8b-a5fc-4f6e-a4c0-024bf4260755] Updating instance_info_cache with network_info: [{"id": "8f34a054-f99b-447b-89d6-ab1ef9bc3c76", "address": "fa:16:3e:06:43:d8", "network": {"id": "b92d2e33-767e-4958-b805-12e11b75c266", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-583710540-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "5281a37e471347b79e5dd6e70e03aa0f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "13e71dbb-4279-427c-b39d-ba5df9895e58", "external-id": "nsx-vlan-transportzone-417", "segmentation_id": 417, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8f34a054-f9", "ovs_interfaceid": "8f34a054-f99b-447b-89d6-ab1ef9bc3c76", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 855.791317] env[62066]: DEBUG oslo_vmware.api [None req-5b85ae79-b544-4bc9-a850-092bba375a6b tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': task-1340939, 'name': CloneVM_Task} progress is 100%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.796279] env[62066]: DEBUG oslo_vmware.api [None req-22d0b1dd-f08e-49e4-a262-9fed18806112 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Task: {'id': task-1340945, 'name': ReconfigVM_Task, 'duration_secs': 0.759297} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.799946] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-22d0b1dd-f08e-49e4-a262-9fed18806112 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] [instance: 3120c9a5-0558-4e83-b0fb-994a989398fb] Reconfigured VM instance instance-0000004d to attach disk [datastore1] 3120c9a5-0558-4e83-b0fb-994a989398fb/3120c9a5-0558-4e83-b0fb-994a989398fb.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 855.801359] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f692540e-6fc9-4f3e-a523-393482b15400 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.807417] env[62066]: DEBUG nova.compute.manager [None req-4c61eb51-497c-4460-951e-d696b73eff28 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 170e184a-f009-42cf-82dc-eb9696a00a24] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 855.814347] env[62066]: DEBUG oslo_vmware.api [None req-22d0b1dd-f08e-49e4-a262-9fed18806112 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Waiting for the task: (returnval){ [ 855.814347] env[62066]: value = "task-1340947" [ 855.814347] env[62066]: _type = "Task" [ 855.814347] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.818226] env[62066]: DEBUG nova.compute.manager [None req-4ccd1f86-3a5c-4e35-bdaa-b8e507c9718e tempest-ServersTestFqdnHostnames-1367628746 tempest-ServersTestFqdnHostnames-1367628746-project-member] [instance: afaca059-fc4e-4de2-8fa0-d2f226ea7051] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 855.833029] env[62066]: DEBUG oslo_vmware.api [None req-22d0b1dd-f08e-49e4-a262-9fed18806112 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Task: {'id': task-1340947, 'name': Rename_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.857577] env[62066]: DEBUG oslo_vmware.api [None req-73e4bc46-8266-447d-984c-3ca2b99ffd23 tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]521025d3-c754-d09a-74f6-9820b52f3c1c, 'name': SearchDatastore_Task, 'duration_secs': 0.030982} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.859164] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4e6e11de-8675-4074-b9a1-85d7eb254810 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.873517] env[62066]: DEBUG oslo_vmware.api [None req-bace0778-9dfe-4de1-8747-93c45a08d78a tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': task-1340940, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.873881] env[62066]: DEBUG oslo_vmware.api [None req-73e4bc46-8266-447d-984c-3ca2b99ffd23 tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] Waiting for the task: (returnval){ [ 855.873881] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]5235cd29-df5c-90ae-0017-202284a2be7b" [ 855.873881] env[62066]: _type = "Task" [ 855.873881] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.884729] env[62066]: DEBUG oslo_vmware.api [None req-73e4bc46-8266-447d-984c-3ca2b99ffd23 tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]5235cd29-df5c-90ae-0017-202284a2be7b, 'name': SearchDatastore_Task, 'duration_secs': 0.009882} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.887188] env[62066]: DEBUG oslo_concurrency.lockutils [None req-73e4bc46-8266-447d-984c-3ca2b99ffd23 tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] Releasing lock "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 855.887452] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-73e4bc46-8266-447d-984c-3ca2b99ffd23 tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk to [datastore1] 0bc31a8b-a5fc-4f6e-a4c0-024bf4260755/0bc31a8b-a5fc-4f6e-a4c0-024bf4260755.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 855.887922] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d3528470-5c6c-48da-880d-497dc156a1d9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.894515] env[62066]: DEBUG oslo_vmware.api [None req-73e4bc46-8266-447d-984c-3ca2b99ffd23 tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] Waiting for the task: (returnval){ [ 855.894515] env[62066]: value = "task-1340948" [ 855.894515] env[62066]: _type = "Task" [ 855.894515] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.904623] env[62066]: DEBUG oslo_vmware.api [None req-73e4bc46-8266-447d-984c-3ca2b99ffd23 tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] Task: {'id': task-1340948, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.041238] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-757e6a4e-6c6c-4a3f-acb6-4637e8289b30 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.049012] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3009a305-5abc-4c34-a58e-c5979ecfd988 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.081638] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fdc69ced-4e32-4367-9df2-db1126c4fcba {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.089273] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f6add3d-8f12-4efa-9f37-1e37ab9ad95a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.102459] env[62066]: DEBUG nova.compute.provider_tree [None req-68383c37-9698-4391-bbf2-b9ebf656dc03 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 856.104204] env[62066]: DEBUG nova.network.neutron [-] [instance: 1e1a2ede-4ad8-4600-851e-6f2046b3f919] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 856.274082] env[62066]: DEBUG oslo_concurrency.lockutils [req-b2ea8705-2cdd-43a8-bc17-8ef80d2acafa req-450fc733-5021-4208-a8f3-a25d0a66183e service nova] Releasing lock "refresh_cache-0bc31a8b-a5fc-4f6e-a4c0-024bf4260755" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 856.285945] env[62066]: DEBUG oslo_vmware.api [None req-5b85ae79-b544-4bc9-a850-092bba375a6b tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': task-1340939, 'name': CloneVM_Task, 'duration_secs': 1.573876} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.286260] env[62066]: INFO nova.virt.vmwareapi.vmops [None req-5b85ae79-b544-4bc9-a850-092bba375a6b tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: b08f27b1-429e-4d05-9a95-f45fd0cb2904] Created linked-clone VM from snapshot [ 856.286983] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a2e057b-d8ed-4607-8eda-aad46c71c6e4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.294323] env[62066]: DEBUG nova.virt.vmwareapi.images [None req-5b85ae79-b544-4bc9-a850-092bba375a6b tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: b08f27b1-429e-4d05-9a95-f45fd0cb2904] Uploading image 1147d77f-02f3-4d4b-a853-4a0e7d8b0a96 {{(pid=62066) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 856.314945] env[62066]: DEBUG oslo_vmware.rw_handles [None req-5b85ae79-b544-4bc9-a850-092bba375a6b tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 856.314945] env[62066]: value = "vm-286069" [ 856.314945] env[62066]: _type = "VirtualMachine" [ 856.314945] env[62066]: }. {{(pid=62066) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 856.315245] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-5f1f0d20-7dce-4ce4-85a9-a151764c1db4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.322243] env[62066]: DEBUG oslo_vmware.rw_handles [None req-5b85ae79-b544-4bc9-a850-092bba375a6b tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Lease: (returnval){ [ 856.322243] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]52441f06-0162-211d-e07f-b56b418dc364" [ 856.322243] env[62066]: _type = "HttpNfcLease" [ 856.322243] env[62066]: } obtained for exporting VM: (result){ [ 856.322243] env[62066]: value = "vm-286069" [ 856.322243] env[62066]: _type = "VirtualMachine" [ 856.322243] env[62066]: }. {{(pid=62066) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 856.322531] env[62066]: DEBUG oslo_vmware.api [None req-5b85ae79-b544-4bc9-a850-092bba375a6b tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Waiting for the lease: (returnval){ [ 856.322531] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]52441f06-0162-211d-e07f-b56b418dc364" [ 856.322531] env[62066]: _type = "HttpNfcLease" [ 856.322531] env[62066]: } to be ready. {{(pid=62066) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 856.335302] env[62066]: DEBUG oslo_vmware.api [None req-22d0b1dd-f08e-49e4-a262-9fed18806112 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Task: {'id': task-1340947, 'name': Rename_Task, 'duration_secs': 0.204415} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.337298] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-22d0b1dd-f08e-49e4-a262-9fed18806112 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] [instance: 3120c9a5-0558-4e83-b0fb-994a989398fb] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 856.337949] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ec704987-9da2-4d06-991f-e93909f562cd {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.340714] env[62066]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 856.340714] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]52441f06-0162-211d-e07f-b56b418dc364" [ 856.340714] env[62066]: _type = "HttpNfcLease" [ 856.340714] env[62066]: } is ready. {{(pid=62066) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 856.341280] env[62066]: DEBUG oslo_vmware.rw_handles [None req-5b85ae79-b544-4bc9-a850-092bba375a6b tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 856.341280] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]52441f06-0162-211d-e07f-b56b418dc364" [ 856.341280] env[62066]: _type = "HttpNfcLease" [ 856.341280] env[62066]: }. {{(pid=62066) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 856.342104] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-426c394f-5dd1-4181-b23e-f9c92c3899f7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.346086] env[62066]: DEBUG oslo_vmware.api [None req-22d0b1dd-f08e-49e4-a262-9fed18806112 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Waiting for the task: (returnval){ [ 856.346086] env[62066]: value = "task-1340950" [ 856.346086] env[62066]: _type = "Task" [ 856.346086] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.347311] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4c61eb51-497c-4460-951e-d696b73eff28 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 856.348268] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4ccd1f86-3a5c-4e35-bdaa-b8e507c9718e tempest-ServersTestFqdnHostnames-1367628746 tempest-ServersTestFqdnHostnames-1367628746-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 856.353807] env[62066]: DEBUG oslo_vmware.rw_handles [None req-5b85ae79-b544-4bc9-a850-092bba375a6b tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52052324-452e-7535-c568-8c4544842db1/disk-0.vmdk from lease info. {{(pid=62066) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 856.353978] env[62066]: DEBUG oslo_vmware.rw_handles [None req-5b85ae79-b544-4bc9-a850-092bba375a6b tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52052324-452e-7535-c568-8c4544842db1/disk-0.vmdk for reading. {{(pid=62066) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 856.422722] env[62066]: DEBUG oslo_vmware.api [None req-22d0b1dd-f08e-49e4-a262-9fed18806112 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Task: {'id': task-1340950, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.429555] env[62066]: DEBUG oslo_vmware.api [None req-bace0778-9dfe-4de1-8747-93c45a08d78a tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': task-1340940, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.433113] env[62066]: DEBUG oslo_vmware.api [None req-73e4bc46-8266-447d-984c-3ca2b99ffd23 tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] Task: {'id': task-1340948, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.490242] env[62066]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-6584208d-d7fa-47b3-8ca6-54115cbe0ac2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.605406] env[62066]: DEBUG nova.scheduler.client.report [None req-68383c37-9698-4391-bbf2-b9ebf656dc03 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 856.608672] env[62066]: INFO nova.compute.manager [-] [instance: 1e1a2ede-4ad8-4600-851e-6f2046b3f919] Took 1.30 seconds to deallocate network for instance. [ 856.857985] env[62066]: DEBUG oslo_vmware.api [None req-22d0b1dd-f08e-49e4-a262-9fed18806112 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Task: {'id': task-1340950, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.923808] env[62066]: DEBUG oslo_vmware.api [None req-73e4bc46-8266-447d-984c-3ca2b99ffd23 tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] Task: {'id': task-1340948, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.882804} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.927224] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-73e4bc46-8266-447d-984c-3ca2b99ffd23 tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk to [datastore1] 0bc31a8b-a5fc-4f6e-a4c0-024bf4260755/0bc31a8b-a5fc-4f6e-a4c0-024bf4260755.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 856.927716] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-73e4bc46-8266-447d-984c-3ca2b99ffd23 tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] [instance: 0bc31a8b-a5fc-4f6e-a4c0-024bf4260755] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 856.928085] env[62066]: DEBUG oslo_vmware.api [None req-bace0778-9dfe-4de1-8747-93c45a08d78a tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': task-1340940, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.928362] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ba758872-d51f-4d27-900b-cbc711c64b0b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.934667] env[62066]: DEBUG oslo_vmware.api [None req-73e4bc46-8266-447d-984c-3ca2b99ffd23 tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] Waiting for the task: (returnval){ [ 856.934667] env[62066]: value = "task-1340951" [ 856.934667] env[62066]: _type = "Task" [ 856.934667] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.943362] env[62066]: DEBUG oslo_vmware.api [None req-73e4bc46-8266-447d-984c-3ca2b99ffd23 tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] Task: {'id': task-1340951, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.109674] env[62066]: DEBUG nova.compute.manager [req-5dd514b4-1186-4ffc-b3f8-7f216a33351f req-b7d311dc-ad44-4e4d-a406-c692d84dc395 service nova] [instance: 1e1a2ede-4ad8-4600-851e-6f2046b3f919] Received event network-vif-deleted-9f537aea-f65c-4bcb-bd50-c419586c6aea {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 857.116382] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5327c80f-b6fd-46f0-ac98-3857740d94e2 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 857.117311] env[62066]: DEBUG oslo_concurrency.lockutils [None req-68383c37-9698-4391-bbf2-b9ebf656dc03 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.357s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 857.118036] env[62066]: DEBUG nova.compute.manager [None req-68383c37-9698-4391-bbf2-b9ebf656dc03 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: c8e0d47c-4421-4e00-9183-206fceeabc40] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 857.125230] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f2684bd2-e68b-4819-84d6-b5b0e8bf10bc tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.572s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 857.127238] env[62066]: INFO nova.compute.claims [None req-f2684bd2-e68b-4819-84d6-b5b0e8bf10bc tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 76306c3f-0674-40fe-9864-d82d11e4bb5e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 857.356716] env[62066]: DEBUG oslo_vmware.api [None req-22d0b1dd-f08e-49e4-a262-9fed18806112 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Task: {'id': task-1340950, 'name': PowerOnVM_Task, 'duration_secs': 0.799428} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.357122] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-22d0b1dd-f08e-49e4-a262-9fed18806112 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] [instance: 3120c9a5-0558-4e83-b0fb-994a989398fb] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 857.357441] env[62066]: INFO nova.compute.manager [None req-22d0b1dd-f08e-49e4-a262-9fed18806112 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] [instance: 3120c9a5-0558-4e83-b0fb-994a989398fb] Took 7.39 seconds to spawn the instance on the hypervisor. [ 857.357664] env[62066]: DEBUG nova.compute.manager [None req-22d0b1dd-f08e-49e4-a262-9fed18806112 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] [instance: 3120c9a5-0558-4e83-b0fb-994a989398fb] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 857.358464] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebb84656-6f0d-4404-a39e-2e8e4965cb7a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.422103] env[62066]: DEBUG oslo_vmware.api [None req-bace0778-9dfe-4de1-8747-93c45a08d78a tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': task-1340940, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.444953] env[62066]: DEBUG oslo_vmware.api [None req-73e4bc46-8266-447d-984c-3ca2b99ffd23 tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] Task: {'id': task-1340951, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067405} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.444953] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-73e4bc46-8266-447d-984c-3ca2b99ffd23 tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] [instance: 0bc31a8b-a5fc-4f6e-a4c0-024bf4260755] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 857.444953] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3833317-3cab-44a2-8544-79bd507c7bfb {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.467027] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-73e4bc46-8266-447d-984c-3ca2b99ffd23 tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] [instance: 0bc31a8b-a5fc-4f6e-a4c0-024bf4260755] Reconfiguring VM instance instance-0000004e to attach disk [datastore1] 0bc31a8b-a5fc-4f6e-a4c0-024bf4260755/0bc31a8b-a5fc-4f6e-a4c0-024bf4260755.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 857.467921] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-23713cb4-9690-4707-a352-fc4f94fa35bb {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.488516] env[62066]: DEBUG oslo_vmware.api [None req-73e4bc46-8266-447d-984c-3ca2b99ffd23 tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] Waiting for the task: (returnval){ [ 857.488516] env[62066]: value = "task-1340952" [ 857.488516] env[62066]: _type = "Task" [ 857.488516] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.496721] env[62066]: DEBUG oslo_vmware.api [None req-73e4bc46-8266-447d-984c-3ca2b99ffd23 tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] Task: {'id': task-1340952, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.639568] env[62066]: DEBUG nova.compute.utils [None req-68383c37-9698-4391-bbf2-b9ebf656dc03 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 857.642043] env[62066]: DEBUG nova.compute.manager [None req-68383c37-9698-4391-bbf2-b9ebf656dc03 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: c8e0d47c-4421-4e00-9183-206fceeabc40] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 857.642264] env[62066]: DEBUG nova.network.neutron [None req-68383c37-9698-4391-bbf2-b9ebf656dc03 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: c8e0d47c-4421-4e00-9183-206fceeabc40] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 857.697840] env[62066]: DEBUG nova.policy [None req-68383c37-9698-4391-bbf2-b9ebf656dc03 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd0d27da4c6fe4c9eb6e2a6e45c231a38', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2252d0f6f18a47fd8a1ddad8020a6db7', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 857.876576] env[62066]: INFO nova.compute.manager [None req-22d0b1dd-f08e-49e4-a262-9fed18806112 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] [instance: 3120c9a5-0558-4e83-b0fb-994a989398fb] Took 19.13 seconds to build instance. [ 857.924561] env[62066]: DEBUG oslo_vmware.api [None req-bace0778-9dfe-4de1-8747-93c45a08d78a tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': task-1340940, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.983355] env[62066]: DEBUG nova.network.neutron [None req-68383c37-9698-4391-bbf2-b9ebf656dc03 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: c8e0d47c-4421-4e00-9183-206fceeabc40] Successfully created port: bb9924e0-8b18-43b2-bda0-3e409fa16718 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 858.001039] env[62066]: DEBUG oslo_vmware.api [None req-73e4bc46-8266-447d-984c-3ca2b99ffd23 tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] Task: {'id': task-1340952, 'name': ReconfigVM_Task, 'duration_secs': 0.27618} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.002042] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-73e4bc46-8266-447d-984c-3ca2b99ffd23 tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] [instance: 0bc31a8b-a5fc-4f6e-a4c0-024bf4260755] Reconfigured VM instance instance-0000004e to attach disk [datastore1] 0bc31a8b-a5fc-4f6e-a4c0-024bf4260755/0bc31a8b-a5fc-4f6e-a4c0-024bf4260755.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 858.002042] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b3052efa-a80e-40b2-a5d6-8615d891efa4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.008585] env[62066]: DEBUG oslo_vmware.api [None req-73e4bc46-8266-447d-984c-3ca2b99ffd23 tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] Waiting for the task: (returnval){ [ 858.008585] env[62066]: value = "task-1340953" [ 858.008585] env[62066]: _type = "Task" [ 858.008585] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.016960] env[62066]: DEBUG oslo_vmware.api [None req-73e4bc46-8266-447d-984c-3ca2b99ffd23 tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] Task: {'id': task-1340953, 'name': Rename_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.142257] env[62066]: DEBUG nova.compute.manager [None req-68383c37-9698-4391-bbf2-b9ebf656dc03 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: c8e0d47c-4421-4e00-9183-206fceeabc40] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 858.378246] env[62066]: DEBUG oslo_concurrency.lockutils [None req-22d0b1dd-f08e-49e4-a262-9fed18806112 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Lock "3120c9a5-0558-4e83-b0fb-994a989398fb" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.642s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 858.424755] env[62066]: DEBUG oslo_vmware.api [None req-bace0778-9dfe-4de1-8747-93c45a08d78a tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': task-1340940, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.426810] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61e250d8-08ad-4f88-92d8-b70071c72f0a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.434874] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8a8fcf5-2b23-4cfd-bc9c-fc5a8523110c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.466572] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fef6158a-0150-46d2-879a-1d0eb7366690 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.474665] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2da657f-2501-40cd-9ebf-a57d47051ef7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.490074] env[62066]: DEBUG nova.compute.provider_tree [None req-f2684bd2-e68b-4819-84d6-b5b0e8bf10bc tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 858.518211] env[62066]: DEBUG oslo_vmware.api [None req-73e4bc46-8266-447d-984c-3ca2b99ffd23 tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] Task: {'id': task-1340953, 'name': Rename_Task, 'duration_secs': 0.148787} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.518561] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-73e4bc46-8266-447d-984c-3ca2b99ffd23 tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] [instance: 0bc31a8b-a5fc-4f6e-a4c0-024bf4260755] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 858.518837] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-57513c29-38c9-4713-a253-a17188ebb127 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.525513] env[62066]: DEBUG oslo_vmware.api [None req-73e4bc46-8266-447d-984c-3ca2b99ffd23 tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] Waiting for the task: (returnval){ [ 858.525513] env[62066]: value = "task-1340954" [ 858.525513] env[62066]: _type = "Task" [ 858.525513] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.533551] env[62066]: DEBUG oslo_vmware.api [None req-73e4bc46-8266-447d-984c-3ca2b99ffd23 tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] Task: {'id': task-1340954, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.609444] env[62066]: INFO nova.compute.manager [None req-6f12e7db-6fdf-4bbe-9b1e-ff4283071f33 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] [instance: 3120c9a5-0558-4e83-b0fb-994a989398fb] Rebuilding instance [ 858.657290] env[62066]: DEBUG nova.compute.manager [None req-6f12e7db-6fdf-4bbe-9b1e-ff4283071f33 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] [instance: 3120c9a5-0558-4e83-b0fb-994a989398fb] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 858.658228] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a26ed8a4-8322-4f9d-a5ad-f98f1017f0e0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.924344] env[62066]: DEBUG oslo_vmware.api [None req-bace0778-9dfe-4de1-8747-93c45a08d78a tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': task-1340940, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.993691] env[62066]: DEBUG nova.scheduler.client.report [None req-f2684bd2-e68b-4819-84d6-b5b0e8bf10bc tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 859.040905] env[62066]: DEBUG oslo_vmware.api [None req-73e4bc46-8266-447d-984c-3ca2b99ffd23 tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] Task: {'id': task-1340954, 'name': PowerOnVM_Task, 'duration_secs': 0.490515} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.041429] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-73e4bc46-8266-447d-984c-3ca2b99ffd23 tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] [instance: 0bc31a8b-a5fc-4f6e-a4c0-024bf4260755] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 859.041796] env[62066]: INFO nova.compute.manager [None req-73e4bc46-8266-447d-984c-3ca2b99ffd23 tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] [instance: 0bc31a8b-a5fc-4f6e-a4c0-024bf4260755] Took 6.62 seconds to spawn the instance on the hypervisor. [ 859.042153] env[62066]: DEBUG nova.compute.manager [None req-73e4bc46-8266-447d-984c-3ca2b99ffd23 tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] [instance: 0bc31a8b-a5fc-4f6e-a4c0-024bf4260755] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 859.043482] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-394a5c03-e002-4332-9bd2-010766ecbce2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.159073] env[62066]: DEBUG nova.compute.manager [None req-68383c37-9698-4391-bbf2-b9ebf656dc03 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: c8e0d47c-4421-4e00-9183-206fceeabc40] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 859.169689] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-6f12e7db-6fdf-4bbe-9b1e-ff4283071f33 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] [instance: 3120c9a5-0558-4e83-b0fb-994a989398fb] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 859.169927] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-df772521-cefc-481e-8977-f32709c856c1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.177493] env[62066]: DEBUG oslo_vmware.api [None req-6f12e7db-6fdf-4bbe-9b1e-ff4283071f33 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Waiting for the task: (returnval){ [ 859.177493] env[62066]: value = "task-1340955" [ 859.177493] env[62066]: _type = "Task" [ 859.177493] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.188679] env[62066]: DEBUG oslo_vmware.api [None req-6f12e7db-6fdf-4bbe-9b1e-ff4283071f33 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Task: {'id': task-1340955, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.190859] env[62066]: DEBUG nova.virt.hardware [None req-68383c37-9698-4391-bbf2-b9ebf656dc03 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-23T13:40:41Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-23T13:40:23Z,direct_url=,disk_format='vmdk',id=50ff584c-3b50-4395-af07-3e66769bc9f7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='eb52f7069a374c61ae946f052007c6d9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-23T13:40:24Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 859.191179] env[62066]: DEBUG nova.virt.hardware [None req-68383c37-9698-4391-bbf2-b9ebf656dc03 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 859.191409] env[62066]: DEBUG nova.virt.hardware [None req-68383c37-9698-4391-bbf2-b9ebf656dc03 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 859.191647] env[62066]: DEBUG nova.virt.hardware [None req-68383c37-9698-4391-bbf2-b9ebf656dc03 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 859.191835] env[62066]: DEBUG nova.virt.hardware [None req-68383c37-9698-4391-bbf2-b9ebf656dc03 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 859.192030] env[62066]: DEBUG nova.virt.hardware [None req-68383c37-9698-4391-bbf2-b9ebf656dc03 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 859.192299] env[62066]: DEBUG nova.virt.hardware [None req-68383c37-9698-4391-bbf2-b9ebf656dc03 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 859.192505] env[62066]: DEBUG nova.virt.hardware [None req-68383c37-9698-4391-bbf2-b9ebf656dc03 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 859.192714] env[62066]: DEBUG nova.virt.hardware [None req-68383c37-9698-4391-bbf2-b9ebf656dc03 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 859.192920] env[62066]: DEBUG nova.virt.hardware [None req-68383c37-9698-4391-bbf2-b9ebf656dc03 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 859.193156] env[62066]: DEBUG nova.virt.hardware [None req-68383c37-9698-4391-bbf2-b9ebf656dc03 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 859.194027] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4a1afef-02d9-432a-8d0f-42be4d612888 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.201754] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f79e1317-351a-485e-bd66-f2d7fcd86776 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.426318] env[62066]: DEBUG oslo_vmware.api [None req-bace0778-9dfe-4de1-8747-93c45a08d78a tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': task-1340940, 'name': ReconfigVM_Task} progress is 18%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.504027] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f2684bd2-e68b-4819-84d6-b5b0e8bf10bc tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.380s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 859.504027] env[62066]: DEBUG nova.compute.manager [None req-f2684bd2-e68b-4819-84d6-b5b0e8bf10bc tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 76306c3f-0674-40fe-9864-d82d11e4bb5e] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 859.507146] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5615ca43-5157-4e54-bcc7-b9e684503f45 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 11.897s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 859.507146] env[62066]: DEBUG nova.objects.instance [None req-5615ca43-5157-4e54-bcc7-b9e684503f45 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Lazy-loading 'resources' on Instance uuid 88daaaa6-f385-4161-bc74-9fffabb1145d {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 859.523016] env[62066]: DEBUG nova.compute.manager [req-e62ba25b-7634-45c9-b441-bd44767e439d req-4b15a0ad-a489-444a-b85f-1f01a5cb18c2 service nova] [instance: c8e0d47c-4421-4e00-9183-206fceeabc40] Received event network-vif-plugged-bb9924e0-8b18-43b2-bda0-3e409fa16718 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 859.523016] env[62066]: DEBUG oslo_concurrency.lockutils [req-e62ba25b-7634-45c9-b441-bd44767e439d req-4b15a0ad-a489-444a-b85f-1f01a5cb18c2 service nova] Acquiring lock "c8e0d47c-4421-4e00-9183-206fceeabc40-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 859.523016] env[62066]: DEBUG oslo_concurrency.lockutils [req-e62ba25b-7634-45c9-b441-bd44767e439d req-4b15a0ad-a489-444a-b85f-1f01a5cb18c2 service nova] Lock "c8e0d47c-4421-4e00-9183-206fceeabc40-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 859.523016] env[62066]: DEBUG oslo_concurrency.lockutils [req-e62ba25b-7634-45c9-b441-bd44767e439d req-4b15a0ad-a489-444a-b85f-1f01a5cb18c2 service nova] Lock "c8e0d47c-4421-4e00-9183-206fceeabc40-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 859.523016] env[62066]: DEBUG nova.compute.manager [req-e62ba25b-7634-45c9-b441-bd44767e439d req-4b15a0ad-a489-444a-b85f-1f01a5cb18c2 service nova] [instance: c8e0d47c-4421-4e00-9183-206fceeabc40] No waiting events found dispatching network-vif-plugged-bb9924e0-8b18-43b2-bda0-3e409fa16718 {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 859.524021] env[62066]: WARNING nova.compute.manager [req-e62ba25b-7634-45c9-b441-bd44767e439d req-4b15a0ad-a489-444a-b85f-1f01a5cb18c2 service nova] [instance: c8e0d47c-4421-4e00-9183-206fceeabc40] Received unexpected event network-vif-plugged-bb9924e0-8b18-43b2-bda0-3e409fa16718 for instance with vm_state building and task_state spawning. [ 859.568612] env[62066]: INFO nova.compute.manager [None req-73e4bc46-8266-447d-984c-3ca2b99ffd23 tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] [instance: 0bc31a8b-a5fc-4f6e-a4c0-024bf4260755] Took 19.43 seconds to build instance. [ 859.690615] env[62066]: DEBUG oslo_vmware.api [None req-6f12e7db-6fdf-4bbe-9b1e-ff4283071f33 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Task: {'id': task-1340955, 'name': PowerOffVM_Task, 'duration_secs': 0.117321} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.691169] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-6f12e7db-6fdf-4bbe-9b1e-ff4283071f33 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] [instance: 3120c9a5-0558-4e83-b0fb-994a989398fb] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 859.691598] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-6f12e7db-6fdf-4bbe-9b1e-ff4283071f33 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] [instance: 3120c9a5-0558-4e83-b0fb-994a989398fb] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 859.693445] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3c0d944-66da-475a-b3d8-96b8ca483b86 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.704218] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-6f12e7db-6fdf-4bbe-9b1e-ff4283071f33 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] [instance: 3120c9a5-0558-4e83-b0fb-994a989398fb] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 859.705031] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f2924efa-38de-4286-bd4d-5daef18feede {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.733626] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-6f12e7db-6fdf-4bbe-9b1e-ff4283071f33 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] [instance: 3120c9a5-0558-4e83-b0fb-994a989398fb] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 859.733626] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-6f12e7db-6fdf-4bbe-9b1e-ff4283071f33 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] [instance: 3120c9a5-0558-4e83-b0fb-994a989398fb] Deleting contents of the VM from datastore datastore1 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 859.733626] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-6f12e7db-6fdf-4bbe-9b1e-ff4283071f33 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Deleting the datastore file [datastore1] 3120c9a5-0558-4e83-b0fb-994a989398fb {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 859.735269] env[62066]: DEBUG nova.network.neutron [None req-68383c37-9698-4391-bbf2-b9ebf656dc03 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: c8e0d47c-4421-4e00-9183-206fceeabc40] Successfully updated port: bb9924e0-8b18-43b2-bda0-3e409fa16718 {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 859.735666] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e3f1c6c3-267e-4d9a-acc1-e7959b8ac4cd {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.742794] env[62066]: DEBUG oslo_vmware.api [None req-6f12e7db-6fdf-4bbe-9b1e-ff4283071f33 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Waiting for the task: (returnval){ [ 859.742794] env[62066]: value = "task-1340957" [ 859.742794] env[62066]: _type = "Task" [ 859.742794] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.753238] env[62066]: DEBUG oslo_vmware.api [None req-6f12e7db-6fdf-4bbe-9b1e-ff4283071f33 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Task: {'id': task-1340957, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.925199] env[62066]: DEBUG oslo_vmware.api [None req-bace0778-9dfe-4de1-8747-93c45a08d78a tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': task-1340940, 'name': ReconfigVM_Task} progress is 18%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.011623] env[62066]: DEBUG nova.compute.utils [None req-f2684bd2-e68b-4819-84d6-b5b0e8bf10bc tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 860.016185] env[62066]: DEBUG nova.compute.manager [None req-f2684bd2-e68b-4819-84d6-b5b0e8bf10bc tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 76306c3f-0674-40fe-9864-d82d11e4bb5e] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 860.016366] env[62066]: DEBUG nova.network.neutron [None req-f2684bd2-e68b-4819-84d6-b5b0e8bf10bc tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 76306c3f-0674-40fe-9864-d82d11e4bb5e] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 860.059363] env[62066]: DEBUG nova.policy [None req-f2684bd2-e68b-4819-84d6-b5b0e8bf10bc tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2ab20cbd0bab4ae7ba46d9135605a509', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd609babc987a42e2a8ddb4bfb9c3b103', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 860.069971] env[62066]: DEBUG oslo_concurrency.lockutils [None req-73e4bc46-8266-447d-984c-3ca2b99ffd23 tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] Lock "0bc31a8b-a5fc-4f6e-a4c0-024bf4260755" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.942s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 860.181790] env[62066]: INFO nova.compute.manager [None req-9a34415f-ab54-458b-b6ce-0d7cde2c3efe tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] [instance: 0bc31a8b-a5fc-4f6e-a4c0-024bf4260755] Rescuing [ 860.182110] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9a34415f-ab54-458b-b6ce-0d7cde2c3efe tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] Acquiring lock "refresh_cache-0bc31a8b-a5fc-4f6e-a4c0-024bf4260755" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 860.182278] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9a34415f-ab54-458b-b6ce-0d7cde2c3efe tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] Acquired lock "refresh_cache-0bc31a8b-a5fc-4f6e-a4c0-024bf4260755" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 860.182463] env[62066]: DEBUG nova.network.neutron [None req-9a34415f-ab54-458b-b6ce-0d7cde2c3efe tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] [instance: 0bc31a8b-a5fc-4f6e-a4c0-024bf4260755] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 860.239772] env[62066]: DEBUG oslo_concurrency.lockutils [None req-68383c37-9698-4391-bbf2-b9ebf656dc03 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Acquiring lock "refresh_cache-c8e0d47c-4421-4e00-9183-206fceeabc40" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 860.239914] env[62066]: DEBUG oslo_concurrency.lockutils [None req-68383c37-9698-4391-bbf2-b9ebf656dc03 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Acquired lock "refresh_cache-c8e0d47c-4421-4e00-9183-206fceeabc40" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 860.240112] env[62066]: DEBUG nova.network.neutron [None req-68383c37-9698-4391-bbf2-b9ebf656dc03 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: c8e0d47c-4421-4e00-9183-206fceeabc40] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 860.264065] env[62066]: DEBUG oslo_vmware.api [None req-6f12e7db-6fdf-4bbe-9b1e-ff4283071f33 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Task: {'id': task-1340957, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.304157} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.264065] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-6f12e7db-6fdf-4bbe-9b1e-ff4283071f33 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 860.264065] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-6f12e7db-6fdf-4bbe-9b1e-ff4283071f33 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] [instance: 3120c9a5-0558-4e83-b0fb-994a989398fb] Deleted contents of the VM from datastore datastore1 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 860.264065] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-6f12e7db-6fdf-4bbe-9b1e-ff4283071f33 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] [instance: 3120c9a5-0558-4e83-b0fb-994a989398fb] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 860.284017] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54dc48fb-3747-4b2a-87e3-e7bd196a92cc {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.291218] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9a595c7-5dfb-44bd-be95-1e8c1e1c7cc0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.328745] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-418d313b-eef2-4ebd-afb7-32b2e4b4fefd {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.339686] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3477ac0-eec8-4a5a-9e74-d56e9192d3b1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.343331] env[62066]: DEBUG nova.network.neutron [None req-f2684bd2-e68b-4819-84d6-b5b0e8bf10bc tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 76306c3f-0674-40fe-9864-d82d11e4bb5e] Successfully created port: 28f3b624-430a-474d-a24b-9e54db92e9e2 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 860.355960] env[62066]: DEBUG nova.compute.provider_tree [None req-5615ca43-5157-4e54-bcc7-b9e684503f45 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 860.427235] env[62066]: DEBUG oslo_vmware.api [None req-bace0778-9dfe-4de1-8747-93c45a08d78a tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': task-1340940, 'name': ReconfigVM_Task, 'duration_secs': 5.81616} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.427452] env[62066]: DEBUG oslo_concurrency.lockutils [None req-bace0778-9dfe-4de1-8747-93c45a08d78a tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Releasing lock "459f432b-e817-405f-9c3e-dfa5d957788c" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 860.427664] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-bace0778-9dfe-4de1-8747-93c45a08d78a tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: 459f432b-e817-405f-9c3e-dfa5d957788c] Reconfigured VM to detach interface {{(pid=62066) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 860.516488] env[62066]: DEBUG nova.compute.manager [None req-f2684bd2-e68b-4819-84d6-b5b0e8bf10bc tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 76306c3f-0674-40fe-9864-d82d11e4bb5e] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 860.785598] env[62066]: DEBUG nova.network.neutron [None req-68383c37-9698-4391-bbf2-b9ebf656dc03 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: c8e0d47c-4421-4e00-9183-206fceeabc40] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 860.862262] env[62066]: DEBUG nova.scheduler.client.report [None req-5615ca43-5157-4e54-bcc7-b9e684503f45 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 860.992778] env[62066]: DEBUG nova.network.neutron [None req-9a34415f-ab54-458b-b6ce-0d7cde2c3efe tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] [instance: 0bc31a8b-a5fc-4f6e-a4c0-024bf4260755] Updating instance_info_cache with network_info: [{"id": "8f34a054-f99b-447b-89d6-ab1ef9bc3c76", "address": "fa:16:3e:06:43:d8", "network": {"id": "b92d2e33-767e-4958-b805-12e11b75c266", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-583710540-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "5281a37e471347b79e5dd6e70e03aa0f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "13e71dbb-4279-427c-b39d-ba5df9895e58", "external-id": "nsx-vlan-transportzone-417", "segmentation_id": 417, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8f34a054-f9", "ovs_interfaceid": "8f34a054-f99b-447b-89d6-ab1ef9bc3c76", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 861.011659] env[62066]: DEBUG nova.network.neutron [None req-68383c37-9698-4391-bbf2-b9ebf656dc03 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: c8e0d47c-4421-4e00-9183-206fceeabc40] Updating instance_info_cache with network_info: [{"id": "bb9924e0-8b18-43b2-bda0-3e409fa16718", "address": "fa:16:3e:b1:e4:41", "network": {"id": "5b89b2d4-71dd-4cd1-9edb-f4040be26540", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-283918837-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2252d0f6f18a47fd8a1ddad8020a6db7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "352165bb-004f-4180-9627-3a275dbe18af", "external-id": "nsx-vlan-transportzone-926", "segmentation_id": 926, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbb9924e0-8b", "ovs_interfaceid": "bb9924e0-8b18-43b2-bda0-3e409fa16718", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 861.310193] env[62066]: DEBUG nova.virt.hardware [None req-6f12e7db-6fdf-4bbe-9b1e-ff4283071f33 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-23T13:40:41Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-23T13:40:23Z,direct_url=,disk_format='vmdk',id=50ff584c-3b50-4395-af07-3e66769bc9f7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='eb52f7069a374c61ae946f052007c6d9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-23T13:40:24Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 861.310416] env[62066]: DEBUG nova.virt.hardware [None req-6f12e7db-6fdf-4bbe-9b1e-ff4283071f33 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 861.310640] env[62066]: DEBUG nova.virt.hardware [None req-6f12e7db-6fdf-4bbe-9b1e-ff4283071f33 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 861.310864] env[62066]: DEBUG nova.virt.hardware [None req-6f12e7db-6fdf-4bbe-9b1e-ff4283071f33 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 861.311032] env[62066]: DEBUG nova.virt.hardware [None req-6f12e7db-6fdf-4bbe-9b1e-ff4283071f33 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 861.311206] env[62066]: DEBUG nova.virt.hardware [None req-6f12e7db-6fdf-4bbe-9b1e-ff4283071f33 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 861.311507] env[62066]: DEBUG nova.virt.hardware [None req-6f12e7db-6fdf-4bbe-9b1e-ff4283071f33 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 861.311692] env[62066]: DEBUG nova.virt.hardware [None req-6f12e7db-6fdf-4bbe-9b1e-ff4283071f33 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 861.311891] env[62066]: DEBUG nova.virt.hardware [None req-6f12e7db-6fdf-4bbe-9b1e-ff4283071f33 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 861.312121] env[62066]: DEBUG nova.virt.hardware [None req-6f12e7db-6fdf-4bbe-9b1e-ff4283071f33 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 861.312314] env[62066]: DEBUG nova.virt.hardware [None req-6f12e7db-6fdf-4bbe-9b1e-ff4283071f33 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 861.313262] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72319ac8-064d-4423-8b68-ebc650201ad7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.324431] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3b9286c-d073-4329-bc57-62a5bfcc3399 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.341762] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-6f12e7db-6fdf-4bbe-9b1e-ff4283071f33 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] [instance: 3120c9a5-0558-4e83-b0fb-994a989398fb] Instance VIF info [] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 861.347818] env[62066]: DEBUG oslo.service.loopingcall [None req-6f12e7db-6fdf-4bbe-9b1e-ff4283071f33 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 861.348186] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3120c9a5-0558-4e83-b0fb-994a989398fb] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 861.348420] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f77bf547-318a-4e46-b780-deb1848b882d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.365168] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5615ca43-5157-4e54-bcc7-b9e684503f45 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.859s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 861.368750] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cf3b688f-a077-42d2-b46f-272004213881 tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 13.636s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 861.369014] env[62066]: DEBUG nova.objects.instance [None req-cf3b688f-a077-42d2-b46f-272004213881 tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] Lazy-loading 'resources' on Instance uuid daffaf51-4c45-44aa-8fc2-4db066a09971 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 861.371074] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 861.371074] env[62066]: value = "task-1340958" [ 861.371074] env[62066]: _type = "Task" [ 861.371074] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.379673] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1340958, 'name': CreateVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.399020] env[62066]: INFO nova.scheduler.client.report [None req-5615ca43-5157-4e54-bcc7-b9e684503f45 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Deleted allocations for instance 88daaaa6-f385-4161-bc74-9fffabb1145d [ 861.496083] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9a34415f-ab54-458b-b6ce-0d7cde2c3efe tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] Releasing lock "refresh_cache-0bc31a8b-a5fc-4f6e-a4c0-024bf4260755" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 861.513952] env[62066]: DEBUG oslo_concurrency.lockutils [None req-68383c37-9698-4391-bbf2-b9ebf656dc03 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Releasing lock "refresh_cache-c8e0d47c-4421-4e00-9183-206fceeabc40" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 861.515031] env[62066]: DEBUG nova.compute.manager [None req-68383c37-9698-4391-bbf2-b9ebf656dc03 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: c8e0d47c-4421-4e00-9183-206fceeabc40] Instance network_info: |[{"id": "bb9924e0-8b18-43b2-bda0-3e409fa16718", "address": "fa:16:3e:b1:e4:41", "network": {"id": "5b89b2d4-71dd-4cd1-9edb-f4040be26540", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-283918837-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2252d0f6f18a47fd8a1ddad8020a6db7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "352165bb-004f-4180-9627-3a275dbe18af", "external-id": "nsx-vlan-transportzone-926", "segmentation_id": 926, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbb9924e0-8b", "ovs_interfaceid": "bb9924e0-8b18-43b2-bda0-3e409fa16718", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 861.515165] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-68383c37-9698-4391-bbf2-b9ebf656dc03 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: c8e0d47c-4421-4e00-9183-206fceeabc40] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b1:e4:41', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '352165bb-004f-4180-9627-3a275dbe18af', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'bb9924e0-8b18-43b2-bda0-3e409fa16718', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 861.523535] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-68383c37-9698-4391-bbf2-b9ebf656dc03 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Creating folder: Project (2252d0f6f18a47fd8a1ddad8020a6db7). Parent ref: group-v285980. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 861.528940] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ee35e2b6-320f-4b24-9798-d36473171f8e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.532072] env[62066]: DEBUG nova.compute.manager [None req-f2684bd2-e68b-4819-84d6-b5b0e8bf10bc tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 76306c3f-0674-40fe-9864-d82d11e4bb5e] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 861.546250] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-68383c37-9698-4391-bbf2-b9ebf656dc03 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Created folder: Project (2252d0f6f18a47fd8a1ddad8020a6db7) in parent group-v285980. [ 861.546507] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-68383c37-9698-4391-bbf2-b9ebf656dc03 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Creating folder: Instances. Parent ref: group-v286074. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 861.546932] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-58ba285f-abfc-428e-85b7-8cb221a0b5b1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.557349] env[62066]: DEBUG nova.virt.hardware [None req-f2684bd2-e68b-4819-84d6-b5b0e8bf10bc tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-23T13:40:41Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-23T13:40:23Z,direct_url=,disk_format='vmdk',id=50ff584c-3b50-4395-af07-3e66769bc9f7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='eb52f7069a374c61ae946f052007c6d9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-23T13:40:24Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 861.557738] env[62066]: DEBUG nova.virt.hardware [None req-f2684bd2-e68b-4819-84d6-b5b0e8bf10bc tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 861.557934] env[62066]: DEBUG nova.virt.hardware [None req-f2684bd2-e68b-4819-84d6-b5b0e8bf10bc tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 861.558144] env[62066]: DEBUG nova.virt.hardware [None req-f2684bd2-e68b-4819-84d6-b5b0e8bf10bc tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 861.558297] env[62066]: DEBUG nova.virt.hardware [None req-f2684bd2-e68b-4819-84d6-b5b0e8bf10bc tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 861.558450] env[62066]: DEBUG nova.virt.hardware [None req-f2684bd2-e68b-4819-84d6-b5b0e8bf10bc tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 861.558693] env[62066]: DEBUG nova.virt.hardware [None req-f2684bd2-e68b-4819-84d6-b5b0e8bf10bc tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 861.558870] env[62066]: DEBUG nova.virt.hardware [None req-f2684bd2-e68b-4819-84d6-b5b0e8bf10bc tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 861.559295] env[62066]: DEBUG nova.virt.hardware [None req-f2684bd2-e68b-4819-84d6-b5b0e8bf10bc tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 861.560133] env[62066]: DEBUG nova.virt.hardware [None req-f2684bd2-e68b-4819-84d6-b5b0e8bf10bc tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 861.560406] env[62066]: DEBUG nova.virt.hardware [None req-f2684bd2-e68b-4819-84d6-b5b0e8bf10bc tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 861.561987] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f900908-b737-4370-8399-05e944b450d7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.565966] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-68383c37-9698-4391-bbf2-b9ebf656dc03 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Created folder: Instances in parent group-v286074. [ 861.566225] env[62066]: DEBUG oslo.service.loopingcall [None req-68383c37-9698-4391-bbf2-b9ebf656dc03 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 861.566786] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c8e0d47c-4421-4e00-9183-206fceeabc40] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 861.567022] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-60621d1e-a373-41c4-af26-aec51b2b04f4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.587221] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12ae2ba0-0bcb-40f6-a600-a6d24333e75d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.592530] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 861.592530] env[62066]: value = "task-1340961" [ 861.592530] env[62066]: _type = "Task" [ 861.592530] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.613183] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1340961, 'name': CreateVM_Task} progress is 10%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.733660] env[62066]: DEBUG nova.compute.manager [req-b3860259-ac6e-4483-a6cc-ef9f6f0b8dc7 req-aa15c4c9-f411-4565-b421-9ae4b26d4133 service nova] [instance: c8e0d47c-4421-4e00-9183-206fceeabc40] Received event network-changed-bb9924e0-8b18-43b2-bda0-3e409fa16718 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 861.733660] env[62066]: DEBUG nova.compute.manager [req-b3860259-ac6e-4483-a6cc-ef9f6f0b8dc7 req-aa15c4c9-f411-4565-b421-9ae4b26d4133 service nova] [instance: c8e0d47c-4421-4e00-9183-206fceeabc40] Refreshing instance network info cache due to event network-changed-bb9924e0-8b18-43b2-bda0-3e409fa16718. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 861.733892] env[62066]: DEBUG oslo_concurrency.lockutils [req-b3860259-ac6e-4483-a6cc-ef9f6f0b8dc7 req-aa15c4c9-f411-4565-b421-9ae4b26d4133 service nova] Acquiring lock "refresh_cache-c8e0d47c-4421-4e00-9183-206fceeabc40" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 861.734194] env[62066]: DEBUG oslo_concurrency.lockutils [req-b3860259-ac6e-4483-a6cc-ef9f6f0b8dc7 req-aa15c4c9-f411-4565-b421-9ae4b26d4133 service nova] Acquired lock "refresh_cache-c8e0d47c-4421-4e00-9183-206fceeabc40" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 861.734456] env[62066]: DEBUG nova.network.neutron [req-b3860259-ac6e-4483-a6cc-ef9f6f0b8dc7 req-aa15c4c9-f411-4565-b421-9ae4b26d4133 service nova] [instance: c8e0d47c-4421-4e00-9183-206fceeabc40] Refreshing network info cache for port bb9924e0-8b18-43b2-bda0-3e409fa16718 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 861.885405] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1340958, 'name': CreateVM_Task} progress is 99%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.908044] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5615ca43-5157-4e54-bcc7-b9e684503f45 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Lock "88daaaa6-f385-4161-bc74-9fffabb1145d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 17.468s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 861.914617] env[62066]: DEBUG oslo_concurrency.lockutils [None req-bace0778-9dfe-4de1-8747-93c45a08d78a tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Acquiring lock "refresh_cache-459f432b-e817-405f-9c3e-dfa5d957788c" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 861.914921] env[62066]: DEBUG oslo_concurrency.lockutils [None req-bace0778-9dfe-4de1-8747-93c45a08d78a tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Acquired lock "refresh_cache-459f432b-e817-405f-9c3e-dfa5d957788c" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 861.915168] env[62066]: DEBUG nova.network.neutron [None req-bace0778-9dfe-4de1-8747-93c45a08d78a tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: 459f432b-e817-405f-9c3e-dfa5d957788c] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 861.952531] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ba2d4aa3-3165-47fd-a67a-a02fe7b33642 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Acquiring lock "459f432b-e817-405f-9c3e-dfa5d957788c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 861.954018] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ba2d4aa3-3165-47fd-a67a-a02fe7b33642 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Lock "459f432b-e817-405f-9c3e-dfa5d957788c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 861.954018] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ba2d4aa3-3165-47fd-a67a-a02fe7b33642 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Acquiring lock "459f432b-e817-405f-9c3e-dfa5d957788c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 861.954018] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ba2d4aa3-3165-47fd-a67a-a02fe7b33642 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Lock "459f432b-e817-405f-9c3e-dfa5d957788c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 861.954018] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ba2d4aa3-3165-47fd-a67a-a02fe7b33642 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Lock "459f432b-e817-405f-9c3e-dfa5d957788c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 861.956418] env[62066]: INFO nova.compute.manager [None req-ba2d4aa3-3165-47fd-a67a-a02fe7b33642 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: 459f432b-e817-405f-9c3e-dfa5d957788c] Terminating instance [ 861.958535] env[62066]: DEBUG nova.compute.manager [None req-ba2d4aa3-3165-47fd-a67a-a02fe7b33642 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: 459f432b-e817-405f-9c3e-dfa5d957788c] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 861.958792] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-ba2d4aa3-3165-47fd-a67a-a02fe7b33642 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: 459f432b-e817-405f-9c3e-dfa5d957788c] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 861.960151] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a251c5a7-bc64-491f-9ab5-e0981e508cf6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.970506] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-ba2d4aa3-3165-47fd-a67a-a02fe7b33642 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: 459f432b-e817-405f-9c3e-dfa5d957788c] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 861.970810] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ac6e9f0f-5ce9-48d0-add2-1e08246459b2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.978944] env[62066]: DEBUG oslo_vmware.api [None req-ba2d4aa3-3165-47fd-a67a-a02fe7b33642 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Waiting for the task: (returnval){ [ 861.978944] env[62066]: value = "task-1340962" [ 861.978944] env[62066]: _type = "Task" [ 861.978944] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.991627] env[62066]: DEBUG oslo_vmware.api [None req-ba2d4aa3-3165-47fd-a67a-a02fe7b33642 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': task-1340962, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.042350] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-9a34415f-ab54-458b-b6ce-0d7cde2c3efe tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] [instance: 0bc31a8b-a5fc-4f6e-a4c0-024bf4260755] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 862.042622] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b15e10e3-b59d-4d35-8f22-d266309c7dc8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.052223] env[62066]: DEBUG oslo_vmware.api [None req-9a34415f-ab54-458b-b6ce-0d7cde2c3efe tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] Waiting for the task: (returnval){ [ 862.052223] env[62066]: value = "task-1340963" [ 862.052223] env[62066]: _type = "Task" [ 862.052223] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.066771] env[62066]: DEBUG oslo_vmware.api [None req-9a34415f-ab54-458b-b6ce-0d7cde2c3efe tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] Task: {'id': task-1340963, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.099255] env[62066]: DEBUG nova.network.neutron [None req-f2684bd2-e68b-4819-84d6-b5b0e8bf10bc tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 76306c3f-0674-40fe-9864-d82d11e4bb5e] Successfully updated port: 28f3b624-430a-474d-a24b-9e54db92e9e2 {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 862.110502] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1340961, 'name': CreateVM_Task, 'duration_secs': 0.43592} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.111283] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c8e0d47c-4421-4e00-9183-206fceeabc40] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 862.112739] env[62066]: DEBUG oslo_concurrency.lockutils [None req-68383c37-9698-4391-bbf2-b9ebf656dc03 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 862.112994] env[62066]: DEBUG oslo_concurrency.lockutils [None req-68383c37-9698-4391-bbf2-b9ebf656dc03 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Acquired lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 862.113406] env[62066]: DEBUG oslo_concurrency.lockutils [None req-68383c37-9698-4391-bbf2-b9ebf656dc03 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 862.113983] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-360c98ac-6067-4eac-9fb0-2cc5382528f8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.123163] env[62066]: DEBUG oslo_vmware.api [None req-68383c37-9698-4391-bbf2-b9ebf656dc03 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Waiting for the task: (returnval){ [ 862.123163] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]520f4fe9-dba4-0fb2-d321-bc9c7b3ed32a" [ 862.123163] env[62066]: _type = "Task" [ 862.123163] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.132046] env[62066]: DEBUG oslo_vmware.api [None req-68383c37-9698-4391-bbf2-b9ebf656dc03 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]520f4fe9-dba4-0fb2-d321-bc9c7b3ed32a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.165863] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a738b16-bf15-4755-85a8-9174bf4b2f9d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.174223] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4cb84dc-1816-4a9e-9d65-c5e2a387685a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.207257] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-046225b2-3a98-4aeb-89de-aa03fb1d1502 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.214842] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e472b8e3-4c84-4d2d-8d9c-38f317601a36 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.228643] env[62066]: DEBUG nova.compute.provider_tree [None req-cf3b688f-a077-42d2-b46f-272004213881 tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 862.380549] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1340958, 'name': CreateVM_Task} progress is 99%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.448845] env[62066]: DEBUG nova.network.neutron [req-b3860259-ac6e-4483-a6cc-ef9f6f0b8dc7 req-aa15c4c9-f411-4565-b421-9ae4b26d4133 service nova] [instance: c8e0d47c-4421-4e00-9183-206fceeabc40] Updated VIF entry in instance network info cache for port bb9924e0-8b18-43b2-bda0-3e409fa16718. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 862.449239] env[62066]: DEBUG nova.network.neutron [req-b3860259-ac6e-4483-a6cc-ef9f6f0b8dc7 req-aa15c4c9-f411-4565-b421-9ae4b26d4133 service nova] [instance: c8e0d47c-4421-4e00-9183-206fceeabc40] Updating instance_info_cache with network_info: [{"id": "bb9924e0-8b18-43b2-bda0-3e409fa16718", "address": "fa:16:3e:b1:e4:41", "network": {"id": "5b89b2d4-71dd-4cd1-9edb-f4040be26540", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-283918837-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2252d0f6f18a47fd8a1ddad8020a6db7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "352165bb-004f-4180-9627-3a275dbe18af", "external-id": "nsx-vlan-transportzone-926", "segmentation_id": 926, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbb9924e0-8b", "ovs_interfaceid": "bb9924e0-8b18-43b2-bda0-3e409fa16718", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 862.488636] env[62066]: DEBUG oslo_vmware.api [None req-ba2d4aa3-3165-47fd-a67a-a02fe7b33642 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': task-1340962, 'name': PowerOffVM_Task, 'duration_secs': 0.204309} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.490974] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-ba2d4aa3-3165-47fd-a67a-a02fe7b33642 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: 459f432b-e817-405f-9c3e-dfa5d957788c] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 862.491185] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-ba2d4aa3-3165-47fd-a67a-a02fe7b33642 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: 459f432b-e817-405f-9c3e-dfa5d957788c] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 862.491713] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d491fca5-fbff-4273-8751-c439089d5de9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.552463] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-ba2d4aa3-3165-47fd-a67a-a02fe7b33642 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: 459f432b-e817-405f-9c3e-dfa5d957788c] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 862.552695] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-ba2d4aa3-3165-47fd-a67a-a02fe7b33642 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: 459f432b-e817-405f-9c3e-dfa5d957788c] Deleting contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 862.552886] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-ba2d4aa3-3165-47fd-a67a-a02fe7b33642 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Deleting the datastore file [datastore2] 459f432b-e817-405f-9c3e-dfa5d957788c {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 862.558633] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-97b42f18-bda0-4915-8f71-cf37dca3ae75 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.564896] env[62066]: DEBUG oslo_vmware.api [None req-9a34415f-ab54-458b-b6ce-0d7cde2c3efe tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] Task: {'id': task-1340963, 'name': PowerOffVM_Task, 'duration_secs': 0.237959} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.566138] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-9a34415f-ab54-458b-b6ce-0d7cde2c3efe tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] [instance: 0bc31a8b-a5fc-4f6e-a4c0-024bf4260755] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 862.566491] env[62066]: DEBUG oslo_vmware.api [None req-ba2d4aa3-3165-47fd-a67a-a02fe7b33642 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Waiting for the task: (returnval){ [ 862.566491] env[62066]: value = "task-1340965" [ 862.566491] env[62066]: _type = "Task" [ 862.566491] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.567181] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4965881a-7a73-4af1-8df2-23e20d34f262 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.576798] env[62066]: DEBUG oslo_vmware.api [None req-ba2d4aa3-3165-47fd-a67a-a02fe7b33642 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': task-1340965, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.593309] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69d40cc1-bda1-4b1d-b675-afef61241bea {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.605979] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f2684bd2-e68b-4819-84d6-b5b0e8bf10bc tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Acquiring lock "refresh_cache-76306c3f-0674-40fe-9864-d82d11e4bb5e" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 862.606167] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f2684bd2-e68b-4819-84d6-b5b0e8bf10bc tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Acquired lock "refresh_cache-76306c3f-0674-40fe-9864-d82d11e4bb5e" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 862.606322] env[62066]: DEBUG nova.network.neutron [None req-f2684bd2-e68b-4819-84d6-b5b0e8bf10bc tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 76306c3f-0674-40fe-9864-d82d11e4bb5e] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 862.628259] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-9a34415f-ab54-458b-b6ce-0d7cde2c3efe tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] [instance: 0bc31a8b-a5fc-4f6e-a4c0-024bf4260755] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 862.628822] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-89ec9fbc-bd27-432c-bf3c-207e588a0607 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.636607] env[62066]: DEBUG oslo_vmware.api [None req-68383c37-9698-4391-bbf2-b9ebf656dc03 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]520f4fe9-dba4-0fb2-d321-bc9c7b3ed32a, 'name': SearchDatastore_Task, 'duration_secs': 0.016398} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.638052] env[62066]: DEBUG oslo_concurrency.lockutils [None req-68383c37-9698-4391-bbf2-b9ebf656dc03 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Releasing lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 862.638313] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-68383c37-9698-4391-bbf2-b9ebf656dc03 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: c8e0d47c-4421-4e00-9183-206fceeabc40] Processing image 50ff584c-3b50-4395-af07-3e66769bc9f7 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 862.638551] env[62066]: DEBUG oslo_concurrency.lockutils [None req-68383c37-9698-4391-bbf2-b9ebf656dc03 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 862.638705] env[62066]: DEBUG oslo_concurrency.lockutils [None req-68383c37-9698-4391-bbf2-b9ebf656dc03 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Acquired lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 862.638890] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-68383c37-9698-4391-bbf2-b9ebf656dc03 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 862.639229] env[62066]: DEBUG oslo_vmware.api [None req-9a34415f-ab54-458b-b6ce-0d7cde2c3efe tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] Waiting for the task: (returnval){ [ 862.639229] env[62066]: value = "task-1340966" [ 862.639229] env[62066]: _type = "Task" [ 862.639229] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.639421] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b4522b25-971b-4bf3-b2d1-4d7dc7f0e9e8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.654137] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-9a34415f-ab54-458b-b6ce-0d7cde2c3efe tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] [instance: 0bc31a8b-a5fc-4f6e-a4c0-024bf4260755] VM already powered off {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 862.654374] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-9a34415f-ab54-458b-b6ce-0d7cde2c3efe tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] [instance: 0bc31a8b-a5fc-4f6e-a4c0-024bf4260755] Processing image 50ff584c-3b50-4395-af07-3e66769bc9f7 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 862.654614] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9a34415f-ab54-458b-b6ce-0d7cde2c3efe tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 862.654771] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9a34415f-ab54-458b-b6ce-0d7cde2c3efe tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] Acquired lock "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 862.654963] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-9a34415f-ab54-458b-b6ce-0d7cde2c3efe tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 862.655541] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9fef56d7-d302-4286-bf54-f97401d03a98 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.661312] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-68383c37-9698-4391-bbf2-b9ebf656dc03 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 862.661512] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-68383c37-9698-4391-bbf2-b9ebf656dc03 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 862.662234] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-762194d4-9c7a-4a64-b212-4c909ffe9d15 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.667365] env[62066]: DEBUG oslo_vmware.api [None req-68383c37-9698-4391-bbf2-b9ebf656dc03 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Waiting for the task: (returnval){ [ 862.667365] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]52d3bc01-8390-17cd-aafe-55db5fb6b0d2" [ 862.667365] env[62066]: _type = "Task" [ 862.667365] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.677673] env[62066]: DEBUG oslo_vmware.api [None req-68383c37-9698-4391-bbf2-b9ebf656dc03 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52d3bc01-8390-17cd-aafe-55db5fb6b0d2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.731944] env[62066]: DEBUG nova.scheduler.client.report [None req-cf3b688f-a077-42d2-b46f-272004213881 tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 862.737618] env[62066]: INFO nova.network.neutron [None req-bace0778-9dfe-4de1-8747-93c45a08d78a tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: 459f432b-e817-405f-9c3e-dfa5d957788c] Port 56efe6b6-2fe3-4c7b-9d59-f8ecf74ac772 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 862.738322] env[62066]: DEBUG nova.network.neutron [None req-bace0778-9dfe-4de1-8747-93c45a08d78a tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: 459f432b-e817-405f-9c3e-dfa5d957788c] Updating instance_info_cache with network_info: [{"id": "cb74e58f-b481-4a0f-861b-96381e2d52dc", "address": "fa:16:3e:87:b9:2f", "network": {"id": "449dfe4e-bb49-43b6-9ba7-b57af74ebfb3", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-282029676-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.159", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0d4e33a99c7741fb8cdd97f4ec5dbbd8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db68bd64-5b56-49af-a075-13dcf85cb2e0", "external-id": "nsx-vlan-transportzone-590", "segmentation_id": 590, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcb74e58f-b4", "ovs_interfaceid": "cb74e58f-b481-4a0f-861b-96381e2d52dc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 862.881835] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1340958, 'name': CreateVM_Task, 'duration_secs': 1.434454} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.882035] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3120c9a5-0558-4e83-b0fb-994a989398fb] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 862.882465] env[62066]: DEBUG oslo_concurrency.lockutils [None req-6f12e7db-6fdf-4bbe-9b1e-ff4283071f33 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 862.882631] env[62066]: DEBUG oslo_concurrency.lockutils [None req-6f12e7db-6fdf-4bbe-9b1e-ff4283071f33 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Acquired lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 862.882975] env[62066]: DEBUG oslo_concurrency.lockutils [None req-6f12e7db-6fdf-4bbe-9b1e-ff4283071f33 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 862.883248] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-966867ff-016e-4ea6-bcd6-6296ee58c311 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.888333] env[62066]: DEBUG oslo_vmware.api [None req-6f12e7db-6fdf-4bbe-9b1e-ff4283071f33 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Waiting for the task: (returnval){ [ 862.888333] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]52566e60-6421-1ba3-b275-a911841e77f8" [ 862.888333] env[62066]: _type = "Task" [ 862.888333] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.896068] env[62066]: DEBUG oslo_vmware.api [None req-6f12e7db-6fdf-4bbe-9b1e-ff4283071f33 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52566e60-6421-1ba3-b275-a911841e77f8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.951975] env[62066]: DEBUG oslo_concurrency.lockutils [req-b3860259-ac6e-4483-a6cc-ef9f6f0b8dc7 req-aa15c4c9-f411-4565-b421-9ae4b26d4133 service nova] Releasing lock "refresh_cache-c8e0d47c-4421-4e00-9183-206fceeabc40" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 862.952345] env[62066]: DEBUG nova.compute.manager [req-b3860259-ac6e-4483-a6cc-ef9f6f0b8dc7 req-aa15c4c9-f411-4565-b421-9ae4b26d4133 service nova] [instance: 459f432b-e817-405f-9c3e-dfa5d957788c] Received event network-vif-deleted-56efe6b6-2fe3-4c7b-9d59-f8ecf74ac772 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 862.952523] env[62066]: INFO nova.compute.manager [req-b3860259-ac6e-4483-a6cc-ef9f6f0b8dc7 req-aa15c4c9-f411-4565-b421-9ae4b26d4133 service nova] [instance: 459f432b-e817-405f-9c3e-dfa5d957788c] Neutron deleted interface 56efe6b6-2fe3-4c7b-9d59-f8ecf74ac772; detaching it from the instance and deleting it from the info cache [ 862.952828] env[62066]: DEBUG nova.network.neutron [req-b3860259-ac6e-4483-a6cc-ef9f6f0b8dc7 req-aa15c4c9-f411-4565-b421-9ae4b26d4133 service nova] [instance: 459f432b-e817-405f-9c3e-dfa5d957788c] Updating instance_info_cache with network_info: [{"id": "cb74e58f-b481-4a0f-861b-96381e2d52dc", "address": "fa:16:3e:87:b9:2f", "network": {"id": "449dfe4e-bb49-43b6-9ba7-b57af74ebfb3", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-282029676-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.159", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0d4e33a99c7741fb8cdd97f4ec5dbbd8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db68bd64-5b56-49af-a075-13dcf85cb2e0", "external-id": "nsx-vlan-transportzone-590", "segmentation_id": 590, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcb74e58f-b4", "ovs_interfaceid": "cb74e58f-b481-4a0f-861b-96381e2d52dc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 863.079069] env[62066]: DEBUG oslo_vmware.api [None req-ba2d4aa3-3165-47fd-a67a-a02fe7b33642 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': task-1340965, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.272083} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.079363] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-ba2d4aa3-3165-47fd-a67a-a02fe7b33642 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 863.079533] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-ba2d4aa3-3165-47fd-a67a-a02fe7b33642 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: 459f432b-e817-405f-9c3e-dfa5d957788c] Deleted contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 863.079802] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-ba2d4aa3-3165-47fd-a67a-a02fe7b33642 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: 459f432b-e817-405f-9c3e-dfa5d957788c] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 863.080023] env[62066]: INFO nova.compute.manager [None req-ba2d4aa3-3165-47fd-a67a-a02fe7b33642 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: 459f432b-e817-405f-9c3e-dfa5d957788c] Took 1.12 seconds to destroy the instance on the hypervisor. [ 863.080462] env[62066]: DEBUG oslo.service.loopingcall [None req-ba2d4aa3-3165-47fd-a67a-a02fe7b33642 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 863.080462] env[62066]: DEBUG nova.compute.manager [-] [instance: 459f432b-e817-405f-9c3e-dfa5d957788c] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 863.080596] env[62066]: DEBUG nova.network.neutron [-] [instance: 459f432b-e817-405f-9c3e-dfa5d957788c] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 863.133266] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-9a34415f-ab54-458b-b6ce-0d7cde2c3efe tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 863.134068] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-9a34415f-ab54-458b-b6ce-0d7cde2c3efe tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 863.134605] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a50a4d94-b280-451c-b7ab-c545a0818e11 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.140376] env[62066]: DEBUG oslo_vmware.api [None req-9a34415f-ab54-458b-b6ce-0d7cde2c3efe tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] Waiting for the task: (returnval){ [ 863.140376] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]5230e7cb-dddd-c525-5e0b-06344348f6da" [ 863.140376] env[62066]: _type = "Task" [ 863.140376] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.144632] env[62066]: DEBUG nova.network.neutron [None req-f2684bd2-e68b-4819-84d6-b5b0e8bf10bc tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 76306c3f-0674-40fe-9864-d82d11e4bb5e] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 863.151206] env[62066]: DEBUG oslo_vmware.api [None req-9a34415f-ab54-458b-b6ce-0d7cde2c3efe tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]5230e7cb-dddd-c525-5e0b-06344348f6da, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.179334] env[62066]: DEBUG oslo_vmware.api [None req-68383c37-9698-4391-bbf2-b9ebf656dc03 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52d3bc01-8390-17cd-aafe-55db5fb6b0d2, 'name': SearchDatastore_Task, 'duration_secs': 0.013638} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.180218] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b0653213-5bed-45ee-9ce0-5930b37b35ad {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.190567] env[62066]: DEBUG oslo_vmware.api [None req-68383c37-9698-4391-bbf2-b9ebf656dc03 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Waiting for the task: (returnval){ [ 863.190567] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]52942f4c-9ba9-b5e3-d9b2-123c8c890dbe" [ 863.190567] env[62066]: _type = "Task" [ 863.190567] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.202012] env[62066]: DEBUG oslo_vmware.api [None req-68383c37-9698-4391-bbf2-b9ebf656dc03 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52942f4c-9ba9-b5e3-d9b2-123c8c890dbe, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.241382] env[62066]: DEBUG oslo_concurrency.lockutils [None req-bace0778-9dfe-4de1-8747-93c45a08d78a tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Releasing lock "refresh_cache-459f432b-e817-405f-9c3e-dfa5d957788c" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 863.243888] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cf3b688f-a077-42d2-b46f-272004213881 tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.875s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 863.246169] env[62066]: DEBUG oslo_concurrency.lockutils [None req-59107a59-1a77-47be-8dae-1ec8f16f1fda tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.202s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 863.247628] env[62066]: INFO nova.compute.claims [None req-59107a59-1a77-47be-8dae-1ec8f16f1fda tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] [instance: 2634fcc0-96bd-4513-8f92-515f83fe23d7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 863.308696] env[62066]: DEBUG nova.network.neutron [None req-f2684bd2-e68b-4819-84d6-b5b0e8bf10bc tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 76306c3f-0674-40fe-9864-d82d11e4bb5e] Updating instance_info_cache with network_info: [{"id": "28f3b624-430a-474d-a24b-9e54db92e9e2", "address": "fa:16:3e:73:9e:3c", "network": {"id": "1c6c51e4-96f8-4175-b7f5-96b8a42fcb13", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-2006089809-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d609babc987a42e2a8ddb4bfb9c3b103", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a965790c-2d2f-4c2a-9ee7-745f4d53039b", "external-id": "nsx-vlan-transportzone-708", "segmentation_id": 708, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap28f3b624-43", "ovs_interfaceid": "28f3b624-430a-474d-a24b-9e54db92e9e2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 863.399539] env[62066]: DEBUG oslo_vmware.api [None req-6f12e7db-6fdf-4bbe-9b1e-ff4283071f33 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52566e60-6421-1ba3-b275-a911841e77f8, 'name': SearchDatastore_Task, 'duration_secs': 0.015124} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.399825] env[62066]: DEBUG oslo_concurrency.lockutils [None req-6f12e7db-6fdf-4bbe-9b1e-ff4283071f33 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Releasing lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 863.400086] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-6f12e7db-6fdf-4bbe-9b1e-ff4283071f33 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] [instance: 3120c9a5-0558-4e83-b0fb-994a989398fb] Processing image 50ff584c-3b50-4395-af07-3e66769bc9f7 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 863.400308] env[62066]: DEBUG oslo_concurrency.lockutils [None req-6f12e7db-6fdf-4bbe-9b1e-ff4283071f33 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 863.455590] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5b7195a1-d294-4cf6-aedd-17e7ca13fdab {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.470034] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3742ccf8-ca25-4d59-9e14-2972315f85a7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.498711] env[62066]: DEBUG nova.compute.manager [req-b3860259-ac6e-4483-a6cc-ef9f6f0b8dc7 req-aa15c4c9-f411-4565-b421-9ae4b26d4133 service nova] [instance: 459f432b-e817-405f-9c3e-dfa5d957788c] Detach interface failed, port_id=56efe6b6-2fe3-4c7b-9d59-f8ecf74ac772, reason: Instance 459f432b-e817-405f-9c3e-dfa5d957788c could not be found. {{(pid=62066) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 863.658530] env[62066]: DEBUG oslo_vmware.api [None req-9a34415f-ab54-458b-b6ce-0d7cde2c3efe tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]5230e7cb-dddd-c525-5e0b-06344348f6da, 'name': SearchDatastore_Task, 'duration_secs': 0.01342} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.658530] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1e7efbe4-dd7d-4f4a-ac37-25e383910a4b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.664395] env[62066]: DEBUG oslo_vmware.api [None req-9a34415f-ab54-458b-b6ce-0d7cde2c3efe tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] Waiting for the task: (returnval){ [ 863.664395] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]52200e09-fe95-7d74-0b7f-ffe0d063c230" [ 863.664395] env[62066]: _type = "Task" [ 863.664395] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.677457] env[62066]: DEBUG oslo_vmware.api [None req-9a34415f-ab54-458b-b6ce-0d7cde2c3efe tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52200e09-fe95-7d74-0b7f-ffe0d063c230, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.692144] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4a3c2cb7-f1e2-437a-b493-e5de08e55742 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Acquiring lock "8ff6fbb9-c90f-498d-9a85-d220a8c2f794" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 863.692432] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4a3c2cb7-f1e2-437a-b493-e5de08e55742 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Lock "8ff6fbb9-c90f-498d-9a85-d220a8c2f794" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 863.692620] env[62066]: INFO nova.compute.manager [None req-4a3c2cb7-f1e2-437a-b493-e5de08e55742 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] [instance: 8ff6fbb9-c90f-498d-9a85-d220a8c2f794] Shelving [ 863.705451] env[62066]: DEBUG oslo_vmware.api [None req-68383c37-9698-4391-bbf2-b9ebf656dc03 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52942f4c-9ba9-b5e3-d9b2-123c8c890dbe, 'name': SearchDatastore_Task, 'duration_secs': 0.01273} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.705722] env[62066]: DEBUG oslo_concurrency.lockutils [None req-68383c37-9698-4391-bbf2-b9ebf656dc03 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Releasing lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 863.706148] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-68383c37-9698-4391-bbf2-b9ebf656dc03 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk to [datastore2] c8e0d47c-4421-4e00-9183-206fceeabc40/c8e0d47c-4421-4e00-9183-206fceeabc40.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 863.706281] env[62066]: DEBUG oslo_concurrency.lockutils [None req-6f12e7db-6fdf-4bbe-9b1e-ff4283071f33 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Acquired lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 863.706476] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-6f12e7db-6fdf-4bbe-9b1e-ff4283071f33 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 863.706688] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-36d9c7db-9496-458b-87eb-3ce780a796f2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.708749] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9f1ac9db-2444-4844-ab8e-36e0024e221e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.716877] env[62066]: DEBUG oslo_vmware.api [None req-68383c37-9698-4391-bbf2-b9ebf656dc03 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Waiting for the task: (returnval){ [ 863.716877] env[62066]: value = "task-1340967" [ 863.716877] env[62066]: _type = "Task" [ 863.716877] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.720731] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-6f12e7db-6fdf-4bbe-9b1e-ff4283071f33 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 863.720915] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-6f12e7db-6fdf-4bbe-9b1e-ff4283071f33 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 863.721979] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d94cb0e9-31d4-4667-a911-fbdff15c402d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.728530] env[62066]: DEBUG oslo_vmware.api [None req-68383c37-9698-4391-bbf2-b9ebf656dc03 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Task: {'id': task-1340967, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.731928] env[62066]: DEBUG oslo_vmware.api [None req-6f12e7db-6fdf-4bbe-9b1e-ff4283071f33 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Waiting for the task: (returnval){ [ 863.731928] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]5232bb20-221e-b680-8efa-a99819bc0d8e" [ 863.731928] env[62066]: _type = "Task" [ 863.731928] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.739459] env[62066]: DEBUG oslo_vmware.api [None req-6f12e7db-6fdf-4bbe-9b1e-ff4283071f33 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]5232bb20-221e-b680-8efa-a99819bc0d8e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.747884] env[62066]: DEBUG oslo_concurrency.lockutils [None req-bace0778-9dfe-4de1-8747-93c45a08d78a tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Lock "interface-459f432b-e817-405f-9c3e-dfa5d957788c-56efe6b6-2fe3-4c7b-9d59-f8ecf74ac772" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 9.961s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 863.767655] env[62066]: DEBUG nova.compute.manager [req-8b585a25-6e64-4ff2-9a2a-8caf552fce2b req-b19604d0-c2b0-4329-bf70-9adb2798768b service nova] [instance: 76306c3f-0674-40fe-9864-d82d11e4bb5e] Received event network-vif-plugged-28f3b624-430a-474d-a24b-9e54db92e9e2 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 863.767869] env[62066]: DEBUG oslo_concurrency.lockutils [req-8b585a25-6e64-4ff2-9a2a-8caf552fce2b req-b19604d0-c2b0-4329-bf70-9adb2798768b service nova] Acquiring lock "76306c3f-0674-40fe-9864-d82d11e4bb5e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 863.768228] env[62066]: DEBUG oslo_concurrency.lockutils [req-8b585a25-6e64-4ff2-9a2a-8caf552fce2b req-b19604d0-c2b0-4329-bf70-9adb2798768b service nova] Lock "76306c3f-0674-40fe-9864-d82d11e4bb5e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 863.768419] env[62066]: DEBUG oslo_concurrency.lockutils [req-8b585a25-6e64-4ff2-9a2a-8caf552fce2b req-b19604d0-c2b0-4329-bf70-9adb2798768b service nova] Lock "76306c3f-0674-40fe-9864-d82d11e4bb5e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 863.768598] env[62066]: DEBUG nova.compute.manager [req-8b585a25-6e64-4ff2-9a2a-8caf552fce2b req-b19604d0-c2b0-4329-bf70-9adb2798768b service nova] [instance: 76306c3f-0674-40fe-9864-d82d11e4bb5e] No waiting events found dispatching network-vif-plugged-28f3b624-430a-474d-a24b-9e54db92e9e2 {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 863.768842] env[62066]: WARNING nova.compute.manager [req-8b585a25-6e64-4ff2-9a2a-8caf552fce2b req-b19604d0-c2b0-4329-bf70-9adb2798768b service nova] [instance: 76306c3f-0674-40fe-9864-d82d11e4bb5e] Received unexpected event network-vif-plugged-28f3b624-430a-474d-a24b-9e54db92e9e2 for instance with vm_state building and task_state spawning. [ 863.769039] env[62066]: DEBUG nova.compute.manager [req-8b585a25-6e64-4ff2-9a2a-8caf552fce2b req-b19604d0-c2b0-4329-bf70-9adb2798768b service nova] [instance: 76306c3f-0674-40fe-9864-d82d11e4bb5e] Received event network-changed-28f3b624-430a-474d-a24b-9e54db92e9e2 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 863.769199] env[62066]: DEBUG nova.compute.manager [req-8b585a25-6e64-4ff2-9a2a-8caf552fce2b req-b19604d0-c2b0-4329-bf70-9adb2798768b service nova] [instance: 76306c3f-0674-40fe-9864-d82d11e4bb5e] Refreshing instance network info cache due to event network-changed-28f3b624-430a-474d-a24b-9e54db92e9e2. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 863.770340] env[62066]: DEBUG oslo_concurrency.lockutils [req-8b585a25-6e64-4ff2-9a2a-8caf552fce2b req-b19604d0-c2b0-4329-bf70-9adb2798768b service nova] Acquiring lock "refresh_cache-76306c3f-0674-40fe-9864-d82d11e4bb5e" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 863.772470] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cf3b688f-a077-42d2-b46f-272004213881 tempest-ServerActionsV293TestJSON-1564821632 tempest-ServerActionsV293TestJSON-1564821632-project-member] Lock "daffaf51-4c45-44aa-8fc2-4db066a09971" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 18.812s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 863.811418] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f2684bd2-e68b-4819-84d6-b5b0e8bf10bc tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Releasing lock "refresh_cache-76306c3f-0674-40fe-9864-d82d11e4bb5e" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 863.811776] env[62066]: DEBUG nova.compute.manager [None req-f2684bd2-e68b-4819-84d6-b5b0e8bf10bc tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 76306c3f-0674-40fe-9864-d82d11e4bb5e] Instance network_info: |[{"id": "28f3b624-430a-474d-a24b-9e54db92e9e2", "address": "fa:16:3e:73:9e:3c", "network": {"id": "1c6c51e4-96f8-4175-b7f5-96b8a42fcb13", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-2006089809-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d609babc987a42e2a8ddb4bfb9c3b103", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a965790c-2d2f-4c2a-9ee7-745f4d53039b", "external-id": "nsx-vlan-transportzone-708", "segmentation_id": 708, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap28f3b624-43", "ovs_interfaceid": "28f3b624-430a-474d-a24b-9e54db92e9e2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 863.813129] env[62066]: DEBUG oslo_concurrency.lockutils [req-8b585a25-6e64-4ff2-9a2a-8caf552fce2b req-b19604d0-c2b0-4329-bf70-9adb2798768b service nova] Acquired lock "refresh_cache-76306c3f-0674-40fe-9864-d82d11e4bb5e" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 863.813334] env[62066]: DEBUG nova.network.neutron [req-8b585a25-6e64-4ff2-9a2a-8caf552fce2b req-b19604d0-c2b0-4329-bf70-9adb2798768b service nova] [instance: 76306c3f-0674-40fe-9864-d82d11e4bb5e] Refreshing network info cache for port 28f3b624-430a-474d-a24b-9e54db92e9e2 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 863.816435] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-f2684bd2-e68b-4819-84d6-b5b0e8bf10bc tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 76306c3f-0674-40fe-9864-d82d11e4bb5e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:73:9e:3c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a965790c-2d2f-4c2a-9ee7-745f4d53039b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '28f3b624-430a-474d-a24b-9e54db92e9e2', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 863.829152] env[62066]: DEBUG oslo.service.loopingcall [None req-f2684bd2-e68b-4819-84d6-b5b0e8bf10bc tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 863.830294] env[62066]: DEBUG nova.compute.manager [req-49e20b8d-dd3a-406d-a504-fc835b8981c7 req-bab8e260-f8a7-4181-b3cd-2a5f5765faf9 service nova] [instance: 459f432b-e817-405f-9c3e-dfa5d957788c] Received event network-vif-deleted-cb74e58f-b481-4a0f-861b-96381e2d52dc {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 863.830568] env[62066]: INFO nova.compute.manager [req-49e20b8d-dd3a-406d-a504-fc835b8981c7 req-bab8e260-f8a7-4181-b3cd-2a5f5765faf9 service nova] [instance: 459f432b-e817-405f-9c3e-dfa5d957788c] Neutron deleted interface cb74e58f-b481-4a0f-861b-96381e2d52dc; detaching it from the instance and deleting it from the info cache [ 863.830675] env[62066]: DEBUG nova.network.neutron [req-49e20b8d-dd3a-406d-a504-fc835b8981c7 req-bab8e260-f8a7-4181-b3cd-2a5f5765faf9 service nova] [instance: 459f432b-e817-405f-9c3e-dfa5d957788c] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 863.835824] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 76306c3f-0674-40fe-9864-d82d11e4bb5e] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 863.836172] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2f86d656-4c5c-441e-8be9-9a69b8c1ef91 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.859541] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 863.859541] env[62066]: value = "task-1340968" [ 863.859541] env[62066]: _type = "Task" [ 863.859541] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.868544] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1340968, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.894122] env[62066]: DEBUG oslo_vmware.rw_handles [None req-5b85ae79-b544-4bc9-a850-092bba375a6b tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52052324-452e-7535-c568-8c4544842db1/disk-0.vmdk. {{(pid=62066) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 863.895706] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-082c63bf-351f-4ab2-8a93-503bb63f33b0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.902631] env[62066]: DEBUG oslo_vmware.rw_handles [None req-5b85ae79-b544-4bc9-a850-092bba375a6b tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52052324-452e-7535-c568-8c4544842db1/disk-0.vmdk is in state: ready. {{(pid=62066) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 863.902925] env[62066]: ERROR oslo_vmware.rw_handles [None req-5b85ae79-b544-4bc9-a850-092bba375a6b tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52052324-452e-7535-c568-8c4544842db1/disk-0.vmdk due to incomplete transfer. [ 863.903226] env[62066]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-6c1f2a73-dd8e-4f25-bd95-72099f966a66 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.914032] env[62066]: DEBUG oslo_vmware.rw_handles [None req-5b85ae79-b544-4bc9-a850-092bba375a6b tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52052324-452e-7535-c568-8c4544842db1/disk-0.vmdk. {{(pid=62066) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 863.914032] env[62066]: DEBUG nova.virt.vmwareapi.images [None req-5b85ae79-b544-4bc9-a850-092bba375a6b tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: b08f27b1-429e-4d05-9a95-f45fd0cb2904] Uploaded image 1147d77f-02f3-4d4b-a853-4a0e7d8b0a96 to the Glance image server {{(pid=62066) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 863.916711] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-5b85ae79-b544-4bc9-a850-092bba375a6b tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: b08f27b1-429e-4d05-9a95-f45fd0cb2904] Destroying the VM {{(pid=62066) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 863.917061] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-297e3977-22bd-4efd-bb2f-1c97a8b14324 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.924115] env[62066]: DEBUG oslo_vmware.api [None req-5b85ae79-b544-4bc9-a850-092bba375a6b tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Waiting for the task: (returnval){ [ 863.924115] env[62066]: value = "task-1340969" [ 863.924115] env[62066]: _type = "Task" [ 863.924115] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.935701] env[62066]: DEBUG oslo_vmware.api [None req-5b85ae79-b544-4bc9-a850-092bba375a6b tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': task-1340969, 'name': Destroy_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.175499] env[62066]: DEBUG oslo_vmware.api [None req-9a34415f-ab54-458b-b6ce-0d7cde2c3efe tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52200e09-fe95-7d74-0b7f-ffe0d063c230, 'name': SearchDatastore_Task, 'duration_secs': 0.062748} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.175862] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9a34415f-ab54-458b-b6ce-0d7cde2c3efe tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] Releasing lock "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 864.176090] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-9a34415f-ab54-458b-b6ce-0d7cde2c3efe tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] Copying virtual disk from [datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk to [datastore1] 0bc31a8b-a5fc-4f6e-a4c0-024bf4260755/50ff584c-3b50-4395-af07-3e66769bc9f7-rescue.vmdk. {{(pid=62066) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 864.176385] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ac7dad4c-3ac3-4f6c-9e87-b20e8776cadb {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.185092] env[62066]: DEBUG oslo_vmware.api [None req-9a34415f-ab54-458b-b6ce-0d7cde2c3efe tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] Waiting for the task: (returnval){ [ 864.185092] env[62066]: value = "task-1340970" [ 864.185092] env[62066]: _type = "Task" [ 864.185092] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.193853] env[62066]: DEBUG oslo_vmware.api [None req-9a34415f-ab54-458b-b6ce-0d7cde2c3efe tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] Task: {'id': task-1340970, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.203820] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-4a3c2cb7-f1e2-437a-b493-e5de08e55742 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] [instance: 8ff6fbb9-c90f-498d-9a85-d220a8c2f794] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 864.204106] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4a07fdfb-6cfe-496d-80be-4cbec5ac4b97 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.211338] env[62066]: DEBUG oslo_vmware.api [None req-4a3c2cb7-f1e2-437a-b493-e5de08e55742 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Waiting for the task: (returnval){ [ 864.211338] env[62066]: value = "task-1340971" [ 864.211338] env[62066]: _type = "Task" [ 864.211338] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.224307] env[62066]: DEBUG oslo_vmware.api [None req-4a3c2cb7-f1e2-437a-b493-e5de08e55742 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Task: {'id': task-1340971, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.229907] env[62066]: DEBUG oslo_vmware.api [None req-68383c37-9698-4391-bbf2-b9ebf656dc03 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Task: {'id': task-1340967, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.502868} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.230246] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-68383c37-9698-4391-bbf2-b9ebf656dc03 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk to [datastore2] c8e0d47c-4421-4e00-9183-206fceeabc40/c8e0d47c-4421-4e00-9183-206fceeabc40.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 864.230480] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-68383c37-9698-4391-bbf2-b9ebf656dc03 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: c8e0d47c-4421-4e00-9183-206fceeabc40] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 864.230764] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b99c0299-bbd8-4a2e-93bf-edd86c3cb216 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.243888] env[62066]: DEBUG nova.network.neutron [-] [instance: 459f432b-e817-405f-9c3e-dfa5d957788c] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 864.245511] env[62066]: DEBUG oslo_vmware.api [None req-6f12e7db-6fdf-4bbe-9b1e-ff4283071f33 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]5232bb20-221e-b680-8efa-a99819bc0d8e, 'name': SearchDatastore_Task, 'duration_secs': 0.014029} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.247902] env[62066]: DEBUG oslo_vmware.api [None req-68383c37-9698-4391-bbf2-b9ebf656dc03 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Waiting for the task: (returnval){ [ 864.247902] env[62066]: value = "task-1340972" [ 864.247902] env[62066]: _type = "Task" [ 864.247902] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.248209] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fd37978c-d14f-4503-8e12-623fd3f42b50 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.263320] env[62066]: DEBUG oslo_vmware.api [None req-68383c37-9698-4391-bbf2-b9ebf656dc03 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Task: {'id': task-1340972, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.263765] env[62066]: DEBUG oslo_vmware.api [None req-6f12e7db-6fdf-4bbe-9b1e-ff4283071f33 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Waiting for the task: (returnval){ [ 864.263765] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]52ec5e74-6824-a5d9-a7e4-fe34395107e6" [ 864.263765] env[62066]: _type = "Task" [ 864.263765] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.275394] env[62066]: DEBUG oslo_vmware.api [None req-6f12e7db-6fdf-4bbe-9b1e-ff4283071f33 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52ec5e74-6824-a5d9-a7e4-fe34395107e6, 'name': SearchDatastore_Task, 'duration_secs': 0.013264} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.275708] env[62066]: DEBUG oslo_concurrency.lockutils [None req-6f12e7db-6fdf-4bbe-9b1e-ff4283071f33 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Releasing lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 864.276038] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-6f12e7db-6fdf-4bbe-9b1e-ff4283071f33 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk to [datastore2] 3120c9a5-0558-4e83-b0fb-994a989398fb/3120c9a5-0558-4e83-b0fb-994a989398fb.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 864.276353] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-489e65be-0711-4176-a4ab-0c30846ca4b8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.288105] env[62066]: DEBUG oslo_vmware.api [None req-6f12e7db-6fdf-4bbe-9b1e-ff4283071f33 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Waiting for the task: (returnval){ [ 864.288105] env[62066]: value = "task-1340973" [ 864.288105] env[62066]: _type = "Task" [ 864.288105] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.295385] env[62066]: DEBUG oslo_vmware.api [None req-6f12e7db-6fdf-4bbe-9b1e-ff4283071f33 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Task: {'id': task-1340973, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.340256] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7ee072a8-ee87-42d7-b1b2-6a3ed8988769 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.348325] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9042e1a-5d5c-473b-8a12-70a06350a114 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.378641] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1340968, 'name': CreateVM_Task} progress is 99%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.393990] env[62066]: DEBUG nova.compute.manager [req-49e20b8d-dd3a-406d-a504-fc835b8981c7 req-bab8e260-f8a7-4181-b3cd-2a5f5765faf9 service nova] [instance: 459f432b-e817-405f-9c3e-dfa5d957788c] Detach interface failed, port_id=cb74e58f-b481-4a0f-861b-96381e2d52dc, reason: Instance 459f432b-e817-405f-9c3e-dfa5d957788c could not be found. {{(pid=62066) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 864.433786] env[62066]: DEBUG oslo_vmware.api [None req-5b85ae79-b544-4bc9-a850-092bba375a6b tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': task-1340969, 'name': Destroy_Task} progress is 33%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.548899] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c27655c6-b072-4de6-bad5-a57f69c9eb88 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.556551] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0dba5c7b-e8b5-4a4f-b024-cdb76bcf484c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.586016] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8345e108-2b2f-4138-ba17-e9ebc19be878 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.593859] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d240b6e7-ec67-47ae-8bdb-e7b1d611a495 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.606980] env[62066]: DEBUG nova.compute.provider_tree [None req-59107a59-1a77-47be-8dae-1ec8f16f1fda tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 864.698915] env[62066]: DEBUG oslo_vmware.api [None req-9a34415f-ab54-458b-b6ce-0d7cde2c3efe tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] Task: {'id': task-1340970, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.723394] env[62066]: DEBUG oslo_vmware.api [None req-4a3c2cb7-f1e2-437a-b493-e5de08e55742 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Task: {'id': task-1340971, 'name': PowerOffVM_Task, 'duration_secs': 0.176976} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.725924] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-4a3c2cb7-f1e2-437a-b493-e5de08e55742 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] [instance: 8ff6fbb9-c90f-498d-9a85-d220a8c2f794] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 864.726883] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fbbb2b1-8bb2-4ca5-94af-2adabfec1ca4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.747590] env[62066]: INFO nova.compute.manager [-] [instance: 459f432b-e817-405f-9c3e-dfa5d957788c] Took 1.67 seconds to deallocate network for instance. [ 864.748482] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8546320e-cdf4-4958-b7d8-7b061a8fd17b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.771153] env[62066]: DEBUG oslo_vmware.api [None req-68383c37-9698-4391-bbf2-b9ebf656dc03 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Task: {'id': task-1340972, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066527} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.771755] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-68383c37-9698-4391-bbf2-b9ebf656dc03 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: c8e0d47c-4421-4e00-9183-206fceeabc40] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 864.772692] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef91b292-bba4-44b7-a4c8-c67487707a67 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.798697] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-68383c37-9698-4391-bbf2-b9ebf656dc03 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: c8e0d47c-4421-4e00-9183-206fceeabc40] Reconfiguring VM instance instance-0000004f to attach disk [datastore2] c8e0d47c-4421-4e00-9183-206fceeabc40/c8e0d47c-4421-4e00-9183-206fceeabc40.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 864.807303] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-42e98930-4953-4f07-b8f7-99de469c30d6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.829958] env[62066]: DEBUG oslo_vmware.api [None req-6f12e7db-6fdf-4bbe-9b1e-ff4283071f33 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Task: {'id': task-1340973, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.831802] env[62066]: DEBUG oslo_vmware.api [None req-68383c37-9698-4391-bbf2-b9ebf656dc03 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Waiting for the task: (returnval){ [ 864.831802] env[62066]: value = "task-1340974" [ 864.831802] env[62066]: _type = "Task" [ 864.831802] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.841664] env[62066]: DEBUG oslo_vmware.api [None req-68383c37-9698-4391-bbf2-b9ebf656dc03 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Task: {'id': task-1340974, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.879321] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1340968, 'name': CreateVM_Task, 'duration_secs': 0.531477} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.879539] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 76306c3f-0674-40fe-9864-d82d11e4bb5e] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 864.880326] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f2684bd2-e68b-4819-84d6-b5b0e8bf10bc tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 864.880547] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f2684bd2-e68b-4819-84d6-b5b0e8bf10bc tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Acquired lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 864.880956] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f2684bd2-e68b-4819-84d6-b5b0e8bf10bc tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 864.881267] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-295aa897-2e38-4881-8e1b-db0d97f534a9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.887145] env[62066]: DEBUG oslo_vmware.api [None req-f2684bd2-e68b-4819-84d6-b5b0e8bf10bc tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Waiting for the task: (returnval){ [ 864.887145] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]52215500-c64d-c583-2e42-8bc8f5b090a6" [ 864.887145] env[62066]: _type = "Task" [ 864.887145] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.899358] env[62066]: DEBUG oslo_vmware.api [None req-f2684bd2-e68b-4819-84d6-b5b0e8bf10bc tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52215500-c64d-c583-2e42-8bc8f5b090a6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.935751] env[62066]: DEBUG oslo_vmware.api [None req-5b85ae79-b544-4bc9-a850-092bba375a6b tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': task-1340969, 'name': Destroy_Task, 'duration_secs': 0.87728} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.936235] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-5b85ae79-b544-4bc9-a850-092bba375a6b tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: b08f27b1-429e-4d05-9a95-f45fd0cb2904] Destroyed the VM [ 864.937759] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-5b85ae79-b544-4bc9-a850-092bba375a6b tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: b08f27b1-429e-4d05-9a95-f45fd0cb2904] Deleting Snapshot of the VM instance {{(pid=62066) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 864.937759] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-9c344ba7-b95d-49d0-9461-b2ed6f24927e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.940123] env[62066]: DEBUG nova.network.neutron [req-8b585a25-6e64-4ff2-9a2a-8caf552fce2b req-b19604d0-c2b0-4329-bf70-9adb2798768b service nova] [instance: 76306c3f-0674-40fe-9864-d82d11e4bb5e] Updated VIF entry in instance network info cache for port 28f3b624-430a-474d-a24b-9e54db92e9e2. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 864.940625] env[62066]: DEBUG nova.network.neutron [req-8b585a25-6e64-4ff2-9a2a-8caf552fce2b req-b19604d0-c2b0-4329-bf70-9adb2798768b service nova] [instance: 76306c3f-0674-40fe-9864-d82d11e4bb5e] Updating instance_info_cache with network_info: [{"id": "28f3b624-430a-474d-a24b-9e54db92e9e2", "address": "fa:16:3e:73:9e:3c", "network": {"id": "1c6c51e4-96f8-4175-b7f5-96b8a42fcb13", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-2006089809-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d609babc987a42e2a8ddb4bfb9c3b103", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a965790c-2d2f-4c2a-9ee7-745f4d53039b", "external-id": "nsx-vlan-transportzone-708", "segmentation_id": 708, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap28f3b624-43", "ovs_interfaceid": "28f3b624-430a-474d-a24b-9e54db92e9e2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 864.947437] env[62066]: DEBUG oslo_vmware.api [None req-5b85ae79-b544-4bc9-a850-092bba375a6b tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Waiting for the task: (returnval){ [ 864.947437] env[62066]: value = "task-1340975" [ 864.947437] env[62066]: _type = "Task" [ 864.947437] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.958128] env[62066]: DEBUG oslo_vmware.api [None req-5b85ae79-b544-4bc9-a850-092bba375a6b tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': task-1340975, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.112509] env[62066]: DEBUG nova.scheduler.client.report [None req-59107a59-1a77-47be-8dae-1ec8f16f1fda tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 865.197841] env[62066]: DEBUG oslo_vmware.api [None req-9a34415f-ab54-458b-b6ce-0d7cde2c3efe tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] Task: {'id': task-1340970, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.986591} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.198068] env[62066]: INFO nova.virt.vmwareapi.ds_util [None req-9a34415f-ab54-458b-b6ce-0d7cde2c3efe tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] Copied virtual disk from [datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk to [datastore1] 0bc31a8b-a5fc-4f6e-a4c0-024bf4260755/50ff584c-3b50-4395-af07-3e66769bc9f7-rescue.vmdk. [ 865.198889] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-053b1a1a-cb90-4eff-8db0-d8c4c50c70d1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.909791] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ba2d4aa3-3165-47fd-a67a-a02fe7b33642 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 865.910874] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-4a3c2cb7-f1e2-437a-b493-e5de08e55742 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] [instance: 8ff6fbb9-c90f-498d-9a85-d220a8c2f794] Creating Snapshot of the VM instance {{(pid=62066) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 865.911411] env[62066]: DEBUG oslo_concurrency.lockutils [req-8b585a25-6e64-4ff2-9a2a-8caf552fce2b req-b19604d0-c2b0-4329-bf70-9adb2798768b service nova] Releasing lock "refresh_cache-76306c3f-0674-40fe-9864-d82d11e4bb5e" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 865.912194] env[62066]: DEBUG oslo_concurrency.lockutils [None req-59107a59-1a77-47be-8dae-1ec8f16f1fda tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.666s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 865.912635] env[62066]: DEBUG nova.compute.manager [None req-59107a59-1a77-47be-8dae-1ec8f16f1fda tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] [instance: 2634fcc0-96bd-4513-8f92-515f83fe23d7] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 865.922255] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-9a34415f-ab54-458b-b6ce-0d7cde2c3efe tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] [instance: 0bc31a8b-a5fc-4f6e-a4c0-024bf4260755] Reconfiguring VM instance instance-0000004e to attach disk [datastore1] 0bc31a8b-a5fc-4f6e-a4c0-024bf4260755/50ff584c-3b50-4395-af07-3e66769bc9f7-rescue.vmdk or device None with type thin {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 865.922971] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-b7edfcd5-2aeb-4362-8621-2057f4036c94 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.936210] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4c61eb51-497c-4460-951e-d696b73eff28 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.589s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 865.937887] env[62066]: INFO nova.compute.claims [None req-4c61eb51-497c-4460-951e-d696b73eff28 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 170e184a-f009-42cf-82dc-eb9696a00a24] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 865.940673] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ef59754a-db1d-4637-82b4-8dff80f183cd {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.974107] env[62066]: DEBUG oslo_vmware.api [None req-6f12e7db-6fdf-4bbe-9b1e-ff4283071f33 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Task: {'id': task-1340973, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.895965} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.974107] env[62066]: DEBUG oslo_vmware.api [None req-f2684bd2-e68b-4819-84d6-b5b0e8bf10bc tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52215500-c64d-c583-2e42-8bc8f5b090a6, 'name': SearchDatastore_Task, 'duration_secs': 0.070347} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.974107] env[62066]: DEBUG oslo_vmware.api [None req-68383c37-9698-4391-bbf2-b9ebf656dc03 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Task: {'id': task-1340974, 'name': ReconfigVM_Task, 'duration_secs': 0.394687} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.974302] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-6f12e7db-6fdf-4bbe-9b1e-ff4283071f33 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk to [datastore2] 3120c9a5-0558-4e83-b0fb-994a989398fb/3120c9a5-0558-4e83-b0fb-994a989398fb.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 865.974608] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-6f12e7db-6fdf-4bbe-9b1e-ff4283071f33 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] [instance: 3120c9a5-0558-4e83-b0fb-994a989398fb] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 865.974929] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f2684bd2-e68b-4819-84d6-b5b0e8bf10bc tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Releasing lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 865.975184] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-f2684bd2-e68b-4819-84d6-b5b0e8bf10bc tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 76306c3f-0674-40fe-9864-d82d11e4bb5e] Processing image 50ff584c-3b50-4395-af07-3e66769bc9f7 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 865.975420] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f2684bd2-e68b-4819-84d6-b5b0e8bf10bc tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 865.975559] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f2684bd2-e68b-4819-84d6-b5b0e8bf10bc tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Acquired lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 865.975732] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-f2684bd2-e68b-4819-84d6-b5b0e8bf10bc tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 865.976049] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-68383c37-9698-4391-bbf2-b9ebf656dc03 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: c8e0d47c-4421-4e00-9183-206fceeabc40] Reconfigured VM instance instance-0000004f to attach disk [datastore2] c8e0d47c-4421-4e00-9183-206fceeabc40/c8e0d47c-4421-4e00-9183-206fceeabc40.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 865.976713] env[62066]: DEBUG oslo_vmware.api [None req-4a3c2cb7-f1e2-437a-b493-e5de08e55742 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Waiting for the task: (returnval){ [ 865.976713] env[62066]: value = "task-1340976" [ 865.976713] env[62066]: _type = "Task" [ 865.976713] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.980843] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-85139aab-0c02-4cec-bb90-25eedbca27d2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.982739] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d7bfb6ec-314d-4090-997b-a356e3acad3c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.985479] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6162d28b-9cc6-4ac4-bec1-9b8e6982614f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.986261] env[62066]: DEBUG oslo_vmware.api [None req-9a34415f-ab54-458b-b6ce-0d7cde2c3efe tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] Waiting for the task: (returnval){ [ 865.986261] env[62066]: value = "task-1340977" [ 865.986261] env[62066]: _type = "Task" [ 865.986261] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.986853] env[62066]: DEBUG oslo_vmware.api [None req-5b85ae79-b544-4bc9-a850-092bba375a6b tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': task-1340975, 'name': RemoveSnapshot_Task, 'duration_secs': 0.993262} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.989855] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-5b85ae79-b544-4bc9-a850-092bba375a6b tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: b08f27b1-429e-4d05-9a95-f45fd0cb2904] Deleted Snapshot of the VM instance {{(pid=62066) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 865.990130] env[62066]: INFO nova.compute.manager [None req-5b85ae79-b544-4bc9-a850-092bba375a6b tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: b08f27b1-429e-4d05-9a95-f45fd0cb2904] Took 13.82 seconds to snapshot the instance on the hypervisor. [ 865.999919] env[62066]: DEBUG oslo_vmware.api [None req-6f12e7db-6fdf-4bbe-9b1e-ff4283071f33 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Waiting for the task: (returnval){ [ 865.999919] env[62066]: value = "task-1340978" [ 865.999919] env[62066]: _type = "Task" [ 865.999919] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.004222] env[62066]: DEBUG oslo_vmware.api [None req-68383c37-9698-4391-bbf2-b9ebf656dc03 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Waiting for the task: (returnval){ [ 866.004222] env[62066]: value = "task-1340979" [ 866.004222] env[62066]: _type = "Task" [ 866.004222] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.004222] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-f2684bd2-e68b-4819-84d6-b5b0e8bf10bc tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 866.004367] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-f2684bd2-e68b-4819-84d6-b5b0e8bf10bc tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 866.004967] env[62066]: DEBUG oslo_vmware.api [None req-4a3c2cb7-f1e2-437a-b493-e5de08e55742 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Task: {'id': task-1340976, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.013746] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fb97f113-0661-4675-945f-6e86cb8cefe4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.016433] env[62066]: DEBUG oslo_vmware.api [None req-9a34415f-ab54-458b-b6ce-0d7cde2c3efe tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] Task: {'id': task-1340977, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.022217] env[62066]: DEBUG oslo_vmware.api [None req-68383c37-9698-4391-bbf2-b9ebf656dc03 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Task: {'id': task-1340979, 'name': Rename_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.026010] env[62066]: DEBUG oslo_vmware.api [None req-f2684bd2-e68b-4819-84d6-b5b0e8bf10bc tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Waiting for the task: (returnval){ [ 866.026010] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]520b00d9-962a-2980-cfe0-c8634b109be8" [ 866.026010] env[62066]: _type = "Task" [ 866.026010] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.026236] env[62066]: DEBUG oslo_vmware.api [None req-6f12e7db-6fdf-4bbe-9b1e-ff4283071f33 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Task: {'id': task-1340978, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.035176] env[62066]: DEBUG oslo_vmware.api [None req-f2684bd2-e68b-4819-84d6-b5b0e8bf10bc tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]520b00d9-962a-2980-cfe0-c8634b109be8, 'name': SearchDatastore_Task, 'duration_secs': 0.008544} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.035763] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-acac4287-ace8-43b1-9114-7459c9f6e68e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.040550] env[62066]: DEBUG oslo_vmware.api [None req-f2684bd2-e68b-4819-84d6-b5b0e8bf10bc tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Waiting for the task: (returnval){ [ 866.040550] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]52979e26-a7b3-215a-e160-00108b2473b8" [ 866.040550] env[62066]: _type = "Task" [ 866.040550] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.049731] env[62066]: DEBUG oslo_vmware.api [None req-f2684bd2-e68b-4819-84d6-b5b0e8bf10bc tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52979e26-a7b3-215a-e160-00108b2473b8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.056678] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f0861cc1-6af6-4429-bc89-cb4b7f228490 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Acquiring lock "8807ce4d-532f-469c-a302-464c61c7efeb" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 866.056907] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f0861cc1-6af6-4429-bc89-cb4b7f228490 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Lock "8807ce4d-532f-469c-a302-464c61c7efeb" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 866.425023] env[62066]: DEBUG nova.compute.utils [None req-59107a59-1a77-47be-8dae-1ec8f16f1fda tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 866.426857] env[62066]: DEBUG nova.compute.manager [None req-59107a59-1a77-47be-8dae-1ec8f16f1fda tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] [instance: 2634fcc0-96bd-4513-8f92-515f83fe23d7] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 866.427101] env[62066]: DEBUG nova.network.neutron [None req-59107a59-1a77-47be-8dae-1ec8f16f1fda tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] [instance: 2634fcc0-96bd-4513-8f92-515f83fe23d7] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 866.484829] env[62066]: DEBUG nova.policy [None req-59107a59-1a77-47be-8dae-1ec8f16f1fda tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'dff959d830ce40869f66802c17b2572d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '59855b1cf91d43e3ba0482f414c4e546', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 866.507792] env[62066]: DEBUG oslo_vmware.api [None req-4a3c2cb7-f1e2-437a-b493-e5de08e55742 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Task: {'id': task-1340976, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.508554] env[62066]: DEBUG oslo_vmware.api [None req-9a34415f-ab54-458b-b6ce-0d7cde2c3efe tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] Task: {'id': task-1340977, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.523201] env[62066]: DEBUG oslo_vmware.api [None req-6f12e7db-6fdf-4bbe-9b1e-ff4283071f33 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Task: {'id': task-1340978, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070396} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.527937] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-6f12e7db-6fdf-4bbe-9b1e-ff4283071f33 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] [instance: 3120c9a5-0558-4e83-b0fb-994a989398fb] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 866.528603] env[62066]: DEBUG oslo_vmware.api [None req-68383c37-9698-4391-bbf2-b9ebf656dc03 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Task: {'id': task-1340979, 'name': Rename_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.529569] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61c70d31-71a5-4a06-b2e5-0633abb4212a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.558456] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-6f12e7db-6fdf-4bbe-9b1e-ff4283071f33 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] [instance: 3120c9a5-0558-4e83-b0fb-994a989398fb] Reconfiguring VM instance instance-0000004d to attach disk [datastore2] 3120c9a5-0558-4e83-b0fb-994a989398fb/3120c9a5-0558-4e83-b0fb-994a989398fb.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 866.563060] env[62066]: DEBUG nova.compute.manager [None req-f0861cc1-6af6-4429-bc89-cb4b7f228490 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 8807ce4d-532f-469c-a302-464c61c7efeb] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 866.566246] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6ff30e7c-df57-4690-89e0-7edf99fb6c20 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.595897] env[62066]: DEBUG oslo_vmware.api [None req-f2684bd2-e68b-4819-84d6-b5b0e8bf10bc tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52979e26-a7b3-215a-e160-00108b2473b8, 'name': SearchDatastore_Task, 'duration_secs': 0.00829} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.597386] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f2684bd2-e68b-4819-84d6-b5b0e8bf10bc tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Releasing lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 866.597493] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-f2684bd2-e68b-4819-84d6-b5b0e8bf10bc tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk to [datastore2] 76306c3f-0674-40fe-9864-d82d11e4bb5e/76306c3f-0674-40fe-9864-d82d11e4bb5e.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 866.601017] env[62066]: DEBUG oslo_vmware.api [None req-6f12e7db-6fdf-4bbe-9b1e-ff4283071f33 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Waiting for the task: (returnval){ [ 866.601017] env[62066]: value = "task-1340980" [ 866.601017] env[62066]: _type = "Task" [ 866.601017] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.601017] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a2b973f8-e7fb-466d-a8ce-fa603117cab4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.608088] env[62066]: DEBUG oslo_vmware.api [None req-6f12e7db-6fdf-4bbe-9b1e-ff4283071f33 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Task: {'id': task-1340980, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.609222] env[62066]: DEBUG oslo_vmware.api [None req-f2684bd2-e68b-4819-84d6-b5b0e8bf10bc tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Waiting for the task: (returnval){ [ 866.609222] env[62066]: value = "task-1340981" [ 866.609222] env[62066]: _type = "Task" [ 866.609222] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.617351] env[62066]: DEBUG oslo_vmware.api [None req-f2684bd2-e68b-4819-84d6-b5b0e8bf10bc tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1340981, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.742569] env[62066]: DEBUG nova.network.neutron [None req-59107a59-1a77-47be-8dae-1ec8f16f1fda tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] [instance: 2634fcc0-96bd-4513-8f92-515f83fe23d7] Successfully created port: cf42ffe4-6621-4c29-9a6d-687ade676346 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 866.931017] env[62066]: DEBUG nova.compute.manager [None req-59107a59-1a77-47be-8dae-1ec8f16f1fda tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] [instance: 2634fcc0-96bd-4513-8f92-515f83fe23d7] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 867.001805] env[62066]: DEBUG oslo_vmware.api [None req-4a3c2cb7-f1e2-437a-b493-e5de08e55742 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Task: {'id': task-1340976, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.010383] env[62066]: DEBUG oslo_vmware.api [None req-9a34415f-ab54-458b-b6ce-0d7cde2c3efe tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] Task: {'id': task-1340977, 'name': ReconfigVM_Task, 'duration_secs': 0.97203} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.014473] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-9a34415f-ab54-458b-b6ce-0d7cde2c3efe tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] [instance: 0bc31a8b-a5fc-4f6e-a4c0-024bf4260755] Reconfigured VM instance instance-0000004e to attach disk [datastore1] 0bc31a8b-a5fc-4f6e-a4c0-024bf4260755/50ff584c-3b50-4395-af07-3e66769bc9f7-rescue.vmdk or device None with type thin {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 867.014827] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0080b0b8-307e-4f95-81f6-ae5d425183d8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.030486] env[62066]: DEBUG oslo_vmware.api [None req-68383c37-9698-4391-bbf2-b9ebf656dc03 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Task: {'id': task-1340979, 'name': Rename_Task, 'duration_secs': 0.775353} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.050164] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-68383c37-9698-4391-bbf2-b9ebf656dc03 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: c8e0d47c-4421-4e00-9183-206fceeabc40] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 867.058091] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9dfb937c-e0cd-44ed-9349-efc520cdd692 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.059909] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2958d8a0-aa5e-4f47-938a-1db9d3ed2c10 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.078912] env[62066]: DEBUG oslo_vmware.api [None req-9a34415f-ab54-458b-b6ce-0d7cde2c3efe tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] Waiting for the task: (returnval){ [ 867.078912] env[62066]: value = "task-1340983" [ 867.078912] env[62066]: _type = "Task" [ 867.078912] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.081088] env[62066]: DEBUG oslo_vmware.api [None req-68383c37-9698-4391-bbf2-b9ebf656dc03 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Waiting for the task: (returnval){ [ 867.081088] env[62066]: value = "task-1340982" [ 867.081088] env[62066]: _type = "Task" [ 867.081088] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.096262] env[62066]: DEBUG oslo_vmware.api [None req-9a34415f-ab54-458b-b6ce-0d7cde2c3efe tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] Task: {'id': task-1340983, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.102823] env[62066]: DEBUG oslo_vmware.api [None req-68383c37-9698-4391-bbf2-b9ebf656dc03 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Task: {'id': task-1340982, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.113346] env[62066]: DEBUG oslo_vmware.api [None req-6f12e7db-6fdf-4bbe-9b1e-ff4283071f33 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Task: {'id': task-1340980, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.114646] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f0861cc1-6af6-4429-bc89-cb4b7f228490 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 867.125589] env[62066]: DEBUG oslo_vmware.api [None req-f2684bd2-e68b-4819-84d6-b5b0e8bf10bc tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1340981, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.494934} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.125949] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-f2684bd2-e68b-4819-84d6-b5b0e8bf10bc tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk to [datastore2] 76306c3f-0674-40fe-9864-d82d11e4bb5e/76306c3f-0674-40fe-9864-d82d11e4bb5e.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 867.126219] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-f2684bd2-e68b-4819-84d6-b5b0e8bf10bc tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 76306c3f-0674-40fe-9864-d82d11e4bb5e] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 867.126516] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-18b56ec6-b4af-45d8-b1ac-9f5aba88c9e1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.132643] env[62066]: DEBUG oslo_vmware.api [None req-f2684bd2-e68b-4819-84d6-b5b0e8bf10bc tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Waiting for the task: (returnval){ [ 867.132643] env[62066]: value = "task-1340984" [ 867.132643] env[62066]: _type = "Task" [ 867.132643] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.144355] env[62066]: DEBUG oslo_vmware.api [None req-f2684bd2-e68b-4819-84d6-b5b0e8bf10bc tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1340984, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.298391] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a57d8cc3-c15b-431a-a42b-b576492229b1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.306171] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6454ab2a-487f-4415-ae0f-9dc3c6bba27d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.336384] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63aaf4c7-a302-49a4-aa50-d2530ce63aa8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.344149] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fab78d15-d7e6-4073-9099-499493f7132e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.358330] env[62066]: DEBUG nova.compute.provider_tree [None req-4c61eb51-497c-4460-951e-d696b73eff28 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 867.496859] env[62066]: DEBUG oslo_vmware.api [None req-4a3c2cb7-f1e2-437a-b493-e5de08e55742 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Task: {'id': task-1340976, 'name': CreateSnapshot_Task, 'duration_secs': 1.286236} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.497210] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-4a3c2cb7-f1e2-437a-b493-e5de08e55742 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] [instance: 8ff6fbb9-c90f-498d-9a85-d220a8c2f794] Created Snapshot of the VM instance {{(pid=62066) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 867.497967] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c766df5-49bb-49bd-8b2d-8e393f384bfd {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.591137] env[62066]: DEBUG oslo_vmware.api [None req-9a34415f-ab54-458b-b6ce-0d7cde2c3efe tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] Task: {'id': task-1340983, 'name': ReconfigVM_Task, 'duration_secs': 0.155814} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.591807] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-9a34415f-ab54-458b-b6ce-0d7cde2c3efe tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] [instance: 0bc31a8b-a5fc-4f6e-a4c0-024bf4260755] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 867.592076] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-37ae8e8c-5238-4c61-ab52-7341e056a515 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.596126] env[62066]: DEBUG oslo_vmware.api [None req-68383c37-9698-4391-bbf2-b9ebf656dc03 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Task: {'id': task-1340982, 'name': PowerOnVM_Task, 'duration_secs': 0.500895} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.596693] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-68383c37-9698-4391-bbf2-b9ebf656dc03 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: c8e0d47c-4421-4e00-9183-206fceeabc40] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 867.596935] env[62066]: INFO nova.compute.manager [None req-68383c37-9698-4391-bbf2-b9ebf656dc03 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: c8e0d47c-4421-4e00-9183-206fceeabc40] Took 8.44 seconds to spawn the instance on the hypervisor. [ 867.597165] env[62066]: DEBUG nova.compute.manager [None req-68383c37-9698-4391-bbf2-b9ebf656dc03 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: c8e0d47c-4421-4e00-9183-206fceeabc40] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 867.597876] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0fef0cf-edf5-4f75-84e7-63a34e19f59d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.601361] env[62066]: DEBUG oslo_vmware.api [None req-9a34415f-ab54-458b-b6ce-0d7cde2c3efe tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] Waiting for the task: (returnval){ [ 867.601361] env[62066]: value = "task-1340985" [ 867.601361] env[62066]: _type = "Task" [ 867.601361] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.618053] env[62066]: DEBUG oslo_vmware.api [None req-9a34415f-ab54-458b-b6ce-0d7cde2c3efe tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] Task: {'id': task-1340985, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.620107] env[62066]: DEBUG oslo_vmware.api [None req-6f12e7db-6fdf-4bbe-9b1e-ff4283071f33 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Task: {'id': task-1340980, 'name': ReconfigVM_Task, 'duration_secs': 0.565863} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.620462] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-6f12e7db-6fdf-4bbe-9b1e-ff4283071f33 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] [instance: 3120c9a5-0558-4e83-b0fb-994a989398fb] Reconfigured VM instance instance-0000004d to attach disk [datastore2] 3120c9a5-0558-4e83-b0fb-994a989398fb/3120c9a5-0558-4e83-b0fb-994a989398fb.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 867.620987] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a4c756c5-8716-4199-80a6-1302e3c67b5f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.625969] env[62066]: DEBUG oslo_vmware.api [None req-6f12e7db-6fdf-4bbe-9b1e-ff4283071f33 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Waiting for the task: (returnval){ [ 867.625969] env[62066]: value = "task-1340986" [ 867.625969] env[62066]: _type = "Task" [ 867.625969] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.633133] env[62066]: DEBUG oslo_vmware.api [None req-6f12e7db-6fdf-4bbe-9b1e-ff4283071f33 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Task: {'id': task-1340986, 'name': Rename_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.642878] env[62066]: DEBUG oslo_vmware.api [None req-f2684bd2-e68b-4819-84d6-b5b0e8bf10bc tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1340984, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.059707} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.643840] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-f2684bd2-e68b-4819-84d6-b5b0e8bf10bc tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 76306c3f-0674-40fe-9864-d82d11e4bb5e] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 867.644626] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6848a57a-8dce-4edc-bb5b-3bc6e3652b65 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.668973] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-f2684bd2-e68b-4819-84d6-b5b0e8bf10bc tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 76306c3f-0674-40fe-9864-d82d11e4bb5e] Reconfiguring VM instance instance-00000050 to attach disk [datastore2] 76306c3f-0674-40fe-9864-d82d11e4bb5e/76306c3f-0674-40fe-9864-d82d11e4bb5e.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 867.669549] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-062d0033-35d0-429a-8965-3e7ca259379b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.688323] env[62066]: DEBUG oslo_vmware.api [None req-f2684bd2-e68b-4819-84d6-b5b0e8bf10bc tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Waiting for the task: (returnval){ [ 867.688323] env[62066]: value = "task-1340987" [ 867.688323] env[62066]: _type = "Task" [ 867.688323] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.695962] env[62066]: DEBUG oslo_vmware.api [None req-f2684bd2-e68b-4819-84d6-b5b0e8bf10bc tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1340987, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.863442] env[62066]: DEBUG nova.scheduler.client.report [None req-4c61eb51-497c-4460-951e-d696b73eff28 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 867.944212] env[62066]: DEBUG nova.compute.manager [None req-59107a59-1a77-47be-8dae-1ec8f16f1fda tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] [instance: 2634fcc0-96bd-4513-8f92-515f83fe23d7] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 867.973651] env[62066]: DEBUG nova.virt.hardware [None req-59107a59-1a77-47be-8dae-1ec8f16f1fda tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-23T13:40:41Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-23T13:40:23Z,direct_url=,disk_format='vmdk',id=50ff584c-3b50-4395-af07-3e66769bc9f7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='eb52f7069a374c61ae946f052007c6d9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-23T13:40:24Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 867.973913] env[62066]: DEBUG nova.virt.hardware [None req-59107a59-1a77-47be-8dae-1ec8f16f1fda tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 867.974085] env[62066]: DEBUG nova.virt.hardware [None req-59107a59-1a77-47be-8dae-1ec8f16f1fda tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 867.974273] env[62066]: DEBUG nova.virt.hardware [None req-59107a59-1a77-47be-8dae-1ec8f16f1fda tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 867.974422] env[62066]: DEBUG nova.virt.hardware [None req-59107a59-1a77-47be-8dae-1ec8f16f1fda tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 867.974593] env[62066]: DEBUG nova.virt.hardware [None req-59107a59-1a77-47be-8dae-1ec8f16f1fda tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 867.974810] env[62066]: DEBUG nova.virt.hardware [None req-59107a59-1a77-47be-8dae-1ec8f16f1fda tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 867.974976] env[62066]: DEBUG nova.virt.hardware [None req-59107a59-1a77-47be-8dae-1ec8f16f1fda tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 867.975164] env[62066]: DEBUG nova.virt.hardware [None req-59107a59-1a77-47be-8dae-1ec8f16f1fda tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 867.975329] env[62066]: DEBUG nova.virt.hardware [None req-59107a59-1a77-47be-8dae-1ec8f16f1fda tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 867.975502] env[62066]: DEBUG nova.virt.hardware [None req-59107a59-1a77-47be-8dae-1ec8f16f1fda tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 867.976776] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f0a0838-e4a0-44fb-8b22-e0ec36da67c3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.984944] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9d85139-57e1-4452-a8fc-094fb550504b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.016107] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-4a3c2cb7-f1e2-437a-b493-e5de08e55742 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] [instance: 8ff6fbb9-c90f-498d-9a85-d220a8c2f794] Creating linked-clone VM from snapshot {{(pid=62066) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 868.016319] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-afa3e84d-055d-49a4-ac10-2c2a4a087d26 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.024108] env[62066]: DEBUG oslo_vmware.api [None req-4a3c2cb7-f1e2-437a-b493-e5de08e55742 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Waiting for the task: (returnval){ [ 868.024108] env[62066]: value = "task-1340988" [ 868.024108] env[62066]: _type = "Task" [ 868.024108] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.032394] env[62066]: DEBUG oslo_vmware.api [None req-4a3c2cb7-f1e2-437a-b493-e5de08e55742 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Task: {'id': task-1340988, 'name': CloneVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.116463] env[62066]: DEBUG oslo_vmware.api [None req-9a34415f-ab54-458b-b6ce-0d7cde2c3efe tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] Task: {'id': task-1340985, 'name': PowerOnVM_Task} progress is 87%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.121046] env[62066]: INFO nova.compute.manager [None req-68383c37-9698-4391-bbf2-b9ebf656dc03 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: c8e0d47c-4421-4e00-9183-206fceeabc40] Took 22.97 seconds to build instance. [ 868.134950] env[62066]: DEBUG oslo_vmware.api [None req-6f12e7db-6fdf-4bbe-9b1e-ff4283071f33 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Task: {'id': task-1340986, 'name': Rename_Task, 'duration_secs': 0.237212} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.135934] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-6f12e7db-6fdf-4bbe-9b1e-ff4283071f33 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] [instance: 3120c9a5-0558-4e83-b0fb-994a989398fb] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 868.136243] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-31c73c52-1e76-493a-a996-6e44b0eb39ec {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.142643] env[62066]: DEBUG oslo_vmware.api [None req-6f12e7db-6fdf-4bbe-9b1e-ff4283071f33 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Waiting for the task: (returnval){ [ 868.142643] env[62066]: value = "task-1340989" [ 868.142643] env[62066]: _type = "Task" [ 868.142643] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.151849] env[62066]: DEBUG oslo_vmware.api [None req-6f12e7db-6fdf-4bbe-9b1e-ff4283071f33 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Task: {'id': task-1340989, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.198183] env[62066]: DEBUG oslo_vmware.api [None req-f2684bd2-e68b-4819-84d6-b5b0e8bf10bc tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1340987, 'name': ReconfigVM_Task, 'duration_secs': 0.273397} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.198513] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-f2684bd2-e68b-4819-84d6-b5b0e8bf10bc tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 76306c3f-0674-40fe-9864-d82d11e4bb5e] Reconfigured VM instance instance-00000050 to attach disk [datastore2] 76306c3f-0674-40fe-9864-d82d11e4bb5e/76306c3f-0674-40fe-9864-d82d11e4bb5e.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 868.199166] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b3b235fa-deea-4151-a8c1-cbd56b9d559d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.205395] env[62066]: DEBUG oslo_vmware.api [None req-f2684bd2-e68b-4819-84d6-b5b0e8bf10bc tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Waiting for the task: (returnval){ [ 868.205395] env[62066]: value = "task-1340990" [ 868.205395] env[62066]: _type = "Task" [ 868.205395] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.213362] env[62066]: DEBUG oslo_vmware.api [None req-f2684bd2-e68b-4819-84d6-b5b0e8bf10bc tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1340990, 'name': Rename_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.368742] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4c61eb51-497c-4460-951e-d696b73eff28 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.432s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 868.369330] env[62066]: DEBUG nova.compute.manager [None req-4c61eb51-497c-4460-951e-d696b73eff28 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 170e184a-f009-42cf-82dc-eb9696a00a24] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 868.372332] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4ccd1f86-3a5c-4e35-bdaa-b8e507c9718e tempest-ServersTestFqdnHostnames-1367628746 tempest-ServersTestFqdnHostnames-1367628746-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.024s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 868.373896] env[62066]: INFO nova.compute.claims [None req-4ccd1f86-3a5c-4e35-bdaa-b8e507c9718e tempest-ServersTestFqdnHostnames-1367628746 tempest-ServersTestFqdnHostnames-1367628746-project-member] [instance: afaca059-fc4e-4de2-8fa0-d2f226ea7051] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 868.460817] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d7bf9b49-5ae5-4613-a969-417764bdb185 tempest-ServersV294TestFqdnHostnames-843300434 tempest-ServersV294TestFqdnHostnames-843300434-project-member] Acquiring lock "2709d0ad-5dd8-4e3c-b1e6-3d1e2e242380" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 868.461147] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d7bf9b49-5ae5-4613-a969-417764bdb185 tempest-ServersV294TestFqdnHostnames-843300434 tempest-ServersV294TestFqdnHostnames-843300434-project-member] Lock "2709d0ad-5dd8-4e3c-b1e6-3d1e2e242380" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 868.461350] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d7bf9b49-5ae5-4613-a969-417764bdb185 tempest-ServersV294TestFqdnHostnames-843300434 tempest-ServersV294TestFqdnHostnames-843300434-project-member] Acquiring lock "2709d0ad-5dd8-4e3c-b1e6-3d1e2e242380-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 868.461561] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d7bf9b49-5ae5-4613-a969-417764bdb185 tempest-ServersV294TestFqdnHostnames-843300434 tempest-ServersV294TestFqdnHostnames-843300434-project-member] Lock "2709d0ad-5dd8-4e3c-b1e6-3d1e2e242380-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 868.461700] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d7bf9b49-5ae5-4613-a969-417764bdb185 tempest-ServersV294TestFqdnHostnames-843300434 tempest-ServersV294TestFqdnHostnames-843300434-project-member] Lock "2709d0ad-5dd8-4e3c-b1e6-3d1e2e242380-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 868.464759] env[62066]: INFO nova.compute.manager [None req-d7bf9b49-5ae5-4613-a969-417764bdb185 tempest-ServersV294TestFqdnHostnames-843300434 tempest-ServersV294TestFqdnHostnames-843300434-project-member] [instance: 2709d0ad-5dd8-4e3c-b1e6-3d1e2e242380] Terminating instance [ 868.468018] env[62066]: DEBUG nova.compute.manager [None req-d7bf9b49-5ae5-4613-a969-417764bdb185 tempest-ServersV294TestFqdnHostnames-843300434 tempest-ServersV294TestFqdnHostnames-843300434-project-member] [instance: 2709d0ad-5dd8-4e3c-b1e6-3d1e2e242380] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 868.468243] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-d7bf9b49-5ae5-4613-a969-417764bdb185 tempest-ServersV294TestFqdnHostnames-843300434 tempest-ServersV294TestFqdnHostnames-843300434-project-member] [instance: 2709d0ad-5dd8-4e3c-b1e6-3d1e2e242380] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 868.469136] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ea99b38-67a8-4e5a-9f6d-d63ef939d21d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.474809] env[62066]: DEBUG nova.compute.manager [req-16c9db02-87d1-4b31-98b3-e14c2e7b1562 req-64a9644a-1685-4630-a8dc-0ac4b613331e service nova] [instance: 2634fcc0-96bd-4513-8f92-515f83fe23d7] Received event network-vif-plugged-cf42ffe4-6621-4c29-9a6d-687ade676346 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 868.475036] env[62066]: DEBUG oslo_concurrency.lockutils [req-16c9db02-87d1-4b31-98b3-e14c2e7b1562 req-64a9644a-1685-4630-a8dc-0ac4b613331e service nova] Acquiring lock "2634fcc0-96bd-4513-8f92-515f83fe23d7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 868.475253] env[62066]: DEBUG oslo_concurrency.lockutils [req-16c9db02-87d1-4b31-98b3-e14c2e7b1562 req-64a9644a-1685-4630-a8dc-0ac4b613331e service nova] Lock "2634fcc0-96bd-4513-8f92-515f83fe23d7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 868.475423] env[62066]: DEBUG oslo_concurrency.lockutils [req-16c9db02-87d1-4b31-98b3-e14c2e7b1562 req-64a9644a-1685-4630-a8dc-0ac4b613331e service nova] Lock "2634fcc0-96bd-4513-8f92-515f83fe23d7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 868.475593] env[62066]: DEBUG nova.compute.manager [req-16c9db02-87d1-4b31-98b3-e14c2e7b1562 req-64a9644a-1685-4630-a8dc-0ac4b613331e service nova] [instance: 2634fcc0-96bd-4513-8f92-515f83fe23d7] No waiting events found dispatching network-vif-plugged-cf42ffe4-6621-4c29-9a6d-687ade676346 {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 868.475761] env[62066]: WARNING nova.compute.manager [req-16c9db02-87d1-4b31-98b3-e14c2e7b1562 req-64a9644a-1685-4630-a8dc-0ac4b613331e service nova] [instance: 2634fcc0-96bd-4513-8f92-515f83fe23d7] Received unexpected event network-vif-plugged-cf42ffe4-6621-4c29-9a6d-687ade676346 for instance with vm_state building and task_state spawning. [ 868.481145] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-d7bf9b49-5ae5-4613-a969-417764bdb185 tempest-ServersV294TestFqdnHostnames-843300434 tempest-ServersV294TestFqdnHostnames-843300434-project-member] [instance: 2709d0ad-5dd8-4e3c-b1e6-3d1e2e242380] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 868.481401] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4bcd167d-5892-4750-a946-7950ddc66925 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.488450] env[62066]: DEBUG oslo_vmware.api [None req-d7bf9b49-5ae5-4613-a969-417764bdb185 tempest-ServersV294TestFqdnHostnames-843300434 tempest-ServersV294TestFqdnHostnames-843300434-project-member] Waiting for the task: (returnval){ [ 868.488450] env[62066]: value = "task-1340991" [ 868.488450] env[62066]: _type = "Task" [ 868.488450] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.496723] env[62066]: DEBUG oslo_vmware.api [None req-d7bf9b49-5ae5-4613-a969-417764bdb185 tempest-ServersV294TestFqdnHostnames-843300434 tempest-ServersV294TestFqdnHostnames-843300434-project-member] Task: {'id': task-1340991, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.534014] env[62066]: DEBUG oslo_vmware.api [None req-4a3c2cb7-f1e2-437a-b493-e5de08e55742 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Task: {'id': task-1340988, 'name': CloneVM_Task} progress is 94%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.612995] env[62066]: DEBUG oslo_vmware.api [None req-9a34415f-ab54-458b-b6ce-0d7cde2c3efe tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] Task: {'id': task-1340985, 'name': PowerOnVM_Task, 'duration_secs': 0.931147} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.613514] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-9a34415f-ab54-458b-b6ce-0d7cde2c3efe tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] [instance: 0bc31a8b-a5fc-4f6e-a4c0-024bf4260755] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 868.617847] env[62066]: DEBUG nova.compute.manager [None req-9a34415f-ab54-458b-b6ce-0d7cde2c3efe tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] [instance: 0bc31a8b-a5fc-4f6e-a4c0-024bf4260755] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 868.618876] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-719d11c8-5712-4dad-b541-7577057e3a8e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.623169] env[62066]: DEBUG nova.network.neutron [None req-59107a59-1a77-47be-8dae-1ec8f16f1fda tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] [instance: 2634fcc0-96bd-4513-8f92-515f83fe23d7] Successfully updated port: cf42ffe4-6621-4c29-9a6d-687ade676346 {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 868.624990] env[62066]: DEBUG oslo_concurrency.lockutils [None req-68383c37-9698-4391-bbf2-b9ebf656dc03 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Lock "c8e0d47c-4421-4e00-9183-206fceeabc40" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 24.485s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 868.653726] env[62066]: DEBUG oslo_vmware.api [None req-6f12e7db-6fdf-4bbe-9b1e-ff4283071f33 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Task: {'id': task-1340989, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.715750] env[62066]: DEBUG oslo_vmware.api [None req-f2684bd2-e68b-4819-84d6-b5b0e8bf10bc tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1340990, 'name': Rename_Task, 'duration_secs': 0.160886} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.716054] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-f2684bd2-e68b-4819-84d6-b5b0e8bf10bc tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 76306c3f-0674-40fe-9864-d82d11e4bb5e] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 868.716307] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ef2370dc-4b9f-42dc-9115-1d21bf738242 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.723374] env[62066]: DEBUG oslo_vmware.api [None req-f2684bd2-e68b-4819-84d6-b5b0e8bf10bc tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Waiting for the task: (returnval){ [ 868.723374] env[62066]: value = "task-1340992" [ 868.723374] env[62066]: _type = "Task" [ 868.723374] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.731363] env[62066]: DEBUG oslo_vmware.api [None req-f2684bd2-e68b-4819-84d6-b5b0e8bf10bc tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1340992, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.881024] env[62066]: DEBUG nova.compute.utils [None req-4c61eb51-497c-4460-951e-d696b73eff28 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 868.882555] env[62066]: DEBUG nova.compute.manager [None req-4c61eb51-497c-4460-951e-d696b73eff28 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 170e184a-f009-42cf-82dc-eb9696a00a24] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 868.882870] env[62066]: DEBUG nova.network.neutron [None req-4c61eb51-497c-4460-951e-d696b73eff28 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 170e184a-f009-42cf-82dc-eb9696a00a24] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 868.966029] env[62066]: DEBUG nova.policy [None req-4c61eb51-497c-4460-951e-d696b73eff28 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8910a229218b4ec5ad72c893badfc598', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6497ab02f327476d8ff81c2ecc0371e5', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 869.000432] env[62066]: DEBUG oslo_vmware.api [None req-d7bf9b49-5ae5-4613-a969-417764bdb185 tempest-ServersV294TestFqdnHostnames-843300434 tempest-ServersV294TestFqdnHostnames-843300434-project-member] Task: {'id': task-1340991, 'name': PowerOffVM_Task, 'duration_secs': 0.187557} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.000432] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-d7bf9b49-5ae5-4613-a969-417764bdb185 tempest-ServersV294TestFqdnHostnames-843300434 tempest-ServersV294TestFqdnHostnames-843300434-project-member] [instance: 2709d0ad-5dd8-4e3c-b1e6-3d1e2e242380] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 869.000432] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-d7bf9b49-5ae5-4613-a969-417764bdb185 tempest-ServersV294TestFqdnHostnames-843300434 tempest-ServersV294TestFqdnHostnames-843300434-project-member] [instance: 2709d0ad-5dd8-4e3c-b1e6-3d1e2e242380] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 869.000432] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0ef52912-fd6c-4e91-b4a9-90be980e82e1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.034294] env[62066]: DEBUG oslo_vmware.api [None req-4a3c2cb7-f1e2-437a-b493-e5de08e55742 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Task: {'id': task-1340988, 'name': CloneVM_Task} progress is 94%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.070036] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-d7bf9b49-5ae5-4613-a969-417764bdb185 tempest-ServersV294TestFqdnHostnames-843300434 tempest-ServersV294TestFqdnHostnames-843300434-project-member] [instance: 2709d0ad-5dd8-4e3c-b1e6-3d1e2e242380] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 869.070036] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-d7bf9b49-5ae5-4613-a969-417764bdb185 tempest-ServersV294TestFqdnHostnames-843300434 tempest-ServersV294TestFqdnHostnames-843300434-project-member] [instance: 2709d0ad-5dd8-4e3c-b1e6-3d1e2e242380] Deleting contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 869.070036] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-d7bf9b49-5ae5-4613-a969-417764bdb185 tempest-ServersV294TestFqdnHostnames-843300434 tempest-ServersV294TestFqdnHostnames-843300434-project-member] Deleting the datastore file [datastore2] 2709d0ad-5dd8-4e3c-b1e6-3d1e2e242380 {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 869.070036] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-75544ad0-7555-4aad-90b6-96d5091890ee {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.077553] env[62066]: DEBUG oslo_vmware.api [None req-d7bf9b49-5ae5-4613-a969-417764bdb185 tempest-ServersV294TestFqdnHostnames-843300434 tempest-ServersV294TestFqdnHostnames-843300434-project-member] Waiting for the task: (returnval){ [ 869.077553] env[62066]: value = "task-1340994" [ 869.077553] env[62066]: _type = "Task" [ 869.077553] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.086020] env[62066]: DEBUG oslo_vmware.api [None req-d7bf9b49-5ae5-4613-a969-417764bdb185 tempest-ServersV294TestFqdnHostnames-843300434 tempest-ServersV294TestFqdnHostnames-843300434-project-member] Task: {'id': task-1340994, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.126109] env[62066]: DEBUG oslo_concurrency.lockutils [None req-59107a59-1a77-47be-8dae-1ec8f16f1fda tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Acquiring lock "refresh_cache-2634fcc0-96bd-4513-8f92-515f83fe23d7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 869.127994] env[62066]: DEBUG oslo_concurrency.lockutils [None req-59107a59-1a77-47be-8dae-1ec8f16f1fda tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Acquired lock "refresh_cache-2634fcc0-96bd-4513-8f92-515f83fe23d7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 869.127994] env[62066]: DEBUG nova.network.neutron [None req-59107a59-1a77-47be-8dae-1ec8f16f1fda tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] [instance: 2634fcc0-96bd-4513-8f92-515f83fe23d7] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 869.156478] env[62066]: DEBUG oslo_vmware.api [None req-6f12e7db-6fdf-4bbe-9b1e-ff4283071f33 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Task: {'id': task-1340989, 'name': PowerOnVM_Task, 'duration_secs': 0.649449} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.157792] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-6f12e7db-6fdf-4bbe-9b1e-ff4283071f33 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] [instance: 3120c9a5-0558-4e83-b0fb-994a989398fb] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 869.161018] env[62066]: DEBUG nova.compute.manager [None req-6f12e7db-6fdf-4bbe-9b1e-ff4283071f33 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] [instance: 3120c9a5-0558-4e83-b0fb-994a989398fb] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 869.161018] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3985086-036a-4816-bf4b-4e4d96750fd7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.236024] env[62066]: DEBUG oslo_vmware.api [None req-f2684bd2-e68b-4819-84d6-b5b0e8bf10bc tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1340992, 'name': PowerOnVM_Task, 'duration_secs': 0.450316} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.236024] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-f2684bd2-e68b-4819-84d6-b5b0e8bf10bc tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 76306c3f-0674-40fe-9864-d82d11e4bb5e] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 869.236024] env[62066]: INFO nova.compute.manager [None req-f2684bd2-e68b-4819-84d6-b5b0e8bf10bc tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 76306c3f-0674-40fe-9864-d82d11e4bb5e] Took 7.70 seconds to spawn the instance on the hypervisor. [ 869.236024] env[62066]: DEBUG nova.compute.manager [None req-f2684bd2-e68b-4819-84d6-b5b0e8bf10bc tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 76306c3f-0674-40fe-9864-d82d11e4bb5e] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 869.236024] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f197da9b-3d28-4844-85e5-6a9e4af2652c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.383308] env[62066]: DEBUG nova.compute.manager [None req-4c61eb51-497c-4460-951e-d696b73eff28 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 170e184a-f009-42cf-82dc-eb9696a00a24] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 869.537592] env[62066]: DEBUG oslo_vmware.api [None req-4a3c2cb7-f1e2-437a-b493-e5de08e55742 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Task: {'id': task-1340988, 'name': CloneVM_Task} progress is 100%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.590770] env[62066]: DEBUG oslo_vmware.api [None req-d7bf9b49-5ae5-4613-a969-417764bdb185 tempest-ServersV294TestFqdnHostnames-843300434 tempest-ServersV294TestFqdnHostnames-843300434-project-member] Task: {'id': task-1340994, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.189478} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.591368] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-d7bf9b49-5ae5-4613-a969-417764bdb185 tempest-ServersV294TestFqdnHostnames-843300434 tempest-ServersV294TestFqdnHostnames-843300434-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 869.591703] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-d7bf9b49-5ae5-4613-a969-417764bdb185 tempest-ServersV294TestFqdnHostnames-843300434 tempest-ServersV294TestFqdnHostnames-843300434-project-member] [instance: 2709d0ad-5dd8-4e3c-b1e6-3d1e2e242380] Deleted contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 869.592053] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-d7bf9b49-5ae5-4613-a969-417764bdb185 tempest-ServersV294TestFqdnHostnames-843300434 tempest-ServersV294TestFqdnHostnames-843300434-project-member] [instance: 2709d0ad-5dd8-4e3c-b1e6-3d1e2e242380] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 869.592370] env[62066]: INFO nova.compute.manager [None req-d7bf9b49-5ae5-4613-a969-417764bdb185 tempest-ServersV294TestFqdnHostnames-843300434 tempest-ServersV294TestFqdnHostnames-843300434-project-member] [instance: 2709d0ad-5dd8-4e3c-b1e6-3d1e2e242380] Took 1.12 seconds to destroy the instance on the hypervisor. [ 869.592757] env[62066]: DEBUG oslo.service.loopingcall [None req-d7bf9b49-5ae5-4613-a969-417764bdb185 tempest-ServersV294TestFqdnHostnames-843300434 tempest-ServersV294TestFqdnHostnames-843300434-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 869.593098] env[62066]: DEBUG nova.compute.manager [-] [instance: 2709d0ad-5dd8-4e3c-b1e6-3d1e2e242380] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 869.595018] env[62066]: DEBUG nova.network.neutron [-] [instance: 2709d0ad-5dd8-4e3c-b1e6-3d1e2e242380] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 869.661048] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86951d7f-cc73-4a89-9d2b-1cc11f290940 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.666647] env[62066]: DEBUG nova.network.neutron [None req-4c61eb51-497c-4460-951e-d696b73eff28 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 170e184a-f009-42cf-82dc-eb9696a00a24] Successfully created port: e8281131-9972-486f-b3be-2452487f99a3 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 869.674637] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a74aa84d-46f3-4cc2-b33e-cedceb2c033b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.685458] env[62066]: DEBUG nova.network.neutron [None req-59107a59-1a77-47be-8dae-1ec8f16f1fda tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] [instance: 2634fcc0-96bd-4513-8f92-515f83fe23d7] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 869.712043] env[62066]: DEBUG oslo_concurrency.lockutils [None req-6f12e7db-6fdf-4bbe-9b1e-ff4283071f33 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 869.715841] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3e351c5-4ae7-4111-8169-293902e112b0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.723951] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbf08e28-9c15-410d-bf04-f617b3b45d5b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.738967] env[62066]: DEBUG nova.compute.provider_tree [None req-4ccd1f86-3a5c-4e35-bdaa-b8e507c9718e tempest-ServersTestFqdnHostnames-1367628746 tempest-ServersTestFqdnHostnames-1367628746-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 869.756479] env[62066]: INFO nova.compute.manager [None req-f2684bd2-e68b-4819-84d6-b5b0e8bf10bc tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 76306c3f-0674-40fe-9864-d82d11e4bb5e] Took 24.23 seconds to build instance. [ 869.927721] env[62066]: DEBUG nova.network.neutron [None req-59107a59-1a77-47be-8dae-1ec8f16f1fda tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] [instance: 2634fcc0-96bd-4513-8f92-515f83fe23d7] Updating instance_info_cache with network_info: [{"id": "cf42ffe4-6621-4c29-9a6d-687ade676346", "address": "fa:16:3e:4c:ae:2f", "network": {"id": "fac1b86c-746c-4b6e-8d7f-21fe6ed404a3", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1827773279-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "59855b1cf91d43e3ba0482f414c4e546", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "88651df2-0506-4f6c-b868-dd30a81f2b1c", "external-id": "nsx-vlan-transportzone-366", "segmentation_id": 366, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcf42ffe4-66", "ovs_interfaceid": "cf42ffe4-6621-4c29-9a6d-687ade676346", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 870.041065] env[62066]: DEBUG oslo_vmware.api [None req-4a3c2cb7-f1e2-437a-b493-e5de08e55742 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Task: {'id': task-1340988, 'name': CloneVM_Task, 'duration_secs': 1.585602} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.041065] env[62066]: INFO nova.virt.vmwareapi.vmops [None req-4a3c2cb7-f1e2-437a-b493-e5de08e55742 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] [instance: 8ff6fbb9-c90f-498d-9a85-d220a8c2f794] Created linked-clone VM from snapshot [ 870.041065] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f69acfdd-4e47-4644-8933-bcc44d41bf1b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.049881] env[62066]: DEBUG nova.virt.vmwareapi.images [None req-4a3c2cb7-f1e2-437a-b493-e5de08e55742 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] [instance: 8ff6fbb9-c90f-498d-9a85-d220a8c2f794] Uploading image 1ebd2ae1-0ae3-4486-822c-c75cafe27fe3 {{(pid=62066) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 870.090207] env[62066]: DEBUG oslo_vmware.rw_handles [None req-4a3c2cb7-f1e2-437a-b493-e5de08e55742 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 870.090207] env[62066]: value = "vm-286079" [ 870.090207] env[62066]: _type = "VirtualMachine" [ 870.090207] env[62066]: }. {{(pid=62066) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 870.092455] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-b54a1a33-ed17-4d36-89d2-85403681309f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.097688] env[62066]: DEBUG oslo_vmware.rw_handles [None req-4a3c2cb7-f1e2-437a-b493-e5de08e55742 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Lease: (returnval){ [ 870.097688] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]52aa643f-0ab2-e74f-d867-3f25486b0633" [ 870.097688] env[62066]: _type = "HttpNfcLease" [ 870.097688] env[62066]: } obtained for exporting VM: (result){ [ 870.097688] env[62066]: value = "vm-286079" [ 870.097688] env[62066]: _type = "VirtualMachine" [ 870.097688] env[62066]: }. {{(pid=62066) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 870.098067] env[62066]: DEBUG oslo_vmware.api [None req-4a3c2cb7-f1e2-437a-b493-e5de08e55742 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Waiting for the lease: (returnval){ [ 870.098067] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]52aa643f-0ab2-e74f-d867-3f25486b0633" [ 870.098067] env[62066]: _type = "HttpNfcLease" [ 870.098067] env[62066]: } to be ready. {{(pid=62066) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 870.104671] env[62066]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 870.104671] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]52aa643f-0ab2-e74f-d867-3f25486b0633" [ 870.104671] env[62066]: _type = "HttpNfcLease" [ 870.104671] env[62066]: } is initializing. {{(pid=62066) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 870.243436] env[62066]: DEBUG nova.scheduler.client.report [None req-4ccd1f86-3a5c-4e35-bdaa-b8e507c9718e tempest-ServersTestFqdnHostnames-1367628746 tempest-ServersTestFqdnHostnames-1367628746-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 870.259826] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f2684bd2-e68b-4819-84d6-b5b0e8bf10bc tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Lock "76306c3f-0674-40fe-9864-d82d11e4bb5e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 25.257s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 870.367374] env[62066]: DEBUG nova.compute.manager [req-28f2e6a9-3d4b-490d-a10c-0637fab88adc req-199dc08d-71e5-47c8-89e3-52c0cb9c2583 service nova] [instance: 2709d0ad-5dd8-4e3c-b1e6-3d1e2e242380] Received event network-vif-deleted-c5cdc663-1e55-4238-bad4-7fb406b34a36 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 870.367596] env[62066]: INFO nova.compute.manager [req-28f2e6a9-3d4b-490d-a10c-0637fab88adc req-199dc08d-71e5-47c8-89e3-52c0cb9c2583 service nova] [instance: 2709d0ad-5dd8-4e3c-b1e6-3d1e2e242380] Neutron deleted interface c5cdc663-1e55-4238-bad4-7fb406b34a36; detaching it from the instance and deleting it from the info cache [ 870.367833] env[62066]: DEBUG nova.network.neutron [req-28f2e6a9-3d4b-490d-a10c-0637fab88adc req-199dc08d-71e5-47c8-89e3-52c0cb9c2583 service nova] [instance: 2709d0ad-5dd8-4e3c-b1e6-3d1e2e242380] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 870.393596] env[62066]: DEBUG nova.compute.manager [None req-4c61eb51-497c-4460-951e-d696b73eff28 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 170e184a-f009-42cf-82dc-eb9696a00a24] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 870.419152] env[62066]: DEBUG nova.virt.hardware [None req-4c61eb51-497c-4460-951e-d696b73eff28 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-23T13:40:41Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-23T13:40:23Z,direct_url=,disk_format='vmdk',id=50ff584c-3b50-4395-af07-3e66769bc9f7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='eb52f7069a374c61ae946f052007c6d9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-23T13:40:24Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 870.419411] env[62066]: DEBUG nova.virt.hardware [None req-4c61eb51-497c-4460-951e-d696b73eff28 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 870.419583] env[62066]: DEBUG nova.virt.hardware [None req-4c61eb51-497c-4460-951e-d696b73eff28 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 870.419797] env[62066]: DEBUG nova.virt.hardware [None req-4c61eb51-497c-4460-951e-d696b73eff28 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 870.419945] env[62066]: DEBUG nova.virt.hardware [None req-4c61eb51-497c-4460-951e-d696b73eff28 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 870.420325] env[62066]: DEBUG nova.virt.hardware [None req-4c61eb51-497c-4460-951e-d696b73eff28 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 870.420639] env[62066]: DEBUG nova.virt.hardware [None req-4c61eb51-497c-4460-951e-d696b73eff28 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 870.420864] env[62066]: DEBUG nova.virt.hardware [None req-4c61eb51-497c-4460-951e-d696b73eff28 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 870.421107] env[62066]: DEBUG nova.virt.hardware [None req-4c61eb51-497c-4460-951e-d696b73eff28 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 870.421322] env[62066]: DEBUG nova.virt.hardware [None req-4c61eb51-497c-4460-951e-d696b73eff28 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 870.421541] env[62066]: DEBUG nova.virt.hardware [None req-4c61eb51-497c-4460-951e-d696b73eff28 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 870.422912] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40404cb8-44eb-4e30-a3f2-6d9c641e14d7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.433549] env[62066]: DEBUG oslo_concurrency.lockutils [None req-59107a59-1a77-47be-8dae-1ec8f16f1fda tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Releasing lock "refresh_cache-2634fcc0-96bd-4513-8f92-515f83fe23d7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 870.433549] env[62066]: DEBUG nova.compute.manager [None req-59107a59-1a77-47be-8dae-1ec8f16f1fda tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] [instance: 2634fcc0-96bd-4513-8f92-515f83fe23d7] Instance network_info: |[{"id": "cf42ffe4-6621-4c29-9a6d-687ade676346", "address": "fa:16:3e:4c:ae:2f", "network": {"id": "fac1b86c-746c-4b6e-8d7f-21fe6ed404a3", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1827773279-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "59855b1cf91d43e3ba0482f414c4e546", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "88651df2-0506-4f6c-b868-dd30a81f2b1c", "external-id": "nsx-vlan-transportzone-366", "segmentation_id": 366, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcf42ffe4-66", "ovs_interfaceid": "cf42ffe4-6621-4c29-9a6d-687ade676346", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 870.433549] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-59107a59-1a77-47be-8dae-1ec8f16f1fda tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] [instance: 2634fcc0-96bd-4513-8f92-515f83fe23d7] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:4c:ae:2f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '88651df2-0506-4f6c-b868-dd30a81f2b1c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'cf42ffe4-6621-4c29-9a6d-687ade676346', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 870.443416] env[62066]: DEBUG oslo.service.loopingcall [None req-59107a59-1a77-47be-8dae-1ec8f16f1fda tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 870.444995] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05996ce7-282b-457c-9be4-c007fa95b56d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.453204] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2634fcc0-96bd-4513-8f92-515f83fe23d7] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 870.453204] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-911904b3-6798-4429-b1ab-4318decd5e2a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.477224] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 870.477224] env[62066]: value = "task-1340996" [ 870.477224] env[62066]: _type = "Task" [ 870.477224] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.487157] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1340996, 'name': CreateVM_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.609874] env[62066]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 870.609874] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]52aa643f-0ab2-e74f-d867-3f25486b0633" [ 870.609874] env[62066]: _type = "HttpNfcLease" [ 870.609874] env[62066]: } is ready. {{(pid=62066) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 870.611026] env[62066]: DEBUG oslo_vmware.rw_handles [None req-4a3c2cb7-f1e2-437a-b493-e5de08e55742 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 870.611026] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]52aa643f-0ab2-e74f-d867-3f25486b0633" [ 870.611026] env[62066]: _type = "HttpNfcLease" [ 870.611026] env[62066]: }. {{(pid=62066) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 870.611941] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e315eb3-de13-4f89-989d-35273f32f9ee {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.623464] env[62066]: DEBUG oslo_vmware.rw_handles [None req-4a3c2cb7-f1e2-437a-b493-e5de08e55742 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/520ab008-37bf-a52c-cfda-92b0db99cbeb/disk-0.vmdk from lease info. {{(pid=62066) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 870.623937] env[62066]: DEBUG oslo_vmware.rw_handles [None req-4a3c2cb7-f1e2-437a-b493-e5de08e55742 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/520ab008-37bf-a52c-cfda-92b0db99cbeb/disk-0.vmdk for reading. {{(pid=62066) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 870.747021] env[62066]: DEBUG nova.network.neutron [-] [instance: 2709d0ad-5dd8-4e3c-b1e6-3d1e2e242380] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 870.752549] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4ccd1f86-3a5c-4e35-bdaa-b8e507c9718e tempest-ServersTestFqdnHostnames-1367628746 tempest-ServersTestFqdnHostnames-1367628746-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.380s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 870.753173] env[62066]: DEBUG nova.compute.manager [None req-4ccd1f86-3a5c-4e35-bdaa-b8e507c9718e tempest-ServersTestFqdnHostnames-1367628746 tempest-ServersTestFqdnHostnames-1367628746-project-member] [instance: afaca059-fc4e-4de2-8fa0-d2f226ea7051] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 870.756320] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5327c80f-b6fd-46f0-ac98-3857740d94e2 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 13.640s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 870.756679] env[62066]: DEBUG nova.objects.instance [None req-5327c80f-b6fd-46f0-ac98-3857740d94e2 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Lazy-loading 'resources' on Instance uuid 1e1a2ede-4ad8-4600-851e-6f2046b3f919 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 870.780851] env[62066]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-330f442f-5494-41f0-9ccc-a2c53333bd2b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.871047] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9dc66c7a-bc5d-438e-985c-6afabf33e47f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.880148] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc9808ca-6dfe-4d49-bbcc-522911a0ff42 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.911536] env[62066]: DEBUG nova.compute.manager [req-28f2e6a9-3d4b-490d-a10c-0637fab88adc req-199dc08d-71e5-47c8-89e3-52c0cb9c2583 service nova] [instance: 2709d0ad-5dd8-4e3c-b1e6-3d1e2e242380] Detach interface failed, port_id=c5cdc663-1e55-4238-bad4-7fb406b34a36, reason: Instance 2709d0ad-5dd8-4e3c-b1e6-3d1e2e242380 could not be found. {{(pid=62066) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 870.979970] env[62066]: DEBUG nova.compute.manager [req-5a81dfe6-f9ee-4753-aa8a-d0144e77bf6d req-4acde67d-511e-4128-91e2-9ef962a1c68a service nova] [instance: 2634fcc0-96bd-4513-8f92-515f83fe23d7] Received event network-changed-cf42ffe4-6621-4c29-9a6d-687ade676346 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 870.980235] env[62066]: DEBUG nova.compute.manager [req-5a81dfe6-f9ee-4753-aa8a-d0144e77bf6d req-4acde67d-511e-4128-91e2-9ef962a1c68a service nova] [instance: 2634fcc0-96bd-4513-8f92-515f83fe23d7] Refreshing instance network info cache due to event network-changed-cf42ffe4-6621-4c29-9a6d-687ade676346. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 870.980471] env[62066]: DEBUG oslo_concurrency.lockutils [req-5a81dfe6-f9ee-4753-aa8a-d0144e77bf6d req-4acde67d-511e-4128-91e2-9ef962a1c68a service nova] Acquiring lock "refresh_cache-2634fcc0-96bd-4513-8f92-515f83fe23d7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 870.980598] env[62066]: DEBUG oslo_concurrency.lockutils [req-5a81dfe6-f9ee-4753-aa8a-d0144e77bf6d req-4acde67d-511e-4128-91e2-9ef962a1c68a service nova] Acquired lock "refresh_cache-2634fcc0-96bd-4513-8f92-515f83fe23d7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 870.980827] env[62066]: DEBUG nova.network.neutron [req-5a81dfe6-f9ee-4753-aa8a-d0144e77bf6d req-4acde67d-511e-4128-91e2-9ef962a1c68a service nova] [instance: 2634fcc0-96bd-4513-8f92-515f83fe23d7] Refreshing network info cache for port cf42ffe4-6621-4c29-9a6d-687ade676346 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 870.991885] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1340996, 'name': CreateVM_Task} progress is 25%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.248780] env[62066]: INFO nova.compute.manager [-] [instance: 2709d0ad-5dd8-4e3c-b1e6-3d1e2e242380] Took 1.66 seconds to deallocate network for instance. [ 871.260191] env[62066]: DEBUG nova.compute.utils [None req-4ccd1f86-3a5c-4e35-bdaa-b8e507c9718e tempest-ServersTestFqdnHostnames-1367628746 tempest-ServersTestFqdnHostnames-1367628746-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 871.265919] env[62066]: DEBUG nova.compute.manager [None req-4ccd1f86-3a5c-4e35-bdaa-b8e507c9718e tempest-ServersTestFqdnHostnames-1367628746 tempest-ServersTestFqdnHostnames-1367628746-project-member] [instance: afaca059-fc4e-4de2-8fa0-d2f226ea7051] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 871.265919] env[62066]: DEBUG nova.network.neutron [None req-4ccd1f86-3a5c-4e35-bdaa-b8e507c9718e tempest-ServersTestFqdnHostnames-1367628746 tempest-ServersTestFqdnHostnames-1367628746-project-member] [instance: afaca059-fc4e-4de2-8fa0-d2f226ea7051] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 871.344830] env[62066]: DEBUG nova.policy [None req-4ccd1f86-3a5c-4e35-bdaa-b8e507c9718e tempest-ServersTestFqdnHostnames-1367628746 tempest-ServersTestFqdnHostnames-1367628746-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '881455975097421cba260b6f1658c530', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '511b5cd53c7745eb8c3e9589b1de01c6', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 871.501645] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1340996, 'name': CreateVM_Task} progress is 25%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.549512] env[62066]: DEBUG oslo_concurrency.lockutils [None req-535b520b-dc22-40fb-a872-f0522e7e73a7 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Acquiring lock "3120c9a5-0558-4e83-b0fb-994a989398fb" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 871.550509] env[62066]: DEBUG oslo_concurrency.lockutils [None req-535b520b-dc22-40fb-a872-f0522e7e73a7 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Lock "3120c9a5-0558-4e83-b0fb-994a989398fb" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 871.550509] env[62066]: DEBUG oslo_concurrency.lockutils [None req-535b520b-dc22-40fb-a872-f0522e7e73a7 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Acquiring lock "3120c9a5-0558-4e83-b0fb-994a989398fb-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 871.551056] env[62066]: DEBUG oslo_concurrency.lockutils [None req-535b520b-dc22-40fb-a872-f0522e7e73a7 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Lock "3120c9a5-0558-4e83-b0fb-994a989398fb-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 871.551213] env[62066]: DEBUG oslo_concurrency.lockutils [None req-535b520b-dc22-40fb-a872-f0522e7e73a7 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Lock "3120c9a5-0558-4e83-b0fb-994a989398fb-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 871.563953] env[62066]: INFO nova.compute.manager [None req-535b520b-dc22-40fb-a872-f0522e7e73a7 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] [instance: 3120c9a5-0558-4e83-b0fb-994a989398fb] Terminating instance [ 871.568225] env[62066]: DEBUG oslo_concurrency.lockutils [None req-535b520b-dc22-40fb-a872-f0522e7e73a7 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Acquiring lock "refresh_cache-3120c9a5-0558-4e83-b0fb-994a989398fb" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 871.568804] env[62066]: DEBUG oslo_concurrency.lockutils [None req-535b520b-dc22-40fb-a872-f0522e7e73a7 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Acquired lock "refresh_cache-3120c9a5-0558-4e83-b0fb-994a989398fb" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 871.569572] env[62066]: DEBUG nova.network.neutron [None req-535b520b-dc22-40fb-a872-f0522e7e73a7 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] [instance: 3120c9a5-0558-4e83-b0fb-994a989398fb] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 871.622620] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3dc36fa-eda5-4171-ba64-80ade059bc9c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.631144] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8d87301-64f5-4ae3-afb4-36220f97518f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.666312] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7d7b686-5b15-410c-a371-c7a2b474e900 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.677586] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b811c812-a1d9-4af2-b930-c42e77e15a99 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.682861] env[62066]: DEBUG nova.network.neutron [None req-4c61eb51-497c-4460-951e-d696b73eff28 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 170e184a-f009-42cf-82dc-eb9696a00a24] Successfully updated port: e8281131-9972-486f-b3be-2452487f99a3 {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 871.695197] env[62066]: DEBUG nova.compute.provider_tree [None req-5327c80f-b6fd-46f0-ac98-3857740d94e2 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 871.756286] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d7bf9b49-5ae5-4613-a969-417764bdb185 tempest-ServersV294TestFqdnHostnames-843300434 tempest-ServersV294TestFqdnHostnames-843300434-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 871.772027] env[62066]: DEBUG nova.compute.manager [None req-4ccd1f86-3a5c-4e35-bdaa-b8e507c9718e tempest-ServersTestFqdnHostnames-1367628746 tempest-ServersTestFqdnHostnames-1367628746-project-member] [instance: afaca059-fc4e-4de2-8fa0-d2f226ea7051] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 871.833494] env[62066]: DEBUG nova.network.neutron [None req-4ccd1f86-3a5c-4e35-bdaa-b8e507c9718e tempest-ServersTestFqdnHostnames-1367628746 tempest-ServersTestFqdnHostnames-1367628746-project-member] [instance: afaca059-fc4e-4de2-8fa0-d2f226ea7051] Successfully created port: 725b6138-b39f-41c3-bf84-1a13a3e30fa9 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 871.917468] env[62066]: DEBUG nova.network.neutron [req-5a81dfe6-f9ee-4753-aa8a-d0144e77bf6d req-4acde67d-511e-4128-91e2-9ef962a1c68a service nova] [instance: 2634fcc0-96bd-4513-8f92-515f83fe23d7] Updated VIF entry in instance network info cache for port cf42ffe4-6621-4c29-9a6d-687ade676346. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 871.917867] env[62066]: DEBUG nova.network.neutron [req-5a81dfe6-f9ee-4753-aa8a-d0144e77bf6d req-4acde67d-511e-4128-91e2-9ef962a1c68a service nova] [instance: 2634fcc0-96bd-4513-8f92-515f83fe23d7] Updating instance_info_cache with network_info: [{"id": "cf42ffe4-6621-4c29-9a6d-687ade676346", "address": "fa:16:3e:4c:ae:2f", "network": {"id": "fac1b86c-746c-4b6e-8d7f-21fe6ed404a3", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1827773279-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "59855b1cf91d43e3ba0482f414c4e546", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "88651df2-0506-4f6c-b868-dd30a81f2b1c", "external-id": "nsx-vlan-transportzone-366", "segmentation_id": 366, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcf42ffe4-66", "ovs_interfaceid": "cf42ffe4-6621-4c29-9a6d-687ade676346", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 872.001546] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1340996, 'name': CreateVM_Task} progress is 25%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.091152] env[62066]: DEBUG nova.network.neutron [None req-535b520b-dc22-40fb-a872-f0522e7e73a7 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] [instance: 3120c9a5-0558-4e83-b0fb-994a989398fb] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 872.171295] env[62066]: DEBUG nova.network.neutron [None req-535b520b-dc22-40fb-a872-f0522e7e73a7 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] [instance: 3120c9a5-0558-4e83-b0fb-994a989398fb] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 872.201129] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4c61eb51-497c-4460-951e-d696b73eff28 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Acquiring lock "refresh_cache-170e184a-f009-42cf-82dc-eb9696a00a24" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 872.201129] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4c61eb51-497c-4460-951e-d696b73eff28 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Acquired lock "refresh_cache-170e184a-f009-42cf-82dc-eb9696a00a24" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 872.201129] env[62066]: DEBUG nova.network.neutron [None req-4c61eb51-497c-4460-951e-d696b73eff28 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 170e184a-f009-42cf-82dc-eb9696a00a24] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 872.203327] env[62066]: DEBUG nova.scheduler.client.report [None req-5327c80f-b6fd-46f0-ac98-3857740d94e2 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 872.421469] env[62066]: DEBUG oslo_concurrency.lockutils [req-5a81dfe6-f9ee-4753-aa8a-d0144e77bf6d req-4acde67d-511e-4128-91e2-9ef962a1c68a service nova] Releasing lock "refresh_cache-2634fcc0-96bd-4513-8f92-515f83fe23d7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 872.422010] env[62066]: DEBUG nova.compute.manager [req-5a81dfe6-f9ee-4753-aa8a-d0144e77bf6d req-4acde67d-511e-4128-91e2-9ef962a1c68a service nova] [instance: c8e0d47c-4421-4e00-9183-206fceeabc40] Received event network-changed-bb9924e0-8b18-43b2-bda0-3e409fa16718 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 872.422236] env[62066]: DEBUG nova.compute.manager [req-5a81dfe6-f9ee-4753-aa8a-d0144e77bf6d req-4acde67d-511e-4128-91e2-9ef962a1c68a service nova] [instance: c8e0d47c-4421-4e00-9183-206fceeabc40] Refreshing instance network info cache due to event network-changed-bb9924e0-8b18-43b2-bda0-3e409fa16718. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 872.422460] env[62066]: DEBUG oslo_concurrency.lockutils [req-5a81dfe6-f9ee-4753-aa8a-d0144e77bf6d req-4acde67d-511e-4128-91e2-9ef962a1c68a service nova] Acquiring lock "refresh_cache-c8e0d47c-4421-4e00-9183-206fceeabc40" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 872.422613] env[62066]: DEBUG oslo_concurrency.lockutils [req-5a81dfe6-f9ee-4753-aa8a-d0144e77bf6d req-4acde67d-511e-4128-91e2-9ef962a1c68a service nova] Acquired lock "refresh_cache-c8e0d47c-4421-4e00-9183-206fceeabc40" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 872.422786] env[62066]: DEBUG nova.network.neutron [req-5a81dfe6-f9ee-4753-aa8a-d0144e77bf6d req-4acde67d-511e-4128-91e2-9ef962a1c68a service nova] [instance: c8e0d47c-4421-4e00-9183-206fceeabc40] Refreshing network info cache for port bb9924e0-8b18-43b2-bda0-3e409fa16718 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 872.504327] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1340996, 'name': CreateVM_Task} progress is 25%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.533091] env[62066]: INFO nova.compute.manager [None req-1dc5d31e-1ccf-4559-b370-76e0d5f04709 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 76306c3f-0674-40fe-9864-d82d11e4bb5e] Rebuilding instance [ 872.586595] env[62066]: DEBUG nova.compute.manager [None req-1dc5d31e-1ccf-4559-b370-76e0d5f04709 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 76306c3f-0674-40fe-9864-d82d11e4bb5e] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 872.587752] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c655e893-41a0-49d6-8234-d2f74f361313 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.674658] env[62066]: DEBUG oslo_concurrency.lockutils [None req-535b520b-dc22-40fb-a872-f0522e7e73a7 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Releasing lock "refresh_cache-3120c9a5-0558-4e83-b0fb-994a989398fb" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 872.675151] env[62066]: DEBUG nova.compute.manager [None req-535b520b-dc22-40fb-a872-f0522e7e73a7 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] [instance: 3120c9a5-0558-4e83-b0fb-994a989398fb] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 872.675407] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-535b520b-dc22-40fb-a872-f0522e7e73a7 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] [instance: 3120c9a5-0558-4e83-b0fb-994a989398fb] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 872.676318] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2658fd2-f81b-4b9c-b29d-1534eff150f6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.686020] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-535b520b-dc22-40fb-a872-f0522e7e73a7 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] [instance: 3120c9a5-0558-4e83-b0fb-994a989398fb] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 872.686020] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-abad2ba9-176a-4e19-8924-c14d70a8e1bb {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.694021] env[62066]: DEBUG oslo_vmware.api [None req-535b520b-dc22-40fb-a872-f0522e7e73a7 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Waiting for the task: (returnval){ [ 872.694021] env[62066]: value = "task-1340997" [ 872.694021] env[62066]: _type = "Task" [ 872.694021] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.701336] env[62066]: DEBUG oslo_vmware.api [None req-535b520b-dc22-40fb-a872-f0522e7e73a7 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Task: {'id': task-1340997, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.710165] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5327c80f-b6fd-46f0-ac98-3857740d94e2 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.954s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 872.712657] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ba2d4aa3-3165-47fd-a67a-a02fe7b33642 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.803s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 872.714686] env[62066]: DEBUG nova.objects.instance [None req-ba2d4aa3-3165-47fd-a67a-a02fe7b33642 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Lazy-loading 'resources' on Instance uuid 459f432b-e817-405f-9c3e-dfa5d957788c {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 872.733251] env[62066]: INFO nova.scheduler.client.report [None req-5327c80f-b6fd-46f0-ac98-3857740d94e2 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Deleted allocations for instance 1e1a2ede-4ad8-4600-851e-6f2046b3f919 [ 872.748169] env[62066]: DEBUG nova.network.neutron [None req-4c61eb51-497c-4460-951e-d696b73eff28 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 170e184a-f009-42cf-82dc-eb9696a00a24] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 872.779942] env[62066]: DEBUG nova.compute.manager [None req-4ccd1f86-3a5c-4e35-bdaa-b8e507c9718e tempest-ServersTestFqdnHostnames-1367628746 tempest-ServersTestFqdnHostnames-1367628746-project-member] [instance: afaca059-fc4e-4de2-8fa0-d2f226ea7051] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 872.806180] env[62066]: DEBUG nova.virt.hardware [None req-4ccd1f86-3a5c-4e35-bdaa-b8e507c9718e tempest-ServersTestFqdnHostnames-1367628746 tempest-ServersTestFqdnHostnames-1367628746-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-23T13:40:41Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-23T13:40:23Z,direct_url=,disk_format='vmdk',id=50ff584c-3b50-4395-af07-3e66769bc9f7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='eb52f7069a374c61ae946f052007c6d9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-23T13:40:24Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 872.806180] env[62066]: DEBUG nova.virt.hardware [None req-4ccd1f86-3a5c-4e35-bdaa-b8e507c9718e tempest-ServersTestFqdnHostnames-1367628746 tempest-ServersTestFqdnHostnames-1367628746-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 872.806180] env[62066]: DEBUG nova.virt.hardware [None req-4ccd1f86-3a5c-4e35-bdaa-b8e507c9718e tempest-ServersTestFqdnHostnames-1367628746 tempest-ServersTestFqdnHostnames-1367628746-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 872.806180] env[62066]: DEBUG nova.virt.hardware [None req-4ccd1f86-3a5c-4e35-bdaa-b8e507c9718e tempest-ServersTestFqdnHostnames-1367628746 tempest-ServersTestFqdnHostnames-1367628746-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 872.806180] env[62066]: DEBUG nova.virt.hardware [None req-4ccd1f86-3a5c-4e35-bdaa-b8e507c9718e tempest-ServersTestFqdnHostnames-1367628746 tempest-ServersTestFqdnHostnames-1367628746-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 872.806577] env[62066]: DEBUG nova.virt.hardware [None req-4ccd1f86-3a5c-4e35-bdaa-b8e507c9718e tempest-ServersTestFqdnHostnames-1367628746 tempest-ServersTestFqdnHostnames-1367628746-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 872.807774] env[62066]: DEBUG nova.virt.hardware [None req-4ccd1f86-3a5c-4e35-bdaa-b8e507c9718e tempest-ServersTestFqdnHostnames-1367628746 tempest-ServersTestFqdnHostnames-1367628746-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 872.808204] env[62066]: DEBUG nova.virt.hardware [None req-4ccd1f86-3a5c-4e35-bdaa-b8e507c9718e tempest-ServersTestFqdnHostnames-1367628746 tempest-ServersTestFqdnHostnames-1367628746-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 872.808570] env[62066]: DEBUG nova.virt.hardware [None req-4ccd1f86-3a5c-4e35-bdaa-b8e507c9718e tempest-ServersTestFqdnHostnames-1367628746 tempest-ServersTestFqdnHostnames-1367628746-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 872.809645] env[62066]: DEBUG nova.virt.hardware [None req-4ccd1f86-3a5c-4e35-bdaa-b8e507c9718e tempest-ServersTestFqdnHostnames-1367628746 tempest-ServersTestFqdnHostnames-1367628746-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 872.810304] env[62066]: DEBUG nova.virt.hardware [None req-4ccd1f86-3a5c-4e35-bdaa-b8e507c9718e tempest-ServersTestFqdnHostnames-1367628746 tempest-ServersTestFqdnHostnames-1367628746-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 872.812409] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ead9458-a9b9-436e-982c-471a64922dac {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.827449] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09b0c1fd-074e-4959-b6a3-4a90c5f1ec35 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.932914] env[62066]: DEBUG nova.network.neutron [None req-4c61eb51-497c-4460-951e-d696b73eff28 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 170e184a-f009-42cf-82dc-eb9696a00a24] Updating instance_info_cache with network_info: [{"id": "e8281131-9972-486f-b3be-2452487f99a3", "address": "fa:16:3e:05:9c:7a", "network": {"id": "9f3396ce-ee48-4eaf-8cc7-3e1711f37537", "bridge": "br-int", "label": "tempest-ServersTestJSON-2129775093-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6497ab02f327476d8ff81c2ecc0371e5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7ab8d568-adb0-4f3b-b6cc-68413e6546ae", "external-id": "nsx-vlan-transportzone-86", "segmentation_id": 86, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape8281131-99", "ovs_interfaceid": "e8281131-9972-486f-b3be-2452487f99a3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 872.999650] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1340996, 'name': CreateVM_Task} progress is 25%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.099075] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-1dc5d31e-1ccf-4559-b370-76e0d5f04709 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 76306c3f-0674-40fe-9864-d82d11e4bb5e] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 873.099407] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4ef75a76-7c1b-42e6-85e8-5b06990fc80c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.108400] env[62066]: DEBUG oslo_vmware.api [None req-1dc5d31e-1ccf-4559-b370-76e0d5f04709 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Waiting for the task: (returnval){ [ 873.108400] env[62066]: value = "task-1340998" [ 873.108400] env[62066]: _type = "Task" [ 873.108400] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.123920] env[62066]: DEBUG oslo_vmware.api [None req-1dc5d31e-1ccf-4559-b370-76e0d5f04709 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1340998, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.205161] env[62066]: DEBUG oslo_vmware.api [None req-535b520b-dc22-40fb-a872-f0522e7e73a7 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Task: {'id': task-1340997, 'name': PowerOffVM_Task, 'duration_secs': 0.190408} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.205635] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-535b520b-dc22-40fb-a872-f0522e7e73a7 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] [instance: 3120c9a5-0558-4e83-b0fb-994a989398fb] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 873.205939] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-535b520b-dc22-40fb-a872-f0522e7e73a7 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] [instance: 3120c9a5-0558-4e83-b0fb-994a989398fb] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 873.206331] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-876ba43e-e6ee-42ae-a5a1-77477afd6a47 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.233529] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-535b520b-dc22-40fb-a872-f0522e7e73a7 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] [instance: 3120c9a5-0558-4e83-b0fb-994a989398fb] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 873.233824] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-535b520b-dc22-40fb-a872-f0522e7e73a7 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] [instance: 3120c9a5-0558-4e83-b0fb-994a989398fb] Deleting contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 873.234125] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-535b520b-dc22-40fb-a872-f0522e7e73a7 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Deleting the datastore file [datastore2] 3120c9a5-0558-4e83-b0fb-994a989398fb {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 873.234469] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2fa043db-679b-403f-b345-07cd9d67a29d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.241617] env[62066]: DEBUG oslo_vmware.api [None req-535b520b-dc22-40fb-a872-f0522e7e73a7 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Waiting for the task: (returnval){ [ 873.241617] env[62066]: value = "task-1341000" [ 873.241617] env[62066]: _type = "Task" [ 873.241617] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.242096] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5327c80f-b6fd-46f0-ac98-3857740d94e2 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Lock "1e1a2ede-4ad8-4600-851e-6f2046b3f919" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 19.194s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 873.260465] env[62066]: DEBUG oslo_vmware.api [None req-535b520b-dc22-40fb-a872-f0522e7e73a7 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Task: {'id': task-1341000, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.316884] env[62066]: DEBUG nova.compute.manager [req-11a21f74-e72b-440c-93b5-0704c4635818 req-d72090e4-7375-4a20-9581-d7bd0a575076 service nova] [instance: 0bc31a8b-a5fc-4f6e-a4c0-024bf4260755] Received event network-changed-8f34a054-f99b-447b-89d6-ab1ef9bc3c76 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 873.317117] env[62066]: DEBUG nova.compute.manager [req-11a21f74-e72b-440c-93b5-0704c4635818 req-d72090e4-7375-4a20-9581-d7bd0a575076 service nova] [instance: 0bc31a8b-a5fc-4f6e-a4c0-024bf4260755] Refreshing instance network info cache due to event network-changed-8f34a054-f99b-447b-89d6-ab1ef9bc3c76. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 873.321639] env[62066]: DEBUG oslo_concurrency.lockutils [req-11a21f74-e72b-440c-93b5-0704c4635818 req-d72090e4-7375-4a20-9581-d7bd0a575076 service nova] Acquiring lock "refresh_cache-0bc31a8b-a5fc-4f6e-a4c0-024bf4260755" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 873.321639] env[62066]: DEBUG oslo_concurrency.lockutils [req-11a21f74-e72b-440c-93b5-0704c4635818 req-d72090e4-7375-4a20-9581-d7bd0a575076 service nova] Acquired lock "refresh_cache-0bc31a8b-a5fc-4f6e-a4c0-024bf4260755" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 873.321639] env[62066]: DEBUG nova.network.neutron [req-11a21f74-e72b-440c-93b5-0704c4635818 req-d72090e4-7375-4a20-9581-d7bd0a575076 service nova] [instance: 0bc31a8b-a5fc-4f6e-a4c0-024bf4260755] Refreshing network info cache for port 8f34a054-f99b-447b-89d6-ab1ef9bc3c76 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 873.350864] env[62066]: DEBUG nova.network.neutron [req-5a81dfe6-f9ee-4753-aa8a-d0144e77bf6d req-4acde67d-511e-4128-91e2-9ef962a1c68a service nova] [instance: c8e0d47c-4421-4e00-9183-206fceeabc40] Updated VIF entry in instance network info cache for port bb9924e0-8b18-43b2-bda0-3e409fa16718. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 873.351384] env[62066]: DEBUG nova.network.neutron [req-5a81dfe6-f9ee-4753-aa8a-d0144e77bf6d req-4acde67d-511e-4128-91e2-9ef962a1c68a service nova] [instance: c8e0d47c-4421-4e00-9183-206fceeabc40] Updating instance_info_cache with network_info: [{"id": "bb9924e0-8b18-43b2-bda0-3e409fa16718", "address": "fa:16:3e:b1:e4:41", "network": {"id": "5b89b2d4-71dd-4cd1-9edb-f4040be26540", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-283918837-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.165", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2252d0f6f18a47fd8a1ddad8020a6db7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "352165bb-004f-4180-9627-3a275dbe18af", "external-id": "nsx-vlan-transportzone-926", "segmentation_id": 926, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbb9924e0-8b", "ovs_interfaceid": "bb9924e0-8b18-43b2-bda0-3e409fa16718", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 873.439801] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4c61eb51-497c-4460-951e-d696b73eff28 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Releasing lock "refresh_cache-170e184a-f009-42cf-82dc-eb9696a00a24" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 873.441983] env[62066]: DEBUG nova.compute.manager [None req-4c61eb51-497c-4460-951e-d696b73eff28 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 170e184a-f009-42cf-82dc-eb9696a00a24] Instance network_info: |[{"id": "e8281131-9972-486f-b3be-2452487f99a3", "address": "fa:16:3e:05:9c:7a", "network": {"id": "9f3396ce-ee48-4eaf-8cc7-3e1711f37537", "bridge": "br-int", "label": "tempest-ServersTestJSON-2129775093-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6497ab02f327476d8ff81c2ecc0371e5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7ab8d568-adb0-4f3b-b6cc-68413e6546ae", "external-id": "nsx-vlan-transportzone-86", "segmentation_id": 86, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape8281131-99", "ovs_interfaceid": "e8281131-9972-486f-b3be-2452487f99a3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 873.441983] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-4c61eb51-497c-4460-951e-d696b73eff28 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 170e184a-f009-42cf-82dc-eb9696a00a24] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:05:9c:7a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7ab8d568-adb0-4f3b-b6cc-68413e6546ae', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e8281131-9972-486f-b3be-2452487f99a3', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 873.449530] env[62066]: DEBUG oslo.service.loopingcall [None req-4c61eb51-497c-4460-951e-d696b73eff28 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 873.449811] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 170e184a-f009-42cf-82dc-eb9696a00a24] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 873.450072] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c01b932e-af1e-4feb-b155-a8790bb18875 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.476025] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 873.476025] env[62066]: value = "task-1341001" [ 873.476025] env[62066]: _type = "Task" [ 873.476025] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.483364] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1341001, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.503817] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1340996, 'name': CreateVM_Task} progress is 25%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.533660] env[62066]: DEBUG nova.network.neutron [None req-4ccd1f86-3a5c-4e35-bdaa-b8e507c9718e tempest-ServersTestFqdnHostnames-1367628746 tempest-ServersTestFqdnHostnames-1367628746-project-member] [instance: afaca059-fc4e-4de2-8fa0-d2f226ea7051] Successfully updated port: 725b6138-b39f-41c3-bf84-1a13a3e30fa9 {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 873.583789] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a8416f8-92f8-4faa-ad39-93359608b314 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.593952] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84b61fed-e9c0-4689-9546-689a0abd25d5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.633640] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b3128e6-6d97-4bbf-a36f-e34997c6855f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.641432] env[62066]: DEBUG oslo_vmware.api [None req-1dc5d31e-1ccf-4559-b370-76e0d5f04709 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1340998, 'name': PowerOffVM_Task, 'duration_secs': 0.223661} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.643904] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-1dc5d31e-1ccf-4559-b370-76e0d5f04709 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 76306c3f-0674-40fe-9864-d82d11e4bb5e] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 873.644234] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-1dc5d31e-1ccf-4559-b370-76e0d5f04709 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 76306c3f-0674-40fe-9864-d82d11e4bb5e] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 873.645073] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38f2887f-d1b6-4fa5-a84e-15c5a7833713 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.648556] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df948e8c-10e7-4471-b068-8eefea724678 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.657081] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-1dc5d31e-1ccf-4559-b370-76e0d5f04709 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 76306c3f-0674-40fe-9864-d82d11e4bb5e] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 873.668183] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-460a7d6d-e09f-406b-9238-e64964748871 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.670591] env[62066]: DEBUG nova.compute.provider_tree [None req-ba2d4aa3-3165-47fd-a67a-a02fe7b33642 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 873.723748] env[62066]: DEBUG nova.compute.manager [req-afa4ac0a-bbe7-4fd2-8bef-cbe74a78cdfc req-7a82618a-fea5-4ec0-b73e-4ac8bd440f70 service nova] [instance: afaca059-fc4e-4de2-8fa0-d2f226ea7051] Received event network-vif-plugged-725b6138-b39f-41c3-bf84-1a13a3e30fa9 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 873.724060] env[62066]: DEBUG oslo_concurrency.lockutils [req-afa4ac0a-bbe7-4fd2-8bef-cbe74a78cdfc req-7a82618a-fea5-4ec0-b73e-4ac8bd440f70 service nova] Acquiring lock "afaca059-fc4e-4de2-8fa0-d2f226ea7051-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 873.724241] env[62066]: DEBUG oslo_concurrency.lockutils [req-afa4ac0a-bbe7-4fd2-8bef-cbe74a78cdfc req-7a82618a-fea5-4ec0-b73e-4ac8bd440f70 service nova] Lock "afaca059-fc4e-4de2-8fa0-d2f226ea7051-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 873.724421] env[62066]: DEBUG oslo_concurrency.lockutils [req-afa4ac0a-bbe7-4fd2-8bef-cbe74a78cdfc req-7a82618a-fea5-4ec0-b73e-4ac8bd440f70 service nova] Lock "afaca059-fc4e-4de2-8fa0-d2f226ea7051-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 873.724584] env[62066]: DEBUG nova.compute.manager [req-afa4ac0a-bbe7-4fd2-8bef-cbe74a78cdfc req-7a82618a-fea5-4ec0-b73e-4ac8bd440f70 service nova] [instance: afaca059-fc4e-4de2-8fa0-d2f226ea7051] No waiting events found dispatching network-vif-plugged-725b6138-b39f-41c3-bf84-1a13a3e30fa9 {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 873.725321] env[62066]: WARNING nova.compute.manager [req-afa4ac0a-bbe7-4fd2-8bef-cbe74a78cdfc req-7a82618a-fea5-4ec0-b73e-4ac8bd440f70 service nova] [instance: afaca059-fc4e-4de2-8fa0-d2f226ea7051] Received unexpected event network-vif-plugged-725b6138-b39f-41c3-bf84-1a13a3e30fa9 for instance with vm_state building and task_state spawning. [ 873.757164] env[62066]: DEBUG oslo_vmware.api [None req-535b520b-dc22-40fb-a872-f0522e7e73a7 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Task: {'id': task-1341000, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.11652} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.757680] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-535b520b-dc22-40fb-a872-f0522e7e73a7 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 873.757986] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-535b520b-dc22-40fb-a872-f0522e7e73a7 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] [instance: 3120c9a5-0558-4e83-b0fb-994a989398fb] Deleted contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 873.758293] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-535b520b-dc22-40fb-a872-f0522e7e73a7 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] [instance: 3120c9a5-0558-4e83-b0fb-994a989398fb] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 873.758576] env[62066]: INFO nova.compute.manager [None req-535b520b-dc22-40fb-a872-f0522e7e73a7 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] [instance: 3120c9a5-0558-4e83-b0fb-994a989398fb] Took 1.08 seconds to destroy the instance on the hypervisor. [ 873.758940] env[62066]: DEBUG oslo.service.loopingcall [None req-535b520b-dc22-40fb-a872-f0522e7e73a7 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 873.759841] env[62066]: DEBUG nova.compute.manager [-] [instance: 3120c9a5-0558-4e83-b0fb-994a989398fb] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 873.760016] env[62066]: DEBUG nova.network.neutron [-] [instance: 3120c9a5-0558-4e83-b0fb-994a989398fb] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 873.772300] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7b75bfaa-cfbe-4d54-8646-a77cb9e92e69 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Acquiring lock "1a4b0637-1a56-41ef-b89b-6b56d24ed206" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 873.772659] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7b75bfaa-cfbe-4d54-8646-a77cb9e92e69 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Lock "1a4b0637-1a56-41ef-b89b-6b56d24ed206" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 873.772813] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7b75bfaa-cfbe-4d54-8646-a77cb9e92e69 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Acquiring lock "1a4b0637-1a56-41ef-b89b-6b56d24ed206-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 873.773037] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7b75bfaa-cfbe-4d54-8646-a77cb9e92e69 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Lock "1a4b0637-1a56-41ef-b89b-6b56d24ed206-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 873.773276] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7b75bfaa-cfbe-4d54-8646-a77cb9e92e69 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Lock "1a4b0637-1a56-41ef-b89b-6b56d24ed206-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 873.776492] env[62066]: INFO nova.compute.manager [None req-7b75bfaa-cfbe-4d54-8646-a77cb9e92e69 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] [instance: 1a4b0637-1a56-41ef-b89b-6b56d24ed206] Terminating instance [ 873.778803] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-1dc5d31e-1ccf-4559-b370-76e0d5f04709 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 76306c3f-0674-40fe-9864-d82d11e4bb5e] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 873.779013] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-1dc5d31e-1ccf-4559-b370-76e0d5f04709 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 76306c3f-0674-40fe-9864-d82d11e4bb5e] Deleting contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 873.779290] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-1dc5d31e-1ccf-4559-b370-76e0d5f04709 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Deleting the datastore file [datastore2] 76306c3f-0674-40fe-9864-d82d11e4bb5e {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 873.779894] env[62066]: DEBUG nova.compute.manager [None req-7b75bfaa-cfbe-4d54-8646-a77cb9e92e69 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] [instance: 1a4b0637-1a56-41ef-b89b-6b56d24ed206] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 873.781025] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-7b75bfaa-cfbe-4d54-8646-a77cb9e92e69 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] [instance: 1a4b0637-1a56-41ef-b89b-6b56d24ed206] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 873.781025] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-00fe5f5b-1944-4b5f-9205-25421eca5d0a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.784124] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63fb61fb-7006-4dd4-a693-7b9290dd4e95 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.791022] env[62066]: DEBUG nova.network.neutron [-] [instance: 3120c9a5-0558-4e83-b0fb-994a989398fb] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 873.794974] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-7b75bfaa-cfbe-4d54-8646-a77cb9e92e69 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] [instance: 1a4b0637-1a56-41ef-b89b-6b56d24ed206] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 873.796920] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-603eb1a3-648c-4e00-a638-40873fe87379 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.798678] env[62066]: DEBUG oslo_vmware.api [None req-1dc5d31e-1ccf-4559-b370-76e0d5f04709 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Waiting for the task: (returnval){ [ 873.798678] env[62066]: value = "task-1341003" [ 873.798678] env[62066]: _type = "Task" [ 873.798678] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.805058] env[62066]: DEBUG oslo_vmware.api [None req-7b75bfaa-cfbe-4d54-8646-a77cb9e92e69 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Waiting for the task: (returnval){ [ 873.805058] env[62066]: value = "task-1341004" [ 873.805058] env[62066]: _type = "Task" [ 873.805058] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.819400] env[62066]: DEBUG oslo_vmware.api [None req-1dc5d31e-1ccf-4559-b370-76e0d5f04709 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1341003, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.829429] env[62066]: DEBUG oslo_vmware.api [None req-7b75bfaa-cfbe-4d54-8646-a77cb9e92e69 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Task: {'id': task-1341004, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.856150] env[62066]: DEBUG oslo_concurrency.lockutils [req-5a81dfe6-f9ee-4753-aa8a-d0144e77bf6d req-4acde67d-511e-4128-91e2-9ef962a1c68a service nova] Releasing lock "refresh_cache-c8e0d47c-4421-4e00-9183-206fceeabc40" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 873.856150] env[62066]: DEBUG nova.compute.manager [req-5a81dfe6-f9ee-4753-aa8a-d0144e77bf6d req-4acde67d-511e-4128-91e2-9ef962a1c68a service nova] [instance: 0bc31a8b-a5fc-4f6e-a4c0-024bf4260755] Received event network-changed-8f34a054-f99b-447b-89d6-ab1ef9bc3c76 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 873.856150] env[62066]: DEBUG nova.compute.manager [req-5a81dfe6-f9ee-4753-aa8a-d0144e77bf6d req-4acde67d-511e-4128-91e2-9ef962a1c68a service nova] [instance: 0bc31a8b-a5fc-4f6e-a4c0-024bf4260755] Refreshing instance network info cache due to event network-changed-8f34a054-f99b-447b-89d6-ab1ef9bc3c76. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 873.856150] env[62066]: DEBUG oslo_concurrency.lockutils [req-5a81dfe6-f9ee-4753-aa8a-d0144e77bf6d req-4acde67d-511e-4128-91e2-9ef962a1c68a service nova] Acquiring lock "refresh_cache-0bc31a8b-a5fc-4f6e-a4c0-024bf4260755" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 873.986820] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1341001, 'name': CreateVM_Task, 'duration_secs': 0.434383} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.988819] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 170e184a-f009-42cf-82dc-eb9696a00a24] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 873.988819] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4c61eb51-497c-4460-951e-d696b73eff28 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 873.988819] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4c61eb51-497c-4460-951e-d696b73eff28 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Acquired lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 873.988819] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4c61eb51-497c-4460-951e-d696b73eff28 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 873.989438] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-021f1473-a75b-4dff-9ef9-387b825eb227 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.997157] env[62066]: DEBUG oslo_vmware.api [None req-4c61eb51-497c-4460-951e-d696b73eff28 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Waiting for the task: (returnval){ [ 873.997157] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]52346759-7ccb-cd4d-9915-4c88eee3af93" [ 873.997157] env[62066]: _type = "Task" [ 873.997157] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.009923] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1340996, 'name': CreateVM_Task} progress is 25%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.014749] env[62066]: DEBUG oslo_vmware.api [None req-4c61eb51-497c-4460-951e-d696b73eff28 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52346759-7ccb-cd4d-9915-4c88eee3af93, 'name': SearchDatastore_Task, 'duration_secs': 0.010424} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.015100] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4c61eb51-497c-4460-951e-d696b73eff28 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Releasing lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 874.015389] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-4c61eb51-497c-4460-951e-d696b73eff28 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 170e184a-f009-42cf-82dc-eb9696a00a24] Processing image 50ff584c-3b50-4395-af07-3e66769bc9f7 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 874.015685] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4c61eb51-497c-4460-951e-d696b73eff28 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 874.015899] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4c61eb51-497c-4460-951e-d696b73eff28 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Acquired lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 874.016184] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-4c61eb51-497c-4460-951e-d696b73eff28 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 874.016508] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b5cfbade-08bb-4bae-bfaa-7c933b2a1c4b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.025688] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-4c61eb51-497c-4460-951e-d696b73eff28 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 874.025688] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-4c61eb51-497c-4460-951e-d696b73eff28 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 874.026681] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2f0c96f3-e1bb-41ed-b9f4-99b74a393812 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.033145] env[62066]: DEBUG oslo_vmware.api [None req-4c61eb51-497c-4460-951e-d696b73eff28 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Waiting for the task: (returnval){ [ 874.033145] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]52fcfb46-6009-8698-8d2e-d2ee84d253b0" [ 874.033145] env[62066]: _type = "Task" [ 874.033145] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.037838] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4ccd1f86-3a5c-4e35-bdaa-b8e507c9718e tempest-ServersTestFqdnHostnames-1367628746 tempest-ServersTestFqdnHostnames-1367628746-project-member] Acquiring lock "refresh_cache-afaca059-fc4e-4de2-8fa0-d2f226ea7051" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 874.038547] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4ccd1f86-3a5c-4e35-bdaa-b8e507c9718e tempest-ServersTestFqdnHostnames-1367628746 tempest-ServersTestFqdnHostnames-1367628746-project-member] Acquired lock "refresh_cache-afaca059-fc4e-4de2-8fa0-d2f226ea7051" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 874.038547] env[62066]: DEBUG nova.network.neutron [None req-4ccd1f86-3a5c-4e35-bdaa-b8e507c9718e tempest-ServersTestFqdnHostnames-1367628746 tempest-ServersTestFqdnHostnames-1367628746-project-member] [instance: afaca059-fc4e-4de2-8fa0-d2f226ea7051] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 874.045803] env[62066]: DEBUG oslo_vmware.api [None req-4c61eb51-497c-4460-951e-d696b73eff28 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52fcfb46-6009-8698-8d2e-d2ee84d253b0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.098124] env[62066]: DEBUG nova.network.neutron [req-11a21f74-e72b-440c-93b5-0704c4635818 req-d72090e4-7375-4a20-9581-d7bd0a575076 service nova] [instance: 0bc31a8b-a5fc-4f6e-a4c0-024bf4260755] Updated VIF entry in instance network info cache for port 8f34a054-f99b-447b-89d6-ab1ef9bc3c76. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 874.098285] env[62066]: DEBUG nova.network.neutron [req-11a21f74-e72b-440c-93b5-0704c4635818 req-d72090e4-7375-4a20-9581-d7bd0a575076 service nova] [instance: 0bc31a8b-a5fc-4f6e-a4c0-024bf4260755] Updating instance_info_cache with network_info: [{"id": "8f34a054-f99b-447b-89d6-ab1ef9bc3c76", "address": "fa:16:3e:06:43:d8", "network": {"id": "b92d2e33-767e-4958-b805-12e11b75c266", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-583710540-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.158", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "5281a37e471347b79e5dd6e70e03aa0f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "13e71dbb-4279-427c-b39d-ba5df9895e58", "external-id": "nsx-vlan-transportzone-417", "segmentation_id": 417, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8f34a054-f9", "ovs_interfaceid": "8f34a054-f99b-447b-89d6-ab1ef9bc3c76", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 874.174123] env[62066]: DEBUG nova.scheduler.client.report [None req-ba2d4aa3-3165-47fd-a67a-a02fe7b33642 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 874.291684] env[62066]: DEBUG nova.network.neutron [-] [instance: 3120c9a5-0558-4e83-b0fb-994a989398fb] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 874.312950] env[62066]: DEBUG oslo_vmware.api [None req-1dc5d31e-1ccf-4559-b370-76e0d5f04709 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1341003, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.174304} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.313568] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-1dc5d31e-1ccf-4559-b370-76e0d5f04709 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 874.313764] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-1dc5d31e-1ccf-4559-b370-76e0d5f04709 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 76306c3f-0674-40fe-9864-d82d11e4bb5e] Deleted contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 874.313949] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-1dc5d31e-1ccf-4559-b370-76e0d5f04709 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 76306c3f-0674-40fe-9864-d82d11e4bb5e] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 874.324021] env[62066]: DEBUG oslo_vmware.api [None req-7b75bfaa-cfbe-4d54-8646-a77cb9e92e69 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Task: {'id': task-1341004, 'name': PowerOffVM_Task, 'duration_secs': 0.199603} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.324021] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-7b75bfaa-cfbe-4d54-8646-a77cb9e92e69 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] [instance: 1a4b0637-1a56-41ef-b89b-6b56d24ed206] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 874.324021] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-7b75bfaa-cfbe-4d54-8646-a77cb9e92e69 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] [instance: 1a4b0637-1a56-41ef-b89b-6b56d24ed206] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 874.324021] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ea9794bb-dde7-4138-bd13-0f85db733215 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.382377] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-7b75bfaa-cfbe-4d54-8646-a77cb9e92e69 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] [instance: 1a4b0637-1a56-41ef-b89b-6b56d24ed206] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 874.382713] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-7b75bfaa-cfbe-4d54-8646-a77cb9e92e69 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] [instance: 1a4b0637-1a56-41ef-b89b-6b56d24ed206] Deleting contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 874.382914] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-7b75bfaa-cfbe-4d54-8646-a77cb9e92e69 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Deleting the datastore file [datastore2] 1a4b0637-1a56-41ef-b89b-6b56d24ed206 {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 874.383698] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-87970de9-ebbe-4d81-9871-c8e1a5346e91 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.390454] env[62066]: DEBUG oslo_vmware.api [None req-7b75bfaa-cfbe-4d54-8646-a77cb9e92e69 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Waiting for the task: (returnval){ [ 874.390454] env[62066]: value = "task-1341006" [ 874.390454] env[62066]: _type = "Task" [ 874.390454] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.398763] env[62066]: DEBUG oslo_vmware.api [None req-7b75bfaa-cfbe-4d54-8646-a77cb9e92e69 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Task: {'id': task-1341006, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.502932] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1340996, 'name': CreateVM_Task, 'duration_secs': 3.797554} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.506751] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2634fcc0-96bd-4513-8f92-515f83fe23d7] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 874.506751] env[62066]: DEBUG oslo_concurrency.lockutils [None req-59107a59-1a77-47be-8dae-1ec8f16f1fda tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 874.506751] env[62066]: DEBUG oslo_concurrency.lockutils [None req-59107a59-1a77-47be-8dae-1ec8f16f1fda tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Acquired lock "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 874.506751] env[62066]: DEBUG oslo_concurrency.lockutils [None req-59107a59-1a77-47be-8dae-1ec8f16f1fda tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 874.506751] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-eb2d76c1-1689-46ed-a0cc-8d19c642a44f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.514605] env[62066]: DEBUG oslo_vmware.api [None req-59107a59-1a77-47be-8dae-1ec8f16f1fda tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Waiting for the task: (returnval){ [ 874.514605] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]52616486-f25f-1dd0-25d4-39e723867fc3" [ 874.514605] env[62066]: _type = "Task" [ 874.514605] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.524975] env[62066]: DEBUG oslo_vmware.api [None req-59107a59-1a77-47be-8dae-1ec8f16f1fda tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52616486-f25f-1dd0-25d4-39e723867fc3, 'name': SearchDatastore_Task, 'duration_secs': 0.009186} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.525351] env[62066]: DEBUG oslo_concurrency.lockutils [None req-59107a59-1a77-47be-8dae-1ec8f16f1fda tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Releasing lock "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 874.525693] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-59107a59-1a77-47be-8dae-1ec8f16f1fda tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] [instance: 2634fcc0-96bd-4513-8f92-515f83fe23d7] Processing image 50ff584c-3b50-4395-af07-3e66769bc9f7 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 874.526046] env[62066]: DEBUG oslo_concurrency.lockutils [None req-59107a59-1a77-47be-8dae-1ec8f16f1fda tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 874.526332] env[62066]: DEBUG oslo_concurrency.lockutils [None req-59107a59-1a77-47be-8dae-1ec8f16f1fda tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Acquired lock "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 874.526620] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-59107a59-1a77-47be-8dae-1ec8f16f1fda tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 874.526987] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1146ddfe-a33e-4ce2-955b-b7718539c2c5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.547248] env[62066]: DEBUG oslo_vmware.api [None req-4c61eb51-497c-4460-951e-d696b73eff28 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52fcfb46-6009-8698-8d2e-d2ee84d253b0, 'name': SearchDatastore_Task, 'duration_secs': 0.010191} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.549767] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-59107a59-1a77-47be-8dae-1ec8f16f1fda tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 874.550122] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-59107a59-1a77-47be-8dae-1ec8f16f1fda tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 874.552172] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7d523704-c9cc-4bbf-a83c-4338bca17605 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.555616] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-16571c9b-03e9-4d0a-b6b3-8d20e043c496 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.564864] env[62066]: DEBUG oslo_vmware.api [None req-4c61eb51-497c-4460-951e-d696b73eff28 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Waiting for the task: (returnval){ [ 874.564864] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]5234f5e9-0d34-58ab-82a2-3a2d75a86f68" [ 874.564864] env[62066]: _type = "Task" [ 874.564864] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.564864] env[62066]: DEBUG oslo_vmware.api [None req-59107a59-1a77-47be-8dae-1ec8f16f1fda tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Waiting for the task: (returnval){ [ 874.564864] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]52aa44ea-9b82-6e33-2bd3-946cc50642b4" [ 874.564864] env[62066]: _type = "Task" [ 874.564864] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.581142] env[62066]: DEBUG oslo_vmware.api [None req-4c61eb51-497c-4460-951e-d696b73eff28 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]5234f5e9-0d34-58ab-82a2-3a2d75a86f68, 'name': SearchDatastore_Task, 'duration_secs': 0.010503} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.584448] env[62066]: DEBUG nova.network.neutron [None req-4ccd1f86-3a5c-4e35-bdaa-b8e507c9718e tempest-ServersTestFqdnHostnames-1367628746 tempest-ServersTestFqdnHostnames-1367628746-project-member] [instance: afaca059-fc4e-4de2-8fa0-d2f226ea7051] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 874.586331] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4c61eb51-497c-4460-951e-d696b73eff28 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Releasing lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 874.586683] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-4c61eb51-497c-4460-951e-d696b73eff28 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk to [datastore2] 170e184a-f009-42cf-82dc-eb9696a00a24/170e184a-f009-42cf-82dc-eb9696a00a24.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 874.587057] env[62066]: DEBUG oslo_vmware.api [None req-59107a59-1a77-47be-8dae-1ec8f16f1fda tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52aa44ea-9b82-6e33-2bd3-946cc50642b4, 'name': SearchDatastore_Task, 'duration_secs': 0.008685} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.587335] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6c86c934-b9e8-4407-bf8b-4c62ca3edc0b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.590475] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-43077231-f03d-489e-af8b-dff342c8a019 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.601149] env[62066]: DEBUG oslo_vmware.api [None req-59107a59-1a77-47be-8dae-1ec8f16f1fda tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Waiting for the task: (returnval){ [ 874.601149] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]521af06f-8fe2-863f-f14a-dd1dfa00d39c" [ 874.601149] env[62066]: _type = "Task" [ 874.601149] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.601149] env[62066]: DEBUG oslo_concurrency.lockutils [req-11a21f74-e72b-440c-93b5-0704c4635818 req-d72090e4-7375-4a20-9581-d7bd0a575076 service nova] Releasing lock "refresh_cache-0bc31a8b-a5fc-4f6e-a4c0-024bf4260755" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 874.601464] env[62066]: DEBUG nova.compute.manager [req-11a21f74-e72b-440c-93b5-0704c4635818 req-d72090e4-7375-4a20-9581-d7bd0a575076 service nova] [instance: 170e184a-f009-42cf-82dc-eb9696a00a24] Received event network-vif-plugged-e8281131-9972-486f-b3be-2452487f99a3 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 874.601749] env[62066]: DEBUG oslo_concurrency.lockutils [req-11a21f74-e72b-440c-93b5-0704c4635818 req-d72090e4-7375-4a20-9581-d7bd0a575076 service nova] Acquiring lock "170e184a-f009-42cf-82dc-eb9696a00a24-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 874.602169] env[62066]: DEBUG oslo_concurrency.lockutils [req-11a21f74-e72b-440c-93b5-0704c4635818 req-d72090e4-7375-4a20-9581-d7bd0a575076 service nova] Lock "170e184a-f009-42cf-82dc-eb9696a00a24-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 874.602771] env[62066]: DEBUG oslo_concurrency.lockutils [req-11a21f74-e72b-440c-93b5-0704c4635818 req-d72090e4-7375-4a20-9581-d7bd0a575076 service nova] Lock "170e184a-f009-42cf-82dc-eb9696a00a24-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 874.603165] env[62066]: DEBUG nova.compute.manager [req-11a21f74-e72b-440c-93b5-0704c4635818 req-d72090e4-7375-4a20-9581-d7bd0a575076 service nova] [instance: 170e184a-f009-42cf-82dc-eb9696a00a24] No waiting events found dispatching network-vif-plugged-e8281131-9972-486f-b3be-2452487f99a3 {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 874.603454] env[62066]: WARNING nova.compute.manager [req-11a21f74-e72b-440c-93b5-0704c4635818 req-d72090e4-7375-4a20-9581-d7bd0a575076 service nova] [instance: 170e184a-f009-42cf-82dc-eb9696a00a24] Received unexpected event network-vif-plugged-e8281131-9972-486f-b3be-2452487f99a3 for instance with vm_state building and task_state spawning. [ 874.603839] env[62066]: DEBUG nova.compute.manager [req-11a21f74-e72b-440c-93b5-0704c4635818 req-d72090e4-7375-4a20-9581-d7bd0a575076 service nova] [instance: 170e184a-f009-42cf-82dc-eb9696a00a24] Received event network-changed-e8281131-9972-486f-b3be-2452487f99a3 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 874.603994] env[62066]: DEBUG nova.compute.manager [req-11a21f74-e72b-440c-93b5-0704c4635818 req-d72090e4-7375-4a20-9581-d7bd0a575076 service nova] [instance: 170e184a-f009-42cf-82dc-eb9696a00a24] Refreshing instance network info cache due to event network-changed-e8281131-9972-486f-b3be-2452487f99a3. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 874.604306] env[62066]: DEBUG oslo_concurrency.lockutils [req-11a21f74-e72b-440c-93b5-0704c4635818 req-d72090e4-7375-4a20-9581-d7bd0a575076 service nova] Acquiring lock "refresh_cache-170e184a-f009-42cf-82dc-eb9696a00a24" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 874.604530] env[62066]: DEBUG oslo_concurrency.lockutils [req-11a21f74-e72b-440c-93b5-0704c4635818 req-d72090e4-7375-4a20-9581-d7bd0a575076 service nova] Acquired lock "refresh_cache-170e184a-f009-42cf-82dc-eb9696a00a24" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 874.604783] env[62066]: DEBUG nova.network.neutron [req-11a21f74-e72b-440c-93b5-0704c4635818 req-d72090e4-7375-4a20-9581-d7bd0a575076 service nova] [instance: 170e184a-f009-42cf-82dc-eb9696a00a24] Refreshing network info cache for port e8281131-9972-486f-b3be-2452487f99a3 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 874.607300] env[62066]: DEBUG oslo_vmware.api [None req-4c61eb51-497c-4460-951e-d696b73eff28 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Waiting for the task: (returnval){ [ 874.607300] env[62066]: value = "task-1341007" [ 874.607300] env[62066]: _type = "Task" [ 874.607300] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.611930] env[62066]: DEBUG oslo_concurrency.lockutils [req-5a81dfe6-f9ee-4753-aa8a-d0144e77bf6d req-4acde67d-511e-4128-91e2-9ef962a1c68a service nova] Acquired lock "refresh_cache-0bc31a8b-a5fc-4f6e-a4c0-024bf4260755" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 874.611930] env[62066]: DEBUG nova.network.neutron [req-5a81dfe6-f9ee-4753-aa8a-d0144e77bf6d req-4acde67d-511e-4128-91e2-9ef962a1c68a service nova] [instance: 0bc31a8b-a5fc-4f6e-a4c0-024bf4260755] Refreshing network info cache for port 8f34a054-f99b-447b-89d6-ab1ef9bc3c76 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 874.629889] env[62066]: DEBUG oslo_vmware.api [None req-4c61eb51-497c-4460-951e-d696b73eff28 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': task-1341007, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.630095] env[62066]: DEBUG oslo_vmware.api [None req-59107a59-1a77-47be-8dae-1ec8f16f1fda tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]521af06f-8fe2-863f-f14a-dd1dfa00d39c, 'name': SearchDatastore_Task, 'duration_secs': 0.010764} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.630350] env[62066]: DEBUG oslo_concurrency.lockutils [None req-59107a59-1a77-47be-8dae-1ec8f16f1fda tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Releasing lock "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 874.630629] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-59107a59-1a77-47be-8dae-1ec8f16f1fda tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk to [datastore1] 2634fcc0-96bd-4513-8f92-515f83fe23d7/2634fcc0-96bd-4513-8f92-515f83fe23d7.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 874.630929] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-069c102d-00f6-40cb-9fc1-9a72649ac4d3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.643053] env[62066]: DEBUG oslo_vmware.api [None req-59107a59-1a77-47be-8dae-1ec8f16f1fda tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Waiting for the task: (returnval){ [ 874.643053] env[62066]: value = "task-1341008" [ 874.643053] env[62066]: _type = "Task" [ 874.643053] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.655320] env[62066]: DEBUG oslo_vmware.api [None req-59107a59-1a77-47be-8dae-1ec8f16f1fda tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Task: {'id': task-1341008, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.679184] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ba2d4aa3-3165-47fd-a67a-a02fe7b33642 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.966s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 874.684363] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f0861cc1-6af6-4429-bc89-cb4b7f228490 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.570s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 874.690020] env[62066]: INFO nova.compute.claims [None req-f0861cc1-6af6-4429-bc89-cb4b7f228490 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 8807ce4d-532f-469c-a302-464c61c7efeb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 874.707470] env[62066]: INFO nova.scheduler.client.report [None req-ba2d4aa3-3165-47fd-a67a-a02fe7b33642 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Deleted allocations for instance 459f432b-e817-405f-9c3e-dfa5d957788c [ 874.772409] env[62066]: DEBUG nova.network.neutron [None req-4ccd1f86-3a5c-4e35-bdaa-b8e507c9718e tempest-ServersTestFqdnHostnames-1367628746 tempest-ServersTestFqdnHostnames-1367628746-project-member] [instance: afaca059-fc4e-4de2-8fa0-d2f226ea7051] Updating instance_info_cache with network_info: [{"id": "725b6138-b39f-41c3-bf84-1a13a3e30fa9", "address": "fa:16:3e:a3:87:a3", "network": {"id": "2d8ba809-853f-4d1b-b4b5-6d255c8a53f3", "bridge": "br-int", "label": "tempest-ServersTestFqdnHostnames-762924253-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "511b5cd53c7745eb8c3e9589b1de01c6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5fdd0624-2edb-4733-8284-225815c07f73", "external-id": "nsx-vlan-transportzone-330", "segmentation_id": 330, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap725b6138-b3", "ovs_interfaceid": "725b6138-b39f-41c3-bf84-1a13a3e30fa9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 874.794263] env[62066]: INFO nova.compute.manager [-] [instance: 3120c9a5-0558-4e83-b0fb-994a989398fb] Took 1.03 seconds to deallocate network for instance. [ 874.902761] env[62066]: DEBUG oslo_vmware.api [None req-7b75bfaa-cfbe-4d54-8646-a77cb9e92e69 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Task: {'id': task-1341006, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.15833} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.902761] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-7b75bfaa-cfbe-4d54-8646-a77cb9e92e69 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 874.902885] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-7b75bfaa-cfbe-4d54-8646-a77cb9e92e69 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] [instance: 1a4b0637-1a56-41ef-b89b-6b56d24ed206] Deleted contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 874.903079] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-7b75bfaa-cfbe-4d54-8646-a77cb9e92e69 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] [instance: 1a4b0637-1a56-41ef-b89b-6b56d24ed206] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 874.903257] env[62066]: INFO nova.compute.manager [None req-7b75bfaa-cfbe-4d54-8646-a77cb9e92e69 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] [instance: 1a4b0637-1a56-41ef-b89b-6b56d24ed206] Took 1.12 seconds to destroy the instance on the hypervisor. [ 874.903656] env[62066]: DEBUG oslo.service.loopingcall [None req-7b75bfaa-cfbe-4d54-8646-a77cb9e92e69 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 874.903799] env[62066]: DEBUG nova.compute.manager [-] [instance: 1a4b0637-1a56-41ef-b89b-6b56d24ed206] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 874.903799] env[62066]: DEBUG nova.network.neutron [-] [instance: 1a4b0637-1a56-41ef-b89b-6b56d24ed206] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 875.127768] env[62066]: DEBUG oslo_vmware.api [None req-4c61eb51-497c-4460-951e-d696b73eff28 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': task-1341007, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.156385] env[62066]: DEBUG oslo_vmware.api [None req-59107a59-1a77-47be-8dae-1ec8f16f1fda tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Task: {'id': task-1341008, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.218034] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ba2d4aa3-3165-47fd-a67a-a02fe7b33642 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Lock "459f432b-e817-405f-9c3e-dfa5d957788c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 13.265s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 875.275145] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4ccd1f86-3a5c-4e35-bdaa-b8e507c9718e tempest-ServersTestFqdnHostnames-1367628746 tempest-ServersTestFqdnHostnames-1367628746-project-member] Releasing lock "refresh_cache-afaca059-fc4e-4de2-8fa0-d2f226ea7051" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 875.275507] env[62066]: DEBUG nova.compute.manager [None req-4ccd1f86-3a5c-4e35-bdaa-b8e507c9718e tempest-ServersTestFqdnHostnames-1367628746 tempest-ServersTestFqdnHostnames-1367628746-project-member] [instance: afaca059-fc4e-4de2-8fa0-d2f226ea7051] Instance network_info: |[{"id": "725b6138-b39f-41c3-bf84-1a13a3e30fa9", "address": "fa:16:3e:a3:87:a3", "network": {"id": "2d8ba809-853f-4d1b-b4b5-6d255c8a53f3", "bridge": "br-int", "label": "tempest-ServersTestFqdnHostnames-762924253-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "511b5cd53c7745eb8c3e9589b1de01c6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5fdd0624-2edb-4733-8284-225815c07f73", "external-id": "nsx-vlan-transportzone-330", "segmentation_id": 330, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap725b6138-b3", "ovs_interfaceid": "725b6138-b39f-41c3-bf84-1a13a3e30fa9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 875.276027] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-4ccd1f86-3a5c-4e35-bdaa-b8e507c9718e tempest-ServersTestFqdnHostnames-1367628746 tempest-ServersTestFqdnHostnames-1367628746-project-member] [instance: afaca059-fc4e-4de2-8fa0-d2f226ea7051] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a3:87:a3', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '5fdd0624-2edb-4733-8284-225815c07f73', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '725b6138-b39f-41c3-bf84-1a13a3e30fa9', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 875.291231] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-4ccd1f86-3a5c-4e35-bdaa-b8e507c9718e tempest-ServersTestFqdnHostnames-1367628746 tempest-ServersTestFqdnHostnames-1367628746-project-member] Creating folder: Project (511b5cd53c7745eb8c3e9589b1de01c6). Parent ref: group-v285980. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 875.296036] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9251ee4f-cc16-419e-b2f9-7ba01e646311 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.301484] env[62066]: DEBUG oslo_concurrency.lockutils [None req-535b520b-dc22-40fb-a872-f0522e7e73a7 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 875.308914] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-4ccd1f86-3a5c-4e35-bdaa-b8e507c9718e tempest-ServersTestFqdnHostnames-1367628746 tempest-ServersTestFqdnHostnames-1367628746-project-member] Created folder: Project (511b5cd53c7745eb8c3e9589b1de01c6) in parent group-v285980. [ 875.309224] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-4ccd1f86-3a5c-4e35-bdaa-b8e507c9718e tempest-ServersTestFqdnHostnames-1367628746 tempest-ServersTestFqdnHostnames-1367628746-project-member] Creating folder: Instances. Parent ref: group-v286082. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 875.309635] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3b2faaf5-5a19-4fb6-b437-79ab73909a7f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.330582] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-4ccd1f86-3a5c-4e35-bdaa-b8e507c9718e tempest-ServersTestFqdnHostnames-1367628746 tempest-ServersTestFqdnHostnames-1367628746-project-member] Created folder: Instances in parent group-v286082. [ 875.330582] env[62066]: DEBUG oslo.service.loopingcall [None req-4ccd1f86-3a5c-4e35-bdaa-b8e507c9718e tempest-ServersTestFqdnHostnames-1367628746 tempest-ServersTestFqdnHostnames-1367628746-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 875.330582] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: afaca059-fc4e-4de2-8fa0-d2f226ea7051] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 875.333372] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-64bd6f80-ad88-40f4-b4f2-3df6715fa464 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.361865] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 875.361865] env[62066]: value = "task-1341011" [ 875.361865] env[62066]: _type = "Task" [ 875.361865] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.365017] env[62066]: DEBUG nova.virt.hardware [None req-1dc5d31e-1ccf-4559-b370-76e0d5f04709 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-23T13:40:41Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-23T13:40:23Z,direct_url=,disk_format='vmdk',id=50ff584c-3b50-4395-af07-3e66769bc9f7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='eb52f7069a374c61ae946f052007c6d9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-23T13:40:24Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 875.365017] env[62066]: DEBUG nova.virt.hardware [None req-1dc5d31e-1ccf-4559-b370-76e0d5f04709 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 875.365017] env[62066]: DEBUG nova.virt.hardware [None req-1dc5d31e-1ccf-4559-b370-76e0d5f04709 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 875.365518] env[62066]: DEBUG nova.virt.hardware [None req-1dc5d31e-1ccf-4559-b370-76e0d5f04709 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 875.365518] env[62066]: DEBUG nova.virt.hardware [None req-1dc5d31e-1ccf-4559-b370-76e0d5f04709 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 875.365518] env[62066]: DEBUG nova.virt.hardware [None req-1dc5d31e-1ccf-4559-b370-76e0d5f04709 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 875.365717] env[62066]: DEBUG nova.virt.hardware [None req-1dc5d31e-1ccf-4559-b370-76e0d5f04709 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 875.365877] env[62066]: DEBUG nova.virt.hardware [None req-1dc5d31e-1ccf-4559-b370-76e0d5f04709 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 875.366061] env[62066]: DEBUG nova.virt.hardware [None req-1dc5d31e-1ccf-4559-b370-76e0d5f04709 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 875.366234] env[62066]: DEBUG nova.virt.hardware [None req-1dc5d31e-1ccf-4559-b370-76e0d5f04709 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 875.366411] env[62066]: DEBUG nova.virt.hardware [None req-1dc5d31e-1ccf-4559-b370-76e0d5f04709 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 875.367458] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38e199e1-8d6c-4608-a97a-d1e4766fec21 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.385778] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3769f17a-cf2c-4d7b-ad2a-2d31a2e9dc21 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.390237] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1341011, 'name': CreateVM_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.402645] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-1dc5d31e-1ccf-4559-b370-76e0d5f04709 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 76306c3f-0674-40fe-9864-d82d11e4bb5e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:73:9e:3c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a965790c-2d2f-4c2a-9ee7-745f4d53039b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '28f3b624-430a-474d-a24b-9e54db92e9e2', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 875.411000] env[62066]: DEBUG oslo.service.loopingcall [None req-1dc5d31e-1ccf-4559-b370-76e0d5f04709 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 875.411460] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 76306c3f-0674-40fe-9864-d82d11e4bb5e] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 875.411796] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-137b4191-149d-4d54-be7e-185c155efb27 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.437824] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 875.437824] env[62066]: value = "task-1341012" [ 875.437824] env[62066]: _type = "Task" [ 875.437824] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.445896] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1341012, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.472559] env[62066]: DEBUG nova.network.neutron [req-5a81dfe6-f9ee-4753-aa8a-d0144e77bf6d req-4acde67d-511e-4128-91e2-9ef962a1c68a service nova] [instance: 0bc31a8b-a5fc-4f6e-a4c0-024bf4260755] Updated VIF entry in instance network info cache for port 8f34a054-f99b-447b-89d6-ab1ef9bc3c76. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 875.473122] env[62066]: DEBUG nova.network.neutron [req-5a81dfe6-f9ee-4753-aa8a-d0144e77bf6d req-4acde67d-511e-4128-91e2-9ef962a1c68a service nova] [instance: 0bc31a8b-a5fc-4f6e-a4c0-024bf4260755] Updating instance_info_cache with network_info: [{"id": "8f34a054-f99b-447b-89d6-ab1ef9bc3c76", "address": "fa:16:3e:06:43:d8", "network": {"id": "b92d2e33-767e-4958-b805-12e11b75c266", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-583710540-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "5281a37e471347b79e5dd6e70e03aa0f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "13e71dbb-4279-427c-b39d-ba5df9895e58", "external-id": "nsx-vlan-transportzone-417", "segmentation_id": 417, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8f34a054-f9", "ovs_interfaceid": "8f34a054-f99b-447b-89d6-ab1ef9bc3c76", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 875.629123] env[62066]: DEBUG oslo_vmware.api [None req-4c61eb51-497c-4460-951e-d696b73eff28 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': task-1341007, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.56667} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.629123] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-4c61eb51-497c-4460-951e-d696b73eff28 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk to [datastore2] 170e184a-f009-42cf-82dc-eb9696a00a24/170e184a-f009-42cf-82dc-eb9696a00a24.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 875.629123] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-4c61eb51-497c-4460-951e-d696b73eff28 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 170e184a-f009-42cf-82dc-eb9696a00a24] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 875.629123] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e6889b77-b311-4c31-9f17-271add9361d8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.633244] env[62066]: DEBUG oslo_concurrency.lockutils [None req-98bb85f9-df22-41ab-a6e4-7c41686713f5 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Acquiring lock "de2e2be7-efdb-45a8-842a-640ab9deb1d9" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 875.633514] env[62066]: DEBUG oslo_concurrency.lockutils [None req-98bb85f9-df22-41ab-a6e4-7c41686713f5 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Lock "de2e2be7-efdb-45a8-842a-640ab9deb1d9" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 875.636740] env[62066]: DEBUG oslo_vmware.api [None req-4c61eb51-497c-4460-951e-d696b73eff28 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Waiting for the task: (returnval){ [ 875.636740] env[62066]: value = "task-1341013" [ 875.636740] env[62066]: _type = "Task" [ 875.636740] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.641980] env[62066]: DEBUG nova.network.neutron [req-11a21f74-e72b-440c-93b5-0704c4635818 req-d72090e4-7375-4a20-9581-d7bd0a575076 service nova] [instance: 170e184a-f009-42cf-82dc-eb9696a00a24] Updated VIF entry in instance network info cache for port e8281131-9972-486f-b3be-2452487f99a3. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 875.642351] env[62066]: DEBUG nova.network.neutron [req-11a21f74-e72b-440c-93b5-0704c4635818 req-d72090e4-7375-4a20-9581-d7bd0a575076 service nova] [instance: 170e184a-f009-42cf-82dc-eb9696a00a24] Updating instance_info_cache with network_info: [{"id": "e8281131-9972-486f-b3be-2452487f99a3", "address": "fa:16:3e:05:9c:7a", "network": {"id": "9f3396ce-ee48-4eaf-8cc7-3e1711f37537", "bridge": "br-int", "label": "tempest-ServersTestJSON-2129775093-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6497ab02f327476d8ff81c2ecc0371e5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7ab8d568-adb0-4f3b-b6cc-68413e6546ae", "external-id": "nsx-vlan-transportzone-86", "segmentation_id": 86, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape8281131-99", "ovs_interfaceid": "e8281131-9972-486f-b3be-2452487f99a3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 875.651071] env[62066]: DEBUG oslo_vmware.api [None req-4c61eb51-497c-4460-951e-d696b73eff28 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': task-1341013, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.662448] env[62066]: DEBUG oslo_vmware.api [None req-59107a59-1a77-47be-8dae-1ec8f16f1fda tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Task: {'id': task-1341008, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.841208} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.662830] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-59107a59-1a77-47be-8dae-1ec8f16f1fda tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk to [datastore1] 2634fcc0-96bd-4513-8f92-515f83fe23d7/2634fcc0-96bd-4513-8f92-515f83fe23d7.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 875.663041] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-59107a59-1a77-47be-8dae-1ec8f16f1fda tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] [instance: 2634fcc0-96bd-4513-8f92-515f83fe23d7] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 875.663291] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7ddd5ead-6b95-4a0d-917f-c8e63945ab21 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.675494] env[62066]: DEBUG oslo_vmware.api [None req-59107a59-1a77-47be-8dae-1ec8f16f1fda tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Waiting for the task: (returnval){ [ 875.675494] env[62066]: value = "task-1341014" [ 875.675494] env[62066]: _type = "Task" [ 875.675494] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.689724] env[62066]: DEBUG oslo_vmware.api [None req-59107a59-1a77-47be-8dae-1ec8f16f1fda tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Task: {'id': task-1341014, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.819869] env[62066]: DEBUG nova.network.neutron [-] [instance: 1a4b0637-1a56-41ef-b89b-6b56d24ed206] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 875.878216] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1341011, 'name': CreateVM_Task, 'duration_secs': 0.501856} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.878434] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: afaca059-fc4e-4de2-8fa0-d2f226ea7051] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 875.879598] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4ccd1f86-3a5c-4e35-bdaa-b8e507c9718e tempest-ServersTestFqdnHostnames-1367628746 tempest-ServersTestFqdnHostnames-1367628746-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 875.879816] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4ccd1f86-3a5c-4e35-bdaa-b8e507c9718e tempest-ServersTestFqdnHostnames-1367628746 tempest-ServersTestFqdnHostnames-1367628746-project-member] Acquired lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 875.880185] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4ccd1f86-3a5c-4e35-bdaa-b8e507c9718e tempest-ServersTestFqdnHostnames-1367628746 tempest-ServersTestFqdnHostnames-1367628746-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 875.880604] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ae91adfd-d077-4498-a809-44fa77c83daf {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.887717] env[62066]: DEBUG oslo_vmware.api [None req-4ccd1f86-3a5c-4e35-bdaa-b8e507c9718e tempest-ServersTestFqdnHostnames-1367628746 tempest-ServersTestFqdnHostnames-1367628746-project-member] Waiting for the task: (returnval){ [ 875.887717] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]5207f687-9ec1-7adf-c759-e78e666b8798" [ 875.887717] env[62066]: _type = "Task" [ 875.887717] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.895655] env[62066]: DEBUG oslo_vmware.api [None req-4ccd1f86-3a5c-4e35-bdaa-b8e507c9718e tempest-ServersTestFqdnHostnames-1367628746 tempest-ServersTestFqdnHostnames-1367628746-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]5207f687-9ec1-7adf-c759-e78e666b8798, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.918729] env[62066]: DEBUG nova.compute.manager [req-fe7cbcbc-8b8d-46e0-8190-3a7fdc4aebcc req-cb7213dc-0c3e-4465-876d-6a6b44359bf9 service nova] [instance: 0bc31a8b-a5fc-4f6e-a4c0-024bf4260755] Received event network-changed-8f34a054-f99b-447b-89d6-ab1ef9bc3c76 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 875.918931] env[62066]: DEBUG nova.compute.manager [req-fe7cbcbc-8b8d-46e0-8190-3a7fdc4aebcc req-cb7213dc-0c3e-4465-876d-6a6b44359bf9 service nova] [instance: 0bc31a8b-a5fc-4f6e-a4c0-024bf4260755] Refreshing instance network info cache due to event network-changed-8f34a054-f99b-447b-89d6-ab1ef9bc3c76. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 875.919149] env[62066]: DEBUG oslo_concurrency.lockutils [req-fe7cbcbc-8b8d-46e0-8190-3a7fdc4aebcc req-cb7213dc-0c3e-4465-876d-6a6b44359bf9 service nova] Acquiring lock "refresh_cache-0bc31a8b-a5fc-4f6e-a4c0-024bf4260755" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 875.932859] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3aa5b41-f1a8-4cb7-a1e1-760a4e5178f7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.945026] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9855cf48-82af-4256-9bb4-e0bf831df854 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.953348] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1341012, 'name': CreateVM_Task, 'duration_secs': 0.395268} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.977216] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 76306c3f-0674-40fe-9864-d82d11e4bb5e] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 875.978183] env[62066]: DEBUG oslo_concurrency.lockutils [req-5a81dfe6-f9ee-4753-aa8a-d0144e77bf6d req-4acde67d-511e-4128-91e2-9ef962a1c68a service nova] Releasing lock "refresh_cache-0bc31a8b-a5fc-4f6e-a4c0-024bf4260755" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 875.979261] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1dc5d31e-1ccf-4559-b370-76e0d5f04709 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 875.979517] env[62066]: DEBUG oslo_concurrency.lockutils [req-fe7cbcbc-8b8d-46e0-8190-3a7fdc4aebcc req-cb7213dc-0c3e-4465-876d-6a6b44359bf9 service nova] Acquired lock "refresh_cache-0bc31a8b-a5fc-4f6e-a4c0-024bf4260755" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 875.979724] env[62066]: DEBUG nova.network.neutron [req-fe7cbcbc-8b8d-46e0-8190-3a7fdc4aebcc req-cb7213dc-0c3e-4465-876d-6a6b44359bf9 service nova] [instance: 0bc31a8b-a5fc-4f6e-a4c0-024bf4260755] Refreshing network info cache for port 8f34a054-f99b-447b-89d6-ab1ef9bc3c76 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 875.981736] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c6f5d70-516e-42da-990c-897a88bf8c4d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.989856] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea5b4316-92c8-4e01-ba23-10873f6410c0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.005844] env[62066]: DEBUG nova.compute.provider_tree [None req-f0861cc1-6af6-4429-bc89-cb4b7f228490 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 876.139180] env[62066]: DEBUG nova.compute.utils [None req-98bb85f9-df22-41ab-a6e4-7c41686713f5 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 876.144933] env[62066]: DEBUG oslo_concurrency.lockutils [req-11a21f74-e72b-440c-93b5-0704c4635818 req-d72090e4-7375-4a20-9581-d7bd0a575076 service nova] Releasing lock "refresh_cache-170e184a-f009-42cf-82dc-eb9696a00a24" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 876.150665] env[62066]: DEBUG oslo_vmware.api [None req-4c61eb51-497c-4460-951e-d696b73eff28 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': task-1341013, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.087084} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.151577] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-4c61eb51-497c-4460-951e-d696b73eff28 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 170e184a-f009-42cf-82dc-eb9696a00a24] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 876.152393] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5564946-7c1d-4acd-b83f-4d6f89b70157 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.176390] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-4c61eb51-497c-4460-951e-d696b73eff28 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 170e184a-f009-42cf-82dc-eb9696a00a24] Reconfiguring VM instance instance-00000052 to attach disk [datastore2] 170e184a-f009-42cf-82dc-eb9696a00a24/170e184a-f009-42cf-82dc-eb9696a00a24.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 876.177852] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4096dc23-d471-4c7f-b076-4d854e16c3b4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.200876] env[62066]: DEBUG oslo_vmware.api [None req-59107a59-1a77-47be-8dae-1ec8f16f1fda tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Task: {'id': task-1341014, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.081291} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.202279] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-59107a59-1a77-47be-8dae-1ec8f16f1fda tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] [instance: 2634fcc0-96bd-4513-8f92-515f83fe23d7] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 876.202643] env[62066]: DEBUG oslo_vmware.api [None req-4c61eb51-497c-4460-951e-d696b73eff28 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Waiting for the task: (returnval){ [ 876.202643] env[62066]: value = "task-1341015" [ 876.202643] env[62066]: _type = "Task" [ 876.202643] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.203364] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bfbaf4c-98dd-4610-aa50-8d394b108d43 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.232255] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-59107a59-1a77-47be-8dae-1ec8f16f1fda tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] [instance: 2634fcc0-96bd-4513-8f92-515f83fe23d7] Reconfiguring VM instance instance-00000051 to attach disk [datastore1] 2634fcc0-96bd-4513-8f92-515f83fe23d7/2634fcc0-96bd-4513-8f92-515f83fe23d7.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 876.232563] env[62066]: DEBUG oslo_vmware.api [None req-4c61eb51-497c-4460-951e-d696b73eff28 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': task-1341015, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.232801] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f3b77d3b-80cf-4721-84c4-9b61f8d61b9e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.252469] env[62066]: DEBUG oslo_vmware.api [None req-59107a59-1a77-47be-8dae-1ec8f16f1fda tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Waiting for the task: (returnval){ [ 876.252469] env[62066]: value = "task-1341016" [ 876.252469] env[62066]: _type = "Task" [ 876.252469] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.262151] env[62066]: DEBUG oslo_vmware.api [None req-59107a59-1a77-47be-8dae-1ec8f16f1fda tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Task: {'id': task-1341016, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.323242] env[62066]: INFO nova.compute.manager [-] [instance: 1a4b0637-1a56-41ef-b89b-6b56d24ed206] Took 1.42 seconds to deallocate network for instance. [ 876.330976] env[62066]: DEBUG nova.compute.manager [req-d66c920c-b1a6-44b0-b07a-49d72edaedab req-77a82869-327c-4d6f-a1f2-04981e33dd63 service nova] [instance: afaca059-fc4e-4de2-8fa0-d2f226ea7051] Received event network-changed-725b6138-b39f-41c3-bf84-1a13a3e30fa9 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 876.331356] env[62066]: DEBUG nova.compute.manager [req-d66c920c-b1a6-44b0-b07a-49d72edaedab req-77a82869-327c-4d6f-a1f2-04981e33dd63 service nova] [instance: afaca059-fc4e-4de2-8fa0-d2f226ea7051] Refreshing instance network info cache due to event network-changed-725b6138-b39f-41c3-bf84-1a13a3e30fa9. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 876.331515] env[62066]: DEBUG oslo_concurrency.lockutils [req-d66c920c-b1a6-44b0-b07a-49d72edaedab req-77a82869-327c-4d6f-a1f2-04981e33dd63 service nova] Acquiring lock "refresh_cache-afaca059-fc4e-4de2-8fa0-d2f226ea7051" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 876.331719] env[62066]: DEBUG oslo_concurrency.lockutils [req-d66c920c-b1a6-44b0-b07a-49d72edaedab req-77a82869-327c-4d6f-a1f2-04981e33dd63 service nova] Acquired lock "refresh_cache-afaca059-fc4e-4de2-8fa0-d2f226ea7051" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 876.332071] env[62066]: DEBUG nova.network.neutron [req-d66c920c-b1a6-44b0-b07a-49d72edaedab req-77a82869-327c-4d6f-a1f2-04981e33dd63 service nova] [instance: afaca059-fc4e-4de2-8fa0-d2f226ea7051] Refreshing network info cache for port 725b6138-b39f-41c3-bf84-1a13a3e30fa9 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 876.397775] env[62066]: DEBUG oslo_vmware.api [None req-4ccd1f86-3a5c-4e35-bdaa-b8e507c9718e tempest-ServersTestFqdnHostnames-1367628746 tempest-ServersTestFqdnHostnames-1367628746-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]5207f687-9ec1-7adf-c759-e78e666b8798, 'name': SearchDatastore_Task, 'duration_secs': 0.010288} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.398117] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4ccd1f86-3a5c-4e35-bdaa-b8e507c9718e tempest-ServersTestFqdnHostnames-1367628746 tempest-ServersTestFqdnHostnames-1367628746-project-member] Releasing lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 876.398394] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-4ccd1f86-3a5c-4e35-bdaa-b8e507c9718e tempest-ServersTestFqdnHostnames-1367628746 tempest-ServersTestFqdnHostnames-1367628746-project-member] [instance: afaca059-fc4e-4de2-8fa0-d2f226ea7051] Processing image 50ff584c-3b50-4395-af07-3e66769bc9f7 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 876.398611] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4ccd1f86-3a5c-4e35-bdaa-b8e507c9718e tempest-ServersTestFqdnHostnames-1367628746 tempest-ServersTestFqdnHostnames-1367628746-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 876.398760] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4ccd1f86-3a5c-4e35-bdaa-b8e507c9718e tempest-ServersTestFqdnHostnames-1367628746 tempest-ServersTestFqdnHostnames-1367628746-project-member] Acquired lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 876.398948] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-4ccd1f86-3a5c-4e35-bdaa-b8e507c9718e tempest-ServersTestFqdnHostnames-1367628746 tempest-ServersTestFqdnHostnames-1367628746-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 876.399282] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1dc5d31e-1ccf-4559-b370-76e0d5f04709 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Acquired lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 876.399628] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1dc5d31e-1ccf-4559-b370-76e0d5f04709 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 876.400180] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ddba9bbb-c71f-4287-8665-7765fb365e34 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.402043] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6df5ddd3-2827-475f-8af9-a90f55ad699f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.406667] env[62066]: DEBUG oslo_vmware.api [None req-1dc5d31e-1ccf-4559-b370-76e0d5f04709 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Waiting for the task: (returnval){ [ 876.406667] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]52fd779d-c4c5-29b1-2f91-7019614bdda1" [ 876.406667] env[62066]: _type = "Task" [ 876.406667] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.410941] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-4ccd1f86-3a5c-4e35-bdaa-b8e507c9718e tempest-ServersTestFqdnHostnames-1367628746 tempest-ServersTestFqdnHostnames-1367628746-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 876.411187] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-4ccd1f86-3a5c-4e35-bdaa-b8e507c9718e tempest-ServersTestFqdnHostnames-1367628746 tempest-ServersTestFqdnHostnames-1367628746-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 876.414528] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-84bc7f42-7612-43f3-bb05-d533f0365bfb {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.416679] env[62066]: DEBUG oslo_vmware.api [None req-1dc5d31e-1ccf-4559-b370-76e0d5f04709 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52fd779d-c4c5-29b1-2f91-7019614bdda1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.419870] env[62066]: DEBUG oslo_vmware.api [None req-4ccd1f86-3a5c-4e35-bdaa-b8e507c9718e tempest-ServersTestFqdnHostnames-1367628746 tempest-ServersTestFqdnHostnames-1367628746-project-member] Waiting for the task: (returnval){ [ 876.419870] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]52ae71a6-d1b4-fd54-2cbb-59ebcc47474d" [ 876.419870] env[62066]: _type = "Task" [ 876.419870] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.429364] env[62066]: DEBUG oslo_vmware.api [None req-4ccd1f86-3a5c-4e35-bdaa-b8e507c9718e tempest-ServersTestFqdnHostnames-1367628746 tempest-ServersTestFqdnHostnames-1367628746-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52ae71a6-d1b4-fd54-2cbb-59ebcc47474d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.509359] env[62066]: DEBUG nova.scheduler.client.report [None req-f0861cc1-6af6-4429-bc89-cb4b7f228490 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 876.645425] env[62066]: DEBUG oslo_concurrency.lockutils [None req-98bb85f9-df22-41ab-a6e4-7c41686713f5 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Lock "de2e2be7-efdb-45a8-842a-640ab9deb1d9" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.012s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 876.716409] env[62066]: DEBUG oslo_vmware.api [None req-4c61eb51-497c-4460-951e-d696b73eff28 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': task-1341015, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.764971] env[62066]: DEBUG oslo_vmware.api [None req-59107a59-1a77-47be-8dae-1ec8f16f1fda tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Task: {'id': task-1341016, 'name': ReconfigVM_Task, 'duration_secs': 0.408111} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.765287] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-59107a59-1a77-47be-8dae-1ec8f16f1fda tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] [instance: 2634fcc0-96bd-4513-8f92-515f83fe23d7] Reconfigured VM instance instance-00000051 to attach disk [datastore1] 2634fcc0-96bd-4513-8f92-515f83fe23d7/2634fcc0-96bd-4513-8f92-515f83fe23d7.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 876.766183] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-71d2ca23-03ef-4ba3-8aff-9dfb1ee0543e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.776083] env[62066]: DEBUG oslo_vmware.api [None req-59107a59-1a77-47be-8dae-1ec8f16f1fda tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Waiting for the task: (returnval){ [ 876.776083] env[62066]: value = "task-1341017" [ 876.776083] env[62066]: _type = "Task" [ 876.776083] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.786572] env[62066]: DEBUG oslo_vmware.api [None req-59107a59-1a77-47be-8dae-1ec8f16f1fda tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Task: {'id': task-1341017, 'name': Rename_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.826884] env[62066]: DEBUG nova.network.neutron [req-fe7cbcbc-8b8d-46e0-8190-3a7fdc4aebcc req-cb7213dc-0c3e-4465-876d-6a6b44359bf9 service nova] [instance: 0bc31a8b-a5fc-4f6e-a4c0-024bf4260755] Updated VIF entry in instance network info cache for port 8f34a054-f99b-447b-89d6-ab1ef9bc3c76. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 876.827277] env[62066]: DEBUG nova.network.neutron [req-fe7cbcbc-8b8d-46e0-8190-3a7fdc4aebcc req-cb7213dc-0c3e-4465-876d-6a6b44359bf9 service nova] [instance: 0bc31a8b-a5fc-4f6e-a4c0-024bf4260755] Updating instance_info_cache with network_info: [{"id": "8f34a054-f99b-447b-89d6-ab1ef9bc3c76", "address": "fa:16:3e:06:43:d8", "network": {"id": "b92d2e33-767e-4958-b805-12e11b75c266", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-583710540-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "5281a37e471347b79e5dd6e70e03aa0f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "13e71dbb-4279-427c-b39d-ba5df9895e58", "external-id": "nsx-vlan-transportzone-417", "segmentation_id": 417, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8f34a054-f9", "ovs_interfaceid": "8f34a054-f99b-447b-89d6-ab1ef9bc3c76", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 876.835440] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7b75bfaa-cfbe-4d54-8646-a77cb9e92e69 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 876.920223] env[62066]: DEBUG oslo_vmware.api [None req-1dc5d31e-1ccf-4559-b370-76e0d5f04709 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52fd779d-c4c5-29b1-2f91-7019614bdda1, 'name': SearchDatastore_Task, 'duration_secs': 0.011051} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.920610] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1dc5d31e-1ccf-4559-b370-76e0d5f04709 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Releasing lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 876.920852] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-1dc5d31e-1ccf-4559-b370-76e0d5f04709 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 76306c3f-0674-40fe-9864-d82d11e4bb5e] Processing image 50ff584c-3b50-4395-af07-3e66769bc9f7 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 876.921084] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1dc5d31e-1ccf-4559-b370-76e0d5f04709 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 876.931235] env[62066]: DEBUG oslo_vmware.api [None req-4ccd1f86-3a5c-4e35-bdaa-b8e507c9718e tempest-ServersTestFqdnHostnames-1367628746 tempest-ServersTestFqdnHostnames-1367628746-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52ae71a6-d1b4-fd54-2cbb-59ebcc47474d, 'name': SearchDatastore_Task, 'duration_secs': 0.010498} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.932100] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-88d0876f-cde5-4c16-b994-ab3591e4badb {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.941944] env[62066]: DEBUG oslo_vmware.api [None req-4ccd1f86-3a5c-4e35-bdaa-b8e507c9718e tempest-ServersTestFqdnHostnames-1367628746 tempest-ServersTestFqdnHostnames-1367628746-project-member] Waiting for the task: (returnval){ [ 876.941944] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]52cfa93e-6430-b919-a596-f8ee9b137b37" [ 876.941944] env[62066]: _type = "Task" [ 876.941944] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.950388] env[62066]: DEBUG oslo_concurrency.lockutils [None req-09aaefa9-33e0-4c34-9b0e-8039240cf379 tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] Acquiring lock "0bc31a8b-a5fc-4f6e-a4c0-024bf4260755" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 876.950629] env[62066]: DEBUG oslo_concurrency.lockutils [None req-09aaefa9-33e0-4c34-9b0e-8039240cf379 tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] Lock "0bc31a8b-a5fc-4f6e-a4c0-024bf4260755" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 876.953893] env[62066]: DEBUG oslo_concurrency.lockutils [None req-09aaefa9-33e0-4c34-9b0e-8039240cf379 tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] Acquiring lock "0bc31a8b-a5fc-4f6e-a4c0-024bf4260755-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 876.953893] env[62066]: DEBUG oslo_concurrency.lockutils [None req-09aaefa9-33e0-4c34-9b0e-8039240cf379 tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] Lock "0bc31a8b-a5fc-4f6e-a4c0-024bf4260755-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 876.953893] env[62066]: DEBUG oslo_concurrency.lockutils [None req-09aaefa9-33e0-4c34-9b0e-8039240cf379 tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] Lock "0bc31a8b-a5fc-4f6e-a4c0-024bf4260755-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 876.954247] env[62066]: INFO nova.compute.manager [None req-09aaefa9-33e0-4c34-9b0e-8039240cf379 tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] [instance: 0bc31a8b-a5fc-4f6e-a4c0-024bf4260755] Terminating instance [ 876.955862] env[62066]: DEBUG nova.compute.manager [None req-09aaefa9-33e0-4c34-9b0e-8039240cf379 tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] [instance: 0bc31a8b-a5fc-4f6e-a4c0-024bf4260755] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 876.956196] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-09aaefa9-33e0-4c34-9b0e-8039240cf379 tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] [instance: 0bc31a8b-a5fc-4f6e-a4c0-024bf4260755] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 876.956378] env[62066]: DEBUG oslo_vmware.api [None req-4ccd1f86-3a5c-4e35-bdaa-b8e507c9718e tempest-ServersTestFqdnHostnames-1367628746 tempest-ServersTestFqdnHostnames-1367628746-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52cfa93e-6430-b919-a596-f8ee9b137b37, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.957227] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b1cf6fc-81fb-41e5-8ccc-dd4dc37ab917 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.965562] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-09aaefa9-33e0-4c34-9b0e-8039240cf379 tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] [instance: 0bc31a8b-a5fc-4f6e-a4c0-024bf4260755] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 876.965845] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3ac9eae5-b1af-43eb-8ee5-475427970160 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.972781] env[62066]: DEBUG oslo_vmware.api [None req-09aaefa9-33e0-4c34-9b0e-8039240cf379 tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] Waiting for the task: (returnval){ [ 876.972781] env[62066]: value = "task-1341018" [ 876.972781] env[62066]: _type = "Task" [ 876.972781] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.982387] env[62066]: DEBUG oslo_vmware.api [None req-09aaefa9-33e0-4c34-9b0e-8039240cf379 tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] Task: {'id': task-1341018, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.014833] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f0861cc1-6af6-4429-bc89-cb4b7f228490 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.330s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 877.015409] env[62066]: DEBUG nova.compute.manager [None req-f0861cc1-6af6-4429-bc89-cb4b7f228490 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 8807ce4d-532f-469c-a302-464c61c7efeb] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 877.018073] env[62066]: DEBUG oslo_concurrency.lockutils [None req-6f12e7db-6fdf-4bbe-9b1e-ff4283071f33 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 7.306s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 877.018305] env[62066]: DEBUG nova.objects.instance [None req-6f12e7db-6fdf-4bbe-9b1e-ff4283071f33 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] [instance: 3120c9a5-0558-4e83-b0fb-994a989398fb] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62066) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 877.149501] env[62066]: DEBUG nova.network.neutron [req-d66c920c-b1a6-44b0-b07a-49d72edaedab req-77a82869-327c-4d6f-a1f2-04981e33dd63 service nova] [instance: afaca059-fc4e-4de2-8fa0-d2f226ea7051] Updated VIF entry in instance network info cache for port 725b6138-b39f-41c3-bf84-1a13a3e30fa9. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 877.149501] env[62066]: DEBUG nova.network.neutron [req-d66c920c-b1a6-44b0-b07a-49d72edaedab req-77a82869-327c-4d6f-a1f2-04981e33dd63 service nova] [instance: afaca059-fc4e-4de2-8fa0-d2f226ea7051] Updating instance_info_cache with network_info: [{"id": "725b6138-b39f-41c3-bf84-1a13a3e30fa9", "address": "fa:16:3e:a3:87:a3", "network": {"id": "2d8ba809-853f-4d1b-b4b5-6d255c8a53f3", "bridge": "br-int", "label": "tempest-ServersTestFqdnHostnames-762924253-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "511b5cd53c7745eb8c3e9589b1de01c6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5fdd0624-2edb-4733-8284-225815c07f73", "external-id": "nsx-vlan-transportzone-330", "segmentation_id": 330, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap725b6138-b3", "ovs_interfaceid": "725b6138-b39f-41c3-bf84-1a13a3e30fa9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 877.219212] env[62066]: DEBUG oslo_vmware.api [None req-4c61eb51-497c-4460-951e-d696b73eff28 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': task-1341015, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.289470] env[62066]: DEBUG oslo_vmware.api [None req-59107a59-1a77-47be-8dae-1ec8f16f1fda tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Task: {'id': task-1341017, 'name': Rename_Task, 'duration_secs': 0.180575} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.289470] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-59107a59-1a77-47be-8dae-1ec8f16f1fda tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] [instance: 2634fcc0-96bd-4513-8f92-515f83fe23d7] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 877.289470] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e18a1904-c4af-4d1e-9bec-a1bf06624acc {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.301690] env[62066]: DEBUG oslo_vmware.api [None req-59107a59-1a77-47be-8dae-1ec8f16f1fda tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Waiting for the task: (returnval){ [ 877.301690] env[62066]: value = "task-1341019" [ 877.301690] env[62066]: _type = "Task" [ 877.301690] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.313479] env[62066]: DEBUG oslo_vmware.api [None req-59107a59-1a77-47be-8dae-1ec8f16f1fda tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Task: {'id': task-1341019, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.334038] env[62066]: DEBUG oslo_concurrency.lockutils [req-fe7cbcbc-8b8d-46e0-8190-3a7fdc4aebcc req-cb7213dc-0c3e-4465-876d-6a6b44359bf9 service nova] Releasing lock "refresh_cache-0bc31a8b-a5fc-4f6e-a4c0-024bf4260755" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 877.334429] env[62066]: DEBUG nova.compute.manager [req-fe7cbcbc-8b8d-46e0-8190-3a7fdc4aebcc req-cb7213dc-0c3e-4465-876d-6a6b44359bf9 service nova] [instance: 0bc31a8b-a5fc-4f6e-a4c0-024bf4260755] Received event network-changed-8f34a054-f99b-447b-89d6-ab1ef9bc3c76 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 877.334605] env[62066]: DEBUG nova.compute.manager [req-fe7cbcbc-8b8d-46e0-8190-3a7fdc4aebcc req-cb7213dc-0c3e-4465-876d-6a6b44359bf9 service nova] [instance: 0bc31a8b-a5fc-4f6e-a4c0-024bf4260755] Refreshing instance network info cache due to event network-changed-8f34a054-f99b-447b-89d6-ab1ef9bc3c76. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 877.334836] env[62066]: DEBUG oslo_concurrency.lockutils [req-fe7cbcbc-8b8d-46e0-8190-3a7fdc4aebcc req-cb7213dc-0c3e-4465-876d-6a6b44359bf9 service nova] Acquiring lock "refresh_cache-0bc31a8b-a5fc-4f6e-a4c0-024bf4260755" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 877.334987] env[62066]: DEBUG oslo_concurrency.lockutils [req-fe7cbcbc-8b8d-46e0-8190-3a7fdc4aebcc req-cb7213dc-0c3e-4465-876d-6a6b44359bf9 service nova] Acquired lock "refresh_cache-0bc31a8b-a5fc-4f6e-a4c0-024bf4260755" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 877.335176] env[62066]: DEBUG nova.network.neutron [req-fe7cbcbc-8b8d-46e0-8190-3a7fdc4aebcc req-cb7213dc-0c3e-4465-876d-6a6b44359bf9 service nova] [instance: 0bc31a8b-a5fc-4f6e-a4c0-024bf4260755] Refreshing network info cache for port 8f34a054-f99b-447b-89d6-ab1ef9bc3c76 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 877.456039] env[62066]: DEBUG oslo_vmware.api [None req-4ccd1f86-3a5c-4e35-bdaa-b8e507c9718e tempest-ServersTestFqdnHostnames-1367628746 tempest-ServersTestFqdnHostnames-1367628746-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52cfa93e-6430-b919-a596-f8ee9b137b37, 'name': SearchDatastore_Task, 'duration_secs': 0.011978} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.456039] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4ccd1f86-3a5c-4e35-bdaa-b8e507c9718e tempest-ServersTestFqdnHostnames-1367628746 tempest-ServersTestFqdnHostnames-1367628746-project-member] Releasing lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 877.456284] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-4ccd1f86-3a5c-4e35-bdaa-b8e507c9718e tempest-ServersTestFqdnHostnames-1367628746 tempest-ServersTestFqdnHostnames-1367628746-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk to [datastore2] afaca059-fc4e-4de2-8fa0-d2f226ea7051/afaca059-fc4e-4de2-8fa0-d2f226ea7051.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 877.456413] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1dc5d31e-1ccf-4559-b370-76e0d5f04709 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Acquired lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 877.456879] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-1dc5d31e-1ccf-4559-b370-76e0d5f04709 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 877.456879] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-db274561-0a0b-43b0-91a0-58b2932abc66 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.460551] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ae36f269-d617-4ba6-930b-e629db2b3806 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.469478] env[62066]: DEBUG oslo_vmware.api [None req-4ccd1f86-3a5c-4e35-bdaa-b8e507c9718e tempest-ServersTestFqdnHostnames-1367628746 tempest-ServersTestFqdnHostnames-1367628746-project-member] Waiting for the task: (returnval){ [ 877.469478] env[62066]: value = "task-1341020" [ 877.469478] env[62066]: _type = "Task" [ 877.469478] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.478104] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-1dc5d31e-1ccf-4559-b370-76e0d5f04709 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 877.478358] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-1dc5d31e-1ccf-4559-b370-76e0d5f04709 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 877.482723] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7614ebe6-2e25-4d80-b7bf-5660c143ff53 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.485635] env[62066]: DEBUG oslo_vmware.api [None req-4ccd1f86-3a5c-4e35-bdaa-b8e507c9718e tempest-ServersTestFqdnHostnames-1367628746 tempest-ServersTestFqdnHostnames-1367628746-project-member] Task: {'id': task-1341020, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.493471] env[62066]: DEBUG oslo_vmware.api [None req-09aaefa9-33e0-4c34-9b0e-8039240cf379 tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] Task: {'id': task-1341018, 'name': PowerOffVM_Task, 'duration_secs': 0.347928} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.494388] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-09aaefa9-33e0-4c34-9b0e-8039240cf379 tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] [instance: 0bc31a8b-a5fc-4f6e-a4c0-024bf4260755] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 877.494585] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-09aaefa9-33e0-4c34-9b0e-8039240cf379 tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] [instance: 0bc31a8b-a5fc-4f6e-a4c0-024bf4260755] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 877.494928] env[62066]: DEBUG oslo_vmware.api [None req-1dc5d31e-1ccf-4559-b370-76e0d5f04709 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Waiting for the task: (returnval){ [ 877.494928] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]524b81f0-b86d-c08c-9085-debce198d185" [ 877.494928] env[62066]: _type = "Task" [ 877.494928] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.495310] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-51729b4e-d257-42d8-b9e3-221419caa686 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.506412] env[62066]: DEBUG oslo_vmware.api [None req-1dc5d31e-1ccf-4559-b370-76e0d5f04709 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]524b81f0-b86d-c08c-9085-debce198d185, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.527043] env[62066]: DEBUG nova.compute.utils [None req-f0861cc1-6af6-4429-bc89-cb4b7f228490 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 877.528316] env[62066]: DEBUG nova.compute.manager [None req-f0861cc1-6af6-4429-bc89-cb4b7f228490 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 8807ce4d-532f-469c-a302-464c61c7efeb] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 877.528569] env[62066]: DEBUG nova.network.neutron [None req-f0861cc1-6af6-4429-bc89-cb4b7f228490 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 8807ce4d-532f-469c-a302-464c61c7efeb] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 877.576090] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-09aaefa9-33e0-4c34-9b0e-8039240cf379 tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] [instance: 0bc31a8b-a5fc-4f6e-a4c0-024bf4260755] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 877.576400] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-09aaefa9-33e0-4c34-9b0e-8039240cf379 tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] [instance: 0bc31a8b-a5fc-4f6e-a4c0-024bf4260755] Deleting contents of the VM from datastore datastore1 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 877.576640] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-09aaefa9-33e0-4c34-9b0e-8039240cf379 tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] Deleting the datastore file [datastore1] 0bc31a8b-a5fc-4f6e-a4c0-024bf4260755 {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 877.577292] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d3421c1c-26f8-4227-8b0b-84635fa0b278 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.585260] env[62066]: DEBUG oslo_vmware.api [None req-09aaefa9-33e0-4c34-9b0e-8039240cf379 tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] Waiting for the task: (returnval){ [ 877.585260] env[62066]: value = "task-1341022" [ 877.585260] env[62066]: _type = "Task" [ 877.585260] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.587431] env[62066]: DEBUG nova.policy [None req-f0861cc1-6af6-4429-bc89-cb4b7f228490 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '60e2bc61839b4299912c53137bc23fbc', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '75f0352c852947369474c9e05766c584', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 877.599165] env[62066]: DEBUG oslo_vmware.api [None req-09aaefa9-33e0-4c34-9b0e-8039240cf379 tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] Task: {'id': task-1341022, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.652888] env[62066]: DEBUG oslo_concurrency.lockutils [req-d66c920c-b1a6-44b0-b07a-49d72edaedab req-77a82869-327c-4d6f-a1f2-04981e33dd63 service nova] Releasing lock "refresh_cache-afaca059-fc4e-4de2-8fa0-d2f226ea7051" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 877.653460] env[62066]: DEBUG nova.compute.manager [req-d66c920c-b1a6-44b0-b07a-49d72edaedab req-77a82869-327c-4d6f-a1f2-04981e33dd63 service nova] [instance: 1a4b0637-1a56-41ef-b89b-6b56d24ed206] Received event network-vif-deleted-84936b80-0721-4235-a92a-7bc4adf450d8 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 877.720057] env[62066]: DEBUG oslo_vmware.api [None req-4c61eb51-497c-4460-951e-d696b73eff28 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': task-1341015, 'name': ReconfigVM_Task, 'duration_secs': 1.139816} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.720413] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-4c61eb51-497c-4460-951e-d696b73eff28 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 170e184a-f009-42cf-82dc-eb9696a00a24] Reconfigured VM instance instance-00000052 to attach disk [datastore2] 170e184a-f009-42cf-82dc-eb9696a00a24/170e184a-f009-42cf-82dc-eb9696a00a24.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 877.721146] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-020560d5-38e3-4739-a369-d8275af0ed8e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.725023] env[62066]: DEBUG oslo_concurrency.lockutils [None req-98bb85f9-df22-41ab-a6e4-7c41686713f5 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Acquiring lock "de2e2be7-efdb-45a8-842a-640ab9deb1d9" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 877.725568] env[62066]: DEBUG oslo_concurrency.lockutils [None req-98bb85f9-df22-41ab-a6e4-7c41686713f5 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Lock "de2e2be7-efdb-45a8-842a-640ab9deb1d9" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 877.725568] env[62066]: INFO nova.compute.manager [None req-98bb85f9-df22-41ab-a6e4-7c41686713f5 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] [instance: de2e2be7-efdb-45a8-842a-640ab9deb1d9] Attaching volume 36cdc9f5-eb4f-44b9-9aac-a7cd3fc5d570 to /dev/sdb [ 877.730713] env[62066]: DEBUG oslo_vmware.api [None req-4c61eb51-497c-4460-951e-d696b73eff28 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Waiting for the task: (returnval){ [ 877.730713] env[62066]: value = "task-1341023" [ 877.730713] env[62066]: _type = "Task" [ 877.730713] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.746516] env[62066]: DEBUG oslo_vmware.api [None req-4c61eb51-497c-4460-951e-d696b73eff28 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': task-1341023, 'name': Rename_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.766289] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a0b318f-6b5e-4fb1-8564-f1c2edf27534 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.775643] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12114d0d-2590-4ef4-9d30-55bfc26ef3b3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.797192] env[62066]: DEBUG nova.virt.block_device [None req-98bb85f9-df22-41ab-a6e4-7c41686713f5 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] [instance: de2e2be7-efdb-45a8-842a-640ab9deb1d9] Updating existing volume attachment record: 3be95a5c-9d96-402c-a175-a64775edc683 {{(pid=62066) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 877.817601] env[62066]: DEBUG oslo_vmware.api [None req-59107a59-1a77-47be-8dae-1ec8f16f1fda tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Task: {'id': task-1341019, 'name': PowerOnVM_Task} progress is 87%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.990191] env[62066]: DEBUG oslo_vmware.api [None req-4ccd1f86-3a5c-4e35-bdaa-b8e507c9718e tempest-ServersTestFqdnHostnames-1367628746 tempest-ServersTestFqdnHostnames-1367628746-project-member] Task: {'id': task-1341020, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.015388] env[62066]: DEBUG oslo_vmware.api [None req-1dc5d31e-1ccf-4559-b370-76e0d5f04709 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]524b81f0-b86d-c08c-9085-debce198d185, 'name': SearchDatastore_Task, 'duration_secs': 0.015379} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.015388] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5d1d64c3-5a4e-488f-97e9-e11a11bbc562 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.023900] env[62066]: DEBUG oslo_vmware.api [None req-1dc5d31e-1ccf-4559-b370-76e0d5f04709 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Waiting for the task: (returnval){ [ 878.023900] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]52d0d9d6-46f4-b06a-97fe-e455e258fd44" [ 878.023900] env[62066]: _type = "Task" [ 878.023900] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.031883] env[62066]: DEBUG oslo_concurrency.lockutils [None req-6f12e7db-6fdf-4bbe-9b1e-ff4283071f33 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.011s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 878.040963] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d7bf9b49-5ae5-4613-a969-417764bdb185 tempest-ServersV294TestFqdnHostnames-843300434 tempest-ServersV294TestFqdnHostnames-843300434-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.282s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 878.040963] env[62066]: DEBUG nova.objects.instance [None req-d7bf9b49-5ae5-4613-a969-417764bdb185 tempest-ServersV294TestFqdnHostnames-843300434 tempest-ServersV294TestFqdnHostnames-843300434-project-member] Lazy-loading 'resources' on Instance uuid 2709d0ad-5dd8-4e3c-b1e6-3d1e2e242380 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 878.040963] env[62066]: DEBUG nova.compute.manager [None req-f0861cc1-6af6-4429-bc89-cb4b7f228490 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 8807ce4d-532f-469c-a302-464c61c7efeb] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 878.044506] env[62066]: DEBUG oslo_vmware.api [None req-1dc5d31e-1ccf-4559-b370-76e0d5f04709 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52d0d9d6-46f4-b06a-97fe-e455e258fd44, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.064231] env[62066]: DEBUG nova.network.neutron [None req-f0861cc1-6af6-4429-bc89-cb4b7f228490 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 8807ce4d-532f-469c-a302-464c61c7efeb] Successfully created port: dba295b3-a285-4890-b73b-ed681d77192a {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 878.102018] env[62066]: DEBUG oslo_vmware.api [None req-09aaefa9-33e0-4c34-9b0e-8039240cf379 tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] Task: {'id': task-1341022, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.232133] env[62066]: DEBUG nova.network.neutron [req-fe7cbcbc-8b8d-46e0-8190-3a7fdc4aebcc req-cb7213dc-0c3e-4465-876d-6a6b44359bf9 service nova] [instance: 0bc31a8b-a5fc-4f6e-a4c0-024bf4260755] Updated VIF entry in instance network info cache for port 8f34a054-f99b-447b-89d6-ab1ef9bc3c76. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 878.232685] env[62066]: DEBUG nova.network.neutron [req-fe7cbcbc-8b8d-46e0-8190-3a7fdc4aebcc req-cb7213dc-0c3e-4465-876d-6a6b44359bf9 service nova] [instance: 0bc31a8b-a5fc-4f6e-a4c0-024bf4260755] Updating instance_info_cache with network_info: [{"id": "8f34a054-f99b-447b-89d6-ab1ef9bc3c76", "address": "fa:16:3e:06:43:d8", "network": {"id": "b92d2e33-767e-4958-b805-12e11b75c266", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-583710540-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "5281a37e471347b79e5dd6e70e03aa0f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "13e71dbb-4279-427c-b39d-ba5df9895e58", "external-id": "nsx-vlan-transportzone-417", "segmentation_id": 417, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8f34a054-f9", "ovs_interfaceid": "8f34a054-f99b-447b-89d6-ab1ef9bc3c76", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 878.257252] env[62066]: DEBUG oslo_vmware.api [None req-4c61eb51-497c-4460-951e-d696b73eff28 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': task-1341023, 'name': Rename_Task, 'duration_secs': 0.287488} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.259965] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-4c61eb51-497c-4460-951e-d696b73eff28 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 170e184a-f009-42cf-82dc-eb9696a00a24] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 878.261942] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8bdb73d4-5658-4b97-b4ef-ded441614d33 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.273676] env[62066]: DEBUG oslo_vmware.api [None req-4c61eb51-497c-4460-951e-d696b73eff28 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Waiting for the task: (returnval){ [ 878.273676] env[62066]: value = "task-1341026" [ 878.273676] env[62066]: _type = "Task" [ 878.273676] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.285525] env[62066]: DEBUG oslo_vmware.api [None req-4c61eb51-497c-4460-951e-d696b73eff28 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': task-1341026, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.313216] env[62066]: DEBUG oslo_vmware.api [None req-59107a59-1a77-47be-8dae-1ec8f16f1fda tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Task: {'id': task-1341019, 'name': PowerOnVM_Task, 'duration_secs': 0.957717} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.313664] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-59107a59-1a77-47be-8dae-1ec8f16f1fda tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] [instance: 2634fcc0-96bd-4513-8f92-515f83fe23d7] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 878.313814] env[62066]: INFO nova.compute.manager [None req-59107a59-1a77-47be-8dae-1ec8f16f1fda tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] [instance: 2634fcc0-96bd-4513-8f92-515f83fe23d7] Took 10.37 seconds to spawn the instance on the hypervisor. [ 878.314060] env[62066]: DEBUG nova.compute.manager [None req-59107a59-1a77-47be-8dae-1ec8f16f1fda tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] [instance: 2634fcc0-96bd-4513-8f92-515f83fe23d7] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 878.315015] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-908ac64f-984c-4885-863c-d8c9b3f16977 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.485516] env[62066]: DEBUG oslo_vmware.api [None req-4ccd1f86-3a5c-4e35-bdaa-b8e507c9718e tempest-ServersTestFqdnHostnames-1367628746 tempest-ServersTestFqdnHostnames-1367628746-project-member] Task: {'id': task-1341020, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.800202} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.485909] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-4ccd1f86-3a5c-4e35-bdaa-b8e507c9718e tempest-ServersTestFqdnHostnames-1367628746 tempest-ServersTestFqdnHostnames-1367628746-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk to [datastore2] afaca059-fc4e-4de2-8fa0-d2f226ea7051/afaca059-fc4e-4de2-8fa0-d2f226ea7051.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 878.486204] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-4ccd1f86-3a5c-4e35-bdaa-b8e507c9718e tempest-ServersTestFqdnHostnames-1367628746 tempest-ServersTestFqdnHostnames-1367628746-project-member] [instance: afaca059-fc4e-4de2-8fa0-d2f226ea7051] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 878.486541] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-98757dda-83f5-413d-8de0-49ba768f7b67 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.503399] env[62066]: DEBUG oslo_vmware.api [None req-4ccd1f86-3a5c-4e35-bdaa-b8e507c9718e tempest-ServersTestFqdnHostnames-1367628746 tempest-ServersTestFqdnHostnames-1367628746-project-member] Waiting for the task: (returnval){ [ 878.503399] env[62066]: value = "task-1341028" [ 878.503399] env[62066]: _type = "Task" [ 878.503399] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.521653] env[62066]: DEBUG oslo_vmware.api [None req-4ccd1f86-3a5c-4e35-bdaa-b8e507c9718e tempest-ServersTestFqdnHostnames-1367628746 tempest-ServersTestFqdnHostnames-1367628746-project-member] Task: {'id': task-1341028, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.537612] env[62066]: DEBUG oslo_vmware.api [None req-1dc5d31e-1ccf-4559-b370-76e0d5f04709 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52d0d9d6-46f4-b06a-97fe-e455e258fd44, 'name': SearchDatastore_Task, 'duration_secs': 0.066521} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.537967] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1dc5d31e-1ccf-4559-b370-76e0d5f04709 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Releasing lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 878.538469] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-1dc5d31e-1ccf-4559-b370-76e0d5f04709 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk to [datastore2] 76306c3f-0674-40fe-9864-d82d11e4bb5e/76306c3f-0674-40fe-9864-d82d11e4bb5e.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 878.539197] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0aeaf1ca-cb8e-4631-99b6-069cc79e4b23 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.557823] env[62066]: DEBUG oslo_vmware.api [None req-1dc5d31e-1ccf-4559-b370-76e0d5f04709 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Waiting for the task: (returnval){ [ 878.557823] env[62066]: value = "task-1341029" [ 878.557823] env[62066]: _type = "Task" [ 878.557823] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.572486] env[62066]: DEBUG oslo_vmware.api [None req-1dc5d31e-1ccf-4559-b370-76e0d5f04709 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1341029, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.605287] env[62066]: DEBUG oslo_vmware.api [None req-09aaefa9-33e0-4c34-9b0e-8039240cf379 tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] Task: {'id': task-1341022, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.685264} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.605609] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-09aaefa9-33e0-4c34-9b0e-8039240cf379 tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 878.605814] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-09aaefa9-33e0-4c34-9b0e-8039240cf379 tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] [instance: 0bc31a8b-a5fc-4f6e-a4c0-024bf4260755] Deleted contents of the VM from datastore datastore1 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 878.606048] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-09aaefa9-33e0-4c34-9b0e-8039240cf379 tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] [instance: 0bc31a8b-a5fc-4f6e-a4c0-024bf4260755] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 878.606226] env[62066]: INFO nova.compute.manager [None req-09aaefa9-33e0-4c34-9b0e-8039240cf379 tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] [instance: 0bc31a8b-a5fc-4f6e-a4c0-024bf4260755] Took 1.65 seconds to destroy the instance on the hypervisor. [ 878.606468] env[62066]: DEBUG oslo.service.loopingcall [None req-09aaefa9-33e0-4c34-9b0e-8039240cf379 tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 878.607040] env[62066]: DEBUG nova.compute.manager [-] [instance: 0bc31a8b-a5fc-4f6e-a4c0-024bf4260755] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 878.607143] env[62066]: DEBUG nova.network.neutron [-] [instance: 0bc31a8b-a5fc-4f6e-a4c0-024bf4260755] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 878.736468] env[62066]: DEBUG oslo_concurrency.lockutils [req-fe7cbcbc-8b8d-46e0-8190-3a7fdc4aebcc req-cb7213dc-0c3e-4465-876d-6a6b44359bf9 service nova] Releasing lock "refresh_cache-0bc31a8b-a5fc-4f6e-a4c0-024bf4260755" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 878.791513] env[62066]: DEBUG oslo_vmware.api [None req-4c61eb51-497c-4460-951e-d696b73eff28 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': task-1341026, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.842980] env[62066]: INFO nova.compute.manager [None req-59107a59-1a77-47be-8dae-1ec8f16f1fda tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] [instance: 2634fcc0-96bd-4513-8f92-515f83fe23d7] Took 29.82 seconds to build instance. [ 878.878559] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86c1e7a8-feb0-4371-9699-41024b8d1bab {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.892153] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a60dda9a-4092-4a58-89ac-c020ecc26c6f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.944393] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bc78436-75d7-4a74-acff-079e0834c717 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.958439] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-026280f0-dd27-4584-82f3-f7ed7f9aac22 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.983618] env[62066]: DEBUG nova.compute.provider_tree [None req-d7bf9b49-5ae5-4613-a969-417764bdb185 tempest-ServersV294TestFqdnHostnames-843300434 tempest-ServersV294TestFqdnHostnames-843300434-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 879.021920] env[62066]: DEBUG oslo_vmware.api [None req-4ccd1f86-3a5c-4e35-bdaa-b8e507c9718e tempest-ServersTestFqdnHostnames-1367628746 tempest-ServersTestFqdnHostnames-1367628746-project-member] Task: {'id': task-1341028, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.113484} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.023148] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-4ccd1f86-3a5c-4e35-bdaa-b8e507c9718e tempest-ServersTestFqdnHostnames-1367628746 tempest-ServersTestFqdnHostnames-1367628746-project-member] [instance: afaca059-fc4e-4de2-8fa0-d2f226ea7051] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 879.023190] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f7aa1fc-2b9d-42f0-b46e-33ba7f93e6fb {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.057104] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-4ccd1f86-3a5c-4e35-bdaa-b8e507c9718e tempest-ServersTestFqdnHostnames-1367628746 tempest-ServersTestFqdnHostnames-1367628746-project-member] [instance: afaca059-fc4e-4de2-8fa0-d2f226ea7051] Reconfiguring VM instance instance-00000053 to attach disk [datastore2] afaca059-fc4e-4de2-8fa0-d2f226ea7051/afaca059-fc4e-4de2-8fa0-d2f226ea7051.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 879.057543] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6c8e74d7-1315-41e8-95fd-d6d904f99b3f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.074725] env[62066]: DEBUG nova.compute.manager [None req-f0861cc1-6af6-4429-bc89-cb4b7f228490 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 8807ce4d-532f-469c-a302-464c61c7efeb] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 879.089386] env[62066]: DEBUG oslo_vmware.api [None req-1dc5d31e-1ccf-4559-b370-76e0d5f04709 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1341029, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.091476] env[62066]: DEBUG oslo_vmware.api [None req-4ccd1f86-3a5c-4e35-bdaa-b8e507c9718e tempest-ServersTestFqdnHostnames-1367628746 tempest-ServersTestFqdnHostnames-1367628746-project-member] Waiting for the task: (returnval){ [ 879.091476] env[62066]: value = "task-1341030" [ 879.091476] env[62066]: _type = "Task" [ 879.091476] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.102469] env[62066]: DEBUG oslo_vmware.api [None req-4ccd1f86-3a5c-4e35-bdaa-b8e507c9718e tempest-ServersTestFqdnHostnames-1367628746 tempest-ServersTestFqdnHostnames-1367628746-project-member] Task: {'id': task-1341030, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.116324] env[62066]: DEBUG nova.virt.hardware [None req-f0861cc1-6af6-4429-bc89-cb4b7f228490 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-23T13:40:41Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='167d2318fe68a6d1c5c831c00bdd672a',container_format='bare',created_at=2024-10-23T13:48:18Z,direct_url=,disk_format='vmdk',id=1147d77f-02f3-4d4b-a853-4a0e7d8b0a96,min_disk=1,min_ram=0,name='tempest-test-snap-1360101594',owner='75f0352c852947369474c9e05766c584',properties=ImageMetaProps,protected=,size=21334016,status='active',tags=,updated_at=2024-10-23T13:48:31Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 879.116668] env[62066]: DEBUG nova.virt.hardware [None req-f0861cc1-6af6-4429-bc89-cb4b7f228490 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 879.117038] env[62066]: DEBUG nova.virt.hardware [None req-f0861cc1-6af6-4429-bc89-cb4b7f228490 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 879.117267] env[62066]: DEBUG nova.virt.hardware [None req-f0861cc1-6af6-4429-bc89-cb4b7f228490 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 879.117430] env[62066]: DEBUG nova.virt.hardware [None req-f0861cc1-6af6-4429-bc89-cb4b7f228490 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 879.117588] env[62066]: DEBUG nova.virt.hardware [None req-f0861cc1-6af6-4429-bc89-cb4b7f228490 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 879.117823] env[62066]: DEBUG nova.virt.hardware [None req-f0861cc1-6af6-4429-bc89-cb4b7f228490 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 879.117993] env[62066]: DEBUG nova.virt.hardware [None req-f0861cc1-6af6-4429-bc89-cb4b7f228490 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 879.118200] env[62066]: DEBUG nova.virt.hardware [None req-f0861cc1-6af6-4429-bc89-cb4b7f228490 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 879.118370] env[62066]: DEBUG nova.virt.hardware [None req-f0861cc1-6af6-4429-bc89-cb4b7f228490 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 879.118544] env[62066]: DEBUG nova.virt.hardware [None req-f0861cc1-6af6-4429-bc89-cb4b7f228490 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 879.119509] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5db1fdd4-1c67-4573-9c56-e13b82a31364 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.126697] env[62066]: DEBUG nova.compute.manager [req-5f260585-d4e8-44ec-808e-e246d6c2f6b5 req-cccbc8b2-7597-4644-b6c9-3c2c66c22ad8 service nova] [instance: 0bc31a8b-a5fc-4f6e-a4c0-024bf4260755] Received event network-vif-deleted-8f34a054-f99b-447b-89d6-ab1ef9bc3c76 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 879.126865] env[62066]: INFO nova.compute.manager [req-5f260585-d4e8-44ec-808e-e246d6c2f6b5 req-cccbc8b2-7597-4644-b6c9-3c2c66c22ad8 service nova] [instance: 0bc31a8b-a5fc-4f6e-a4c0-024bf4260755] Neutron deleted interface 8f34a054-f99b-447b-89d6-ab1ef9bc3c76; detaching it from the instance and deleting it from the info cache [ 879.127009] env[62066]: DEBUG nova.network.neutron [req-5f260585-d4e8-44ec-808e-e246d6c2f6b5 req-cccbc8b2-7597-4644-b6c9-3c2c66c22ad8 service nova] [instance: 0bc31a8b-a5fc-4f6e-a4c0-024bf4260755] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 879.132556] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a81c7943-6c5b-4979-9650-f6bcbb55fecb {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.287421] env[62066]: DEBUG oslo_vmware.api [None req-4c61eb51-497c-4460-951e-d696b73eff28 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': task-1341026, 'name': PowerOnVM_Task, 'duration_secs': 0.824801} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.287707] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-4c61eb51-497c-4460-951e-d696b73eff28 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 170e184a-f009-42cf-82dc-eb9696a00a24] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 879.287941] env[62066]: INFO nova.compute.manager [None req-4c61eb51-497c-4460-951e-d696b73eff28 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 170e184a-f009-42cf-82dc-eb9696a00a24] Took 8.89 seconds to spawn the instance on the hypervisor. [ 879.288120] env[62066]: DEBUG nova.compute.manager [None req-4c61eb51-497c-4460-951e-d696b73eff28 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 170e184a-f009-42cf-82dc-eb9696a00a24] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 879.289034] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-392a84d2-6627-4d8b-aca8-a792fe0e9de1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.347055] env[62066]: DEBUG oslo_concurrency.lockutils [None req-59107a59-1a77-47be-8dae-1ec8f16f1fda tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Lock "2634fcc0-96bd-4513-8f92-515f83fe23d7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 31.336s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 879.454184] env[62066]: DEBUG nova.network.neutron [-] [instance: 0bc31a8b-a5fc-4f6e-a4c0-024bf4260755] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 879.487284] env[62066]: DEBUG nova.scheduler.client.report [None req-d7bf9b49-5ae5-4613-a969-417764bdb185 tempest-ServersV294TestFqdnHostnames-843300434 tempest-ServersV294TestFqdnHostnames-843300434-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 879.587201] env[62066]: DEBUG oslo_vmware.api [None req-1dc5d31e-1ccf-4559-b370-76e0d5f04709 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1341029, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.785019} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.587504] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-1dc5d31e-1ccf-4559-b370-76e0d5f04709 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk to [datastore2] 76306c3f-0674-40fe-9864-d82d11e4bb5e/76306c3f-0674-40fe-9864-d82d11e4bb5e.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 879.587729] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-1dc5d31e-1ccf-4559-b370-76e0d5f04709 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 76306c3f-0674-40fe-9864-d82d11e4bb5e] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 879.587998] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-116b62a6-c8c7-44e5-b8a7-41c332c58acb {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.600739] env[62066]: DEBUG oslo_vmware.api [None req-4ccd1f86-3a5c-4e35-bdaa-b8e507c9718e tempest-ServersTestFqdnHostnames-1367628746 tempest-ServersTestFqdnHostnames-1367628746-project-member] Task: {'id': task-1341030, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.602278] env[62066]: DEBUG oslo_vmware.api [None req-1dc5d31e-1ccf-4559-b370-76e0d5f04709 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Waiting for the task: (returnval){ [ 879.602278] env[62066]: value = "task-1341031" [ 879.602278] env[62066]: _type = "Task" [ 879.602278] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.612150] env[62066]: DEBUG oslo_vmware.api [None req-1dc5d31e-1ccf-4559-b370-76e0d5f04709 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1341031, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.630232] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6755fa92-20fc-49e4-92dd-2411861cf1cc {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.643048] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-415e729c-11fe-44d9-837f-b9443e7271a0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.676128] env[62066]: DEBUG nova.compute.manager [req-5f260585-d4e8-44ec-808e-e246d6c2f6b5 req-cccbc8b2-7597-4644-b6c9-3c2c66c22ad8 service nova] [instance: 0bc31a8b-a5fc-4f6e-a4c0-024bf4260755] Detach interface failed, port_id=8f34a054-f99b-447b-89d6-ab1ef9bc3c76, reason: Instance 0bc31a8b-a5fc-4f6e-a4c0-024bf4260755 could not be found. {{(pid=62066) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 879.710274] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c06200c8-d207-47fe-8d97-8cb6c611ec30 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Acquiring lock "b718b09b-a7fa-445b-8be3-5b3ebca210a2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 879.710546] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c06200c8-d207-47fe-8d97-8cb6c611ec30 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Lock "b718b09b-a7fa-445b-8be3-5b3ebca210a2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 879.808921] env[62066]: INFO nova.compute.manager [None req-4c61eb51-497c-4460-951e-d696b73eff28 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 170e184a-f009-42cf-82dc-eb9696a00a24] Took 23.48 seconds to build instance. [ 879.936546] env[62066]: DEBUG nova.compute.manager [req-0770bb30-eac4-462e-bead-811845aea42e req-3d27ca0c-a1a7-4e48-a8ec-39af614051de service nova] [instance: 8807ce4d-532f-469c-a302-464c61c7efeb] Received event network-vif-plugged-dba295b3-a285-4890-b73b-ed681d77192a {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 879.936546] env[62066]: DEBUG oslo_concurrency.lockutils [req-0770bb30-eac4-462e-bead-811845aea42e req-3d27ca0c-a1a7-4e48-a8ec-39af614051de service nova] Acquiring lock "8807ce4d-532f-469c-a302-464c61c7efeb-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 879.937828] env[62066]: DEBUG oslo_concurrency.lockutils [req-0770bb30-eac4-462e-bead-811845aea42e req-3d27ca0c-a1a7-4e48-a8ec-39af614051de service nova] Lock "8807ce4d-532f-469c-a302-464c61c7efeb-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.002s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 879.938014] env[62066]: DEBUG oslo_concurrency.lockutils [req-0770bb30-eac4-462e-bead-811845aea42e req-3d27ca0c-a1a7-4e48-a8ec-39af614051de service nova] Lock "8807ce4d-532f-469c-a302-464c61c7efeb-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 879.938211] env[62066]: DEBUG nova.compute.manager [req-0770bb30-eac4-462e-bead-811845aea42e req-3d27ca0c-a1a7-4e48-a8ec-39af614051de service nova] [instance: 8807ce4d-532f-469c-a302-464c61c7efeb] No waiting events found dispatching network-vif-plugged-dba295b3-a285-4890-b73b-ed681d77192a {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 879.938756] env[62066]: WARNING nova.compute.manager [req-0770bb30-eac4-462e-bead-811845aea42e req-3d27ca0c-a1a7-4e48-a8ec-39af614051de service nova] [instance: 8807ce4d-532f-469c-a302-464c61c7efeb] Received unexpected event network-vif-plugged-dba295b3-a285-4890-b73b-ed681d77192a for instance with vm_state building and task_state spawning. [ 879.941306] env[62066]: INFO nova.compute.manager [None req-e9fb52ef-3f9c-460e-85ba-59eb989b863f tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] [instance: 2634fcc0-96bd-4513-8f92-515f83fe23d7] Rescuing [ 879.941575] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e9fb52ef-3f9c-460e-85ba-59eb989b863f tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Acquiring lock "refresh_cache-2634fcc0-96bd-4513-8f92-515f83fe23d7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 879.941754] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e9fb52ef-3f9c-460e-85ba-59eb989b863f tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Acquired lock "refresh_cache-2634fcc0-96bd-4513-8f92-515f83fe23d7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 879.941927] env[62066]: DEBUG nova.network.neutron [None req-e9fb52ef-3f9c-460e-85ba-59eb989b863f tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] [instance: 2634fcc0-96bd-4513-8f92-515f83fe23d7] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 879.956306] env[62066]: INFO nova.compute.manager [-] [instance: 0bc31a8b-a5fc-4f6e-a4c0-024bf4260755] Took 1.35 seconds to deallocate network for instance. [ 879.991879] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d7bf9b49-5ae5-4613-a969-417764bdb185 tempest-ServersV294TestFqdnHostnames-843300434 tempest-ServersV294TestFqdnHostnames-843300434-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.953s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 879.994734] env[62066]: DEBUG oslo_concurrency.lockutils [None req-535b520b-dc22-40fb-a872-f0522e7e73a7 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.693s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 879.994978] env[62066]: DEBUG nova.objects.instance [None req-535b520b-dc22-40fb-a872-f0522e7e73a7 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Lazy-loading 'resources' on Instance uuid 3120c9a5-0558-4e83-b0fb-994a989398fb {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 880.028076] env[62066]: INFO nova.scheduler.client.report [None req-d7bf9b49-5ae5-4613-a969-417764bdb185 tempest-ServersV294TestFqdnHostnames-843300434 tempest-ServersV294TestFqdnHostnames-843300434-project-member] Deleted allocations for instance 2709d0ad-5dd8-4e3c-b1e6-3d1e2e242380 [ 880.038975] env[62066]: DEBUG nova.network.neutron [None req-f0861cc1-6af6-4429-bc89-cb4b7f228490 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 8807ce4d-532f-469c-a302-464c61c7efeb] Successfully updated port: dba295b3-a285-4890-b73b-ed681d77192a {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 880.102744] env[62066]: DEBUG oslo_vmware.api [None req-4ccd1f86-3a5c-4e35-bdaa-b8e507c9718e tempest-ServersTestFqdnHostnames-1367628746 tempest-ServersTestFqdnHostnames-1367628746-project-member] Task: {'id': task-1341030, 'name': ReconfigVM_Task, 'duration_secs': 0.824632} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.103133] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-4ccd1f86-3a5c-4e35-bdaa-b8e507c9718e tempest-ServersTestFqdnHostnames-1367628746 tempest-ServersTestFqdnHostnames-1367628746-project-member] [instance: afaca059-fc4e-4de2-8fa0-d2f226ea7051] Reconfigured VM instance instance-00000053 to attach disk [datastore2] afaca059-fc4e-4de2-8fa0-d2f226ea7051/afaca059-fc4e-4de2-8fa0-d2f226ea7051.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 880.106562] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a41ba5e4-7e62-4c49-b5f3-82611693c55a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.117238] env[62066]: DEBUG oslo_vmware.api [None req-1dc5d31e-1ccf-4559-b370-76e0d5f04709 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1341031, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.078949} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.118727] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-1dc5d31e-1ccf-4559-b370-76e0d5f04709 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 76306c3f-0674-40fe-9864-d82d11e4bb5e] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 880.119098] env[62066]: DEBUG oslo_vmware.api [None req-4ccd1f86-3a5c-4e35-bdaa-b8e507c9718e tempest-ServersTestFqdnHostnames-1367628746 tempest-ServersTestFqdnHostnames-1367628746-project-member] Waiting for the task: (returnval){ [ 880.119098] env[62066]: value = "task-1341032" [ 880.119098] env[62066]: _type = "Task" [ 880.119098] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.119795] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18f9d264-c8d2-4725-a0ab-7302680a307a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.130840] env[62066]: DEBUG oslo_vmware.api [None req-4ccd1f86-3a5c-4e35-bdaa-b8e507c9718e tempest-ServersTestFqdnHostnames-1367628746 tempest-ServersTestFqdnHostnames-1367628746-project-member] Task: {'id': task-1341032, 'name': Rename_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.148639] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-1dc5d31e-1ccf-4559-b370-76e0d5f04709 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 76306c3f-0674-40fe-9864-d82d11e4bb5e] Reconfiguring VM instance instance-00000050 to attach disk [datastore2] 76306c3f-0674-40fe-9864-d82d11e4bb5e/76306c3f-0674-40fe-9864-d82d11e4bb5e.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 880.149377] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-81e33e9c-9f39-4191-ac68-6b78f68a6331 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.170650] env[62066]: DEBUG oslo_vmware.api [None req-1dc5d31e-1ccf-4559-b370-76e0d5f04709 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Waiting for the task: (returnval){ [ 880.170650] env[62066]: value = "task-1341033" [ 880.170650] env[62066]: _type = "Task" [ 880.170650] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.180552] env[62066]: DEBUG oslo_vmware.api [None req-1dc5d31e-1ccf-4559-b370-76e0d5f04709 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1341033, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.213198] env[62066]: DEBUG nova.compute.manager [None req-c06200c8-d207-47fe-8d97-8cb6c611ec30 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: b718b09b-a7fa-445b-8be3-5b3ebca210a2] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 880.314706] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4c61eb51-497c-4460-951e-d696b73eff28 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Lock "170e184a-f009-42cf-82dc-eb9696a00a24" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 25.016s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 880.463377] env[62066]: DEBUG oslo_concurrency.lockutils [None req-09aaefa9-33e0-4c34-9b0e-8039240cf379 tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 880.546873] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f0861cc1-6af6-4429-bc89-cb4b7f228490 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Acquiring lock "refresh_cache-8807ce4d-532f-469c-a302-464c61c7efeb" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 880.546873] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f0861cc1-6af6-4429-bc89-cb4b7f228490 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Acquired lock "refresh_cache-8807ce4d-532f-469c-a302-464c61c7efeb" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 880.546873] env[62066]: DEBUG nova.network.neutron [None req-f0861cc1-6af6-4429-bc89-cb4b7f228490 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 8807ce4d-532f-469c-a302-464c61c7efeb] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 880.546873] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d7bf9b49-5ae5-4613-a969-417764bdb185 tempest-ServersV294TestFqdnHostnames-843300434 tempest-ServersV294TestFqdnHostnames-843300434-project-member] Lock "2709d0ad-5dd8-4e3c-b1e6-3d1e2e242380" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 12.085s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 880.635462] env[62066]: DEBUG oslo_vmware.api [None req-4ccd1f86-3a5c-4e35-bdaa-b8e507c9718e tempest-ServersTestFqdnHostnames-1367628746 tempest-ServersTestFqdnHostnames-1367628746-project-member] Task: {'id': task-1341032, 'name': Rename_Task, 'duration_secs': 0.329143} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.635764] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-4ccd1f86-3a5c-4e35-bdaa-b8e507c9718e tempest-ServersTestFqdnHostnames-1367628746 tempest-ServersTestFqdnHostnames-1367628746-project-member] [instance: afaca059-fc4e-4de2-8fa0-d2f226ea7051] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 880.636079] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a091486d-995d-4457-b89a-200017fe3d4e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.644983] env[62066]: DEBUG oslo_vmware.api [None req-4ccd1f86-3a5c-4e35-bdaa-b8e507c9718e tempest-ServersTestFqdnHostnames-1367628746 tempest-ServersTestFqdnHostnames-1367628746-project-member] Waiting for the task: (returnval){ [ 880.644983] env[62066]: value = "task-1341035" [ 880.644983] env[62066]: _type = "Task" [ 880.644983] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.654529] env[62066]: DEBUG oslo_vmware.api [None req-4ccd1f86-3a5c-4e35-bdaa-b8e507c9718e tempest-ServersTestFqdnHostnames-1367628746 tempest-ServersTestFqdnHostnames-1367628746-project-member] Task: {'id': task-1341035, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.684287] env[62066]: DEBUG oslo_vmware.api [None req-1dc5d31e-1ccf-4559-b370-76e0d5f04709 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1341033, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.713386] env[62066]: DEBUG oslo_vmware.rw_handles [None req-4a3c2cb7-f1e2-437a-b493-e5de08e55742 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/520ab008-37bf-a52c-cfda-92b0db99cbeb/disk-0.vmdk. {{(pid=62066) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 880.714600] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ada55f9b-38c4-4915-be16-1a1c6f918fc9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.727278] env[62066]: DEBUG oslo_vmware.rw_handles [None req-4a3c2cb7-f1e2-437a-b493-e5de08e55742 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/520ab008-37bf-a52c-cfda-92b0db99cbeb/disk-0.vmdk is in state: ready. {{(pid=62066) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 880.727473] env[62066]: ERROR oslo_vmware.rw_handles [None req-4a3c2cb7-f1e2-437a-b493-e5de08e55742 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/520ab008-37bf-a52c-cfda-92b0db99cbeb/disk-0.vmdk due to incomplete transfer. [ 880.727699] env[62066]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-5ac5bdcf-ab24-4c2f-b1ba-052b96c810ee {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.735215] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c06200c8-d207-47fe-8d97-8cb6c611ec30 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 880.736813] env[62066]: DEBUG oslo_vmware.rw_handles [None req-4a3c2cb7-f1e2-437a-b493-e5de08e55742 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/520ab008-37bf-a52c-cfda-92b0db99cbeb/disk-0.vmdk. {{(pid=62066) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 880.737020] env[62066]: DEBUG nova.virt.vmwareapi.images [None req-4a3c2cb7-f1e2-437a-b493-e5de08e55742 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] [instance: 8ff6fbb9-c90f-498d-9a85-d220a8c2f794] Uploaded image 1ebd2ae1-0ae3-4486-822c-c75cafe27fe3 to the Glance image server {{(pid=62066) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 880.741659] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-4a3c2cb7-f1e2-437a-b493-e5de08e55742 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] [instance: 8ff6fbb9-c90f-498d-9a85-d220a8c2f794] Destroying the VM {{(pid=62066) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 880.741659] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-8fac8aa0-5062-4af9-b306-1a9c925ba81d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.745175] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd4c7e0e-bf53-447c-8c5f-59bfebb98bc1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.755020] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccdfeeb5-07b6-4329-b86e-b688aa62f8a9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.756764] env[62066]: DEBUG oslo_vmware.api [None req-4a3c2cb7-f1e2-437a-b493-e5de08e55742 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Waiting for the task: (returnval){ [ 880.756764] env[62066]: value = "task-1341036" [ 880.756764] env[62066]: _type = "Task" [ 880.756764] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.786319] env[62066]: DEBUG nova.network.neutron [None req-e9fb52ef-3f9c-460e-85ba-59eb989b863f tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] [instance: 2634fcc0-96bd-4513-8f92-515f83fe23d7] Updating instance_info_cache with network_info: [{"id": "cf42ffe4-6621-4c29-9a6d-687ade676346", "address": "fa:16:3e:4c:ae:2f", "network": {"id": "fac1b86c-746c-4b6e-8d7f-21fe6ed404a3", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1827773279-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "59855b1cf91d43e3ba0482f414c4e546", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "88651df2-0506-4f6c-b868-dd30a81f2b1c", "external-id": "nsx-vlan-transportzone-366", "segmentation_id": 366, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcf42ffe4-66", "ovs_interfaceid": "cf42ffe4-6621-4c29-9a6d-687ade676346", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 880.788458] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51228392-cf4d-479c-9bbf-8028748ea83a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.794188] env[62066]: DEBUG oslo_vmware.api [None req-4a3c2cb7-f1e2-437a-b493-e5de08e55742 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Task: {'id': task-1341036, 'name': Destroy_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.799894] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-352d1676-d4e8-47f2-a4ea-f836eda77a66 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.819317] env[62066]: DEBUG nova.compute.provider_tree [None req-535b520b-dc22-40fb-a872-f0522e7e73a7 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 880.896361] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ebe8ad74-3036-4fe0-ab5d-5d9cbfc986ca tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Acquiring lock "170e184a-f009-42cf-82dc-eb9696a00a24" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 880.896701] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ebe8ad74-3036-4fe0-ab5d-5d9cbfc986ca tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Lock "170e184a-f009-42cf-82dc-eb9696a00a24" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 880.896965] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ebe8ad74-3036-4fe0-ab5d-5d9cbfc986ca tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Acquiring lock "170e184a-f009-42cf-82dc-eb9696a00a24-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 880.897227] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ebe8ad74-3036-4fe0-ab5d-5d9cbfc986ca tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Lock "170e184a-f009-42cf-82dc-eb9696a00a24-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 880.897418] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ebe8ad74-3036-4fe0-ab5d-5d9cbfc986ca tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Lock "170e184a-f009-42cf-82dc-eb9696a00a24-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 880.900871] env[62066]: INFO nova.compute.manager [None req-ebe8ad74-3036-4fe0-ab5d-5d9cbfc986ca tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 170e184a-f009-42cf-82dc-eb9696a00a24] Terminating instance [ 880.903235] env[62066]: DEBUG nova.compute.manager [None req-ebe8ad74-3036-4fe0-ab5d-5d9cbfc986ca tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 170e184a-f009-42cf-82dc-eb9696a00a24] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 880.903490] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-ebe8ad74-3036-4fe0-ab5d-5d9cbfc986ca tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 170e184a-f009-42cf-82dc-eb9696a00a24] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 880.904353] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3127a2e-4b94-409b-9137-8b4ac048ade5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.913616] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-ebe8ad74-3036-4fe0-ab5d-5d9cbfc986ca tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 170e184a-f009-42cf-82dc-eb9696a00a24] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 880.913949] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5ff27f33-bdde-45cc-bba9-f24a5f2f671e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.922311] env[62066]: DEBUG oslo_vmware.api [None req-ebe8ad74-3036-4fe0-ab5d-5d9cbfc986ca tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Waiting for the task: (returnval){ [ 880.922311] env[62066]: value = "task-1341037" [ 880.922311] env[62066]: _type = "Task" [ 880.922311] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.932154] env[62066]: DEBUG oslo_vmware.api [None req-ebe8ad74-3036-4fe0-ab5d-5d9cbfc986ca tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': task-1341037, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.085476] env[62066]: DEBUG nova.network.neutron [None req-f0861cc1-6af6-4429-bc89-cb4b7f228490 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 8807ce4d-532f-469c-a302-464c61c7efeb] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 881.155390] env[62066]: DEBUG oslo_vmware.api [None req-4ccd1f86-3a5c-4e35-bdaa-b8e507c9718e tempest-ServersTestFqdnHostnames-1367628746 tempest-ServersTestFqdnHostnames-1367628746-project-member] Task: {'id': task-1341035, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.184502] env[62066]: DEBUG oslo_vmware.api [None req-1dc5d31e-1ccf-4559-b370-76e0d5f04709 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1341033, 'name': ReconfigVM_Task, 'duration_secs': 0.733465} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.187452] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-1dc5d31e-1ccf-4559-b370-76e0d5f04709 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 76306c3f-0674-40fe-9864-d82d11e4bb5e] Reconfigured VM instance instance-00000050 to attach disk [datastore2] 76306c3f-0674-40fe-9864-d82d11e4bb5e/76306c3f-0674-40fe-9864-d82d11e4bb5e.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 881.187452] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-69dedbbc-b8d5-42bd-a622-baaa6181c0ec {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.193304] env[62066]: DEBUG oslo_vmware.api [None req-1dc5d31e-1ccf-4559-b370-76e0d5f04709 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Waiting for the task: (returnval){ [ 881.193304] env[62066]: value = "task-1341038" [ 881.193304] env[62066]: _type = "Task" [ 881.193304] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.201906] env[62066]: DEBUG oslo_vmware.api [None req-1dc5d31e-1ccf-4559-b370-76e0d5f04709 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1341038, 'name': Rename_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.262360] env[62066]: DEBUG nova.network.neutron [None req-f0861cc1-6af6-4429-bc89-cb4b7f228490 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 8807ce4d-532f-469c-a302-464c61c7efeb] Updating instance_info_cache with network_info: [{"id": "dba295b3-a285-4890-b73b-ed681d77192a", "address": "fa:16:3e:36:70:cf", "network": {"id": "a3852490-7a1f-499c-813a-46cbf52318da", "bridge": "br-int", "label": "tempest-ImagesTestJSON-735840736-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "75f0352c852947369474c9e05766c584", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "50171613-b419-45e3-9ada-fcb6cd921428", "external-id": "nsx-vlan-transportzone-914", "segmentation_id": 914, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdba295b3-a2", "ovs_interfaceid": "dba295b3-a285-4890-b73b-ed681d77192a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 881.269570] env[62066]: DEBUG oslo_vmware.api [None req-4a3c2cb7-f1e2-437a-b493-e5de08e55742 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Task: {'id': task-1341036, 'name': Destroy_Task, 'duration_secs': 0.471641} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.269926] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-4a3c2cb7-f1e2-437a-b493-e5de08e55742 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] [instance: 8ff6fbb9-c90f-498d-9a85-d220a8c2f794] Destroyed the VM [ 881.270210] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-4a3c2cb7-f1e2-437a-b493-e5de08e55742 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] [instance: 8ff6fbb9-c90f-498d-9a85-d220a8c2f794] Deleting Snapshot of the VM instance {{(pid=62066) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 881.270492] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-38594774-f53e-487b-b0b2-8921758157da {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.277049] env[62066]: DEBUG oslo_vmware.api [None req-4a3c2cb7-f1e2-437a-b493-e5de08e55742 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Waiting for the task: (returnval){ [ 881.277049] env[62066]: value = "task-1341039" [ 881.277049] env[62066]: _type = "Task" [ 881.277049] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.285782] env[62066]: DEBUG oslo_vmware.api [None req-4a3c2cb7-f1e2-437a-b493-e5de08e55742 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Task: {'id': task-1341039, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.292099] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e9fb52ef-3f9c-460e-85ba-59eb989b863f tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Releasing lock "refresh_cache-2634fcc0-96bd-4513-8f92-515f83fe23d7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 881.320052] env[62066]: DEBUG nova.scheduler.client.report [None req-535b520b-dc22-40fb-a872-f0522e7e73a7 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 881.436554] env[62066]: DEBUG oslo_vmware.api [None req-ebe8ad74-3036-4fe0-ab5d-5d9cbfc986ca tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': task-1341037, 'name': PowerOffVM_Task, 'duration_secs': 0.23077} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.436554] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-ebe8ad74-3036-4fe0-ab5d-5d9cbfc986ca tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 170e184a-f009-42cf-82dc-eb9696a00a24] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 881.436554] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-ebe8ad74-3036-4fe0-ab5d-5d9cbfc986ca tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 170e184a-f009-42cf-82dc-eb9696a00a24] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 881.436554] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9d35269c-9c0a-4810-ac46-ec897810de36 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.518923] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-ebe8ad74-3036-4fe0-ab5d-5d9cbfc986ca tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 170e184a-f009-42cf-82dc-eb9696a00a24] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 881.519182] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-ebe8ad74-3036-4fe0-ab5d-5d9cbfc986ca tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 170e184a-f009-42cf-82dc-eb9696a00a24] Deleting contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 881.519366] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-ebe8ad74-3036-4fe0-ab5d-5d9cbfc986ca tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Deleting the datastore file [datastore2] 170e184a-f009-42cf-82dc-eb9696a00a24 {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 881.519632] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-56770dea-5170-4fe9-92d1-f49234f2d8be {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.525730] env[62066]: DEBUG oslo_vmware.api [None req-ebe8ad74-3036-4fe0-ab5d-5d9cbfc986ca tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Waiting for the task: (returnval){ [ 881.525730] env[62066]: value = "task-1341041" [ 881.525730] env[62066]: _type = "Task" [ 881.525730] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.532948] env[62066]: DEBUG oslo_vmware.api [None req-ebe8ad74-3036-4fe0-ab5d-5d9cbfc986ca tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': task-1341041, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.655119] env[62066]: DEBUG oslo_vmware.api [None req-4ccd1f86-3a5c-4e35-bdaa-b8e507c9718e tempest-ServersTestFqdnHostnames-1367628746 tempest-ServersTestFqdnHostnames-1367628746-project-member] Task: {'id': task-1341035, 'name': PowerOnVM_Task, 'duration_secs': 0.64917} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.655471] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-4ccd1f86-3a5c-4e35-bdaa-b8e507c9718e tempest-ServersTestFqdnHostnames-1367628746 tempest-ServersTestFqdnHostnames-1367628746-project-member] [instance: afaca059-fc4e-4de2-8fa0-d2f226ea7051] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 881.655599] env[62066]: INFO nova.compute.manager [None req-4ccd1f86-3a5c-4e35-bdaa-b8e507c9718e tempest-ServersTestFqdnHostnames-1367628746 tempest-ServersTestFqdnHostnames-1367628746-project-member] [instance: afaca059-fc4e-4de2-8fa0-d2f226ea7051] Took 8.87 seconds to spawn the instance on the hypervisor. [ 881.655782] env[62066]: DEBUG nova.compute.manager [None req-4ccd1f86-3a5c-4e35-bdaa-b8e507c9718e tempest-ServersTestFqdnHostnames-1367628746 tempest-ServersTestFqdnHostnames-1367628746-project-member] [instance: afaca059-fc4e-4de2-8fa0-d2f226ea7051] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 881.656573] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61a3ebe6-c15e-4578-94a9-229655dd7e4d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.701443] env[62066]: DEBUG oslo_vmware.api [None req-1dc5d31e-1ccf-4559-b370-76e0d5f04709 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1341038, 'name': Rename_Task, 'duration_secs': 0.276174} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.701768] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-1dc5d31e-1ccf-4559-b370-76e0d5f04709 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 76306c3f-0674-40fe-9864-d82d11e4bb5e] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 881.702125] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f8781b56-7c74-4ffc-a916-5057ec525619 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.710271] env[62066]: DEBUG oslo_vmware.api [None req-1dc5d31e-1ccf-4559-b370-76e0d5f04709 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Waiting for the task: (returnval){ [ 881.710271] env[62066]: value = "task-1341042" [ 881.710271] env[62066]: _type = "Task" [ 881.710271] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.716565] env[62066]: DEBUG oslo_vmware.api [None req-1dc5d31e-1ccf-4559-b370-76e0d5f04709 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1341042, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.765617] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f0861cc1-6af6-4429-bc89-cb4b7f228490 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Releasing lock "refresh_cache-8807ce4d-532f-469c-a302-464c61c7efeb" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 881.766125] env[62066]: DEBUG nova.compute.manager [None req-f0861cc1-6af6-4429-bc89-cb4b7f228490 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 8807ce4d-532f-469c-a302-464c61c7efeb] Instance network_info: |[{"id": "dba295b3-a285-4890-b73b-ed681d77192a", "address": "fa:16:3e:36:70:cf", "network": {"id": "a3852490-7a1f-499c-813a-46cbf52318da", "bridge": "br-int", "label": "tempest-ImagesTestJSON-735840736-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "75f0352c852947369474c9e05766c584", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "50171613-b419-45e3-9ada-fcb6cd921428", "external-id": "nsx-vlan-transportzone-914", "segmentation_id": 914, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdba295b3-a2", "ovs_interfaceid": "dba295b3-a285-4890-b73b-ed681d77192a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 881.766745] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-f0861cc1-6af6-4429-bc89-cb4b7f228490 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 8807ce4d-532f-469c-a302-464c61c7efeb] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:36:70:cf', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '50171613-b419-45e3-9ada-fcb6cd921428', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'dba295b3-a285-4890-b73b-ed681d77192a', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 881.778730] env[62066]: DEBUG oslo.service.loopingcall [None req-f0861cc1-6af6-4429-bc89-cb4b7f228490 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 881.779676] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8807ce4d-532f-469c-a302-464c61c7efeb] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 881.784421] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ec97e2de-1777-4004-8a7a-e96838820b36 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.814183] env[62066]: DEBUG oslo_vmware.api [None req-4a3c2cb7-f1e2-437a-b493-e5de08e55742 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Task: {'id': task-1341039, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.816095] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 881.816095] env[62066]: value = "task-1341043" [ 881.816095] env[62066]: _type = "Task" [ 881.816095] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.829414] env[62066]: DEBUG oslo_concurrency.lockutils [None req-535b520b-dc22-40fb-a872-f0522e7e73a7 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.834s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 881.832495] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1341043, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.833311] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7b75bfaa-cfbe-4d54-8646-a77cb9e92e69 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.998s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 881.833715] env[62066]: DEBUG nova.objects.instance [None req-7b75bfaa-cfbe-4d54-8646-a77cb9e92e69 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Lazy-loading 'resources' on Instance uuid 1a4b0637-1a56-41ef-b89b-6b56d24ed206 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 881.838913] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-e9fb52ef-3f9c-460e-85ba-59eb989b863f tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] [instance: 2634fcc0-96bd-4513-8f92-515f83fe23d7] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 881.839355] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9e0f3173-3562-402e-8585-4c159501e857 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.847850] env[62066]: DEBUG oslo_vmware.api [None req-e9fb52ef-3f9c-460e-85ba-59eb989b863f tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Waiting for the task: (returnval){ [ 881.847850] env[62066]: value = "task-1341044" [ 881.847850] env[62066]: _type = "Task" [ 881.847850] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.860992] env[62066]: DEBUG oslo_vmware.api [None req-e9fb52ef-3f9c-460e-85ba-59eb989b863f tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Task: {'id': task-1341044, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.862701] env[62066]: INFO nova.scheduler.client.report [None req-535b520b-dc22-40fb-a872-f0522e7e73a7 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Deleted allocations for instance 3120c9a5-0558-4e83-b0fb-994a989398fb [ 882.036101] env[62066]: DEBUG oslo_vmware.api [None req-ebe8ad74-3036-4fe0-ab5d-5d9cbfc986ca tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': task-1341041, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.151545} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.036442] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-ebe8ad74-3036-4fe0-ab5d-5d9cbfc986ca tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 882.036743] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-ebe8ad74-3036-4fe0-ab5d-5d9cbfc986ca tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 170e184a-f009-42cf-82dc-eb9696a00a24] Deleted contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 882.036834] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-ebe8ad74-3036-4fe0-ab5d-5d9cbfc986ca tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 170e184a-f009-42cf-82dc-eb9696a00a24] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 882.037018] env[62066]: INFO nova.compute.manager [None req-ebe8ad74-3036-4fe0-ab5d-5d9cbfc986ca tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 170e184a-f009-42cf-82dc-eb9696a00a24] Took 1.13 seconds to destroy the instance on the hypervisor. [ 882.037270] env[62066]: DEBUG oslo.service.loopingcall [None req-ebe8ad74-3036-4fe0-ab5d-5d9cbfc986ca tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 882.037471] env[62066]: DEBUG nova.compute.manager [-] [instance: 170e184a-f009-42cf-82dc-eb9696a00a24] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 882.037565] env[62066]: DEBUG nova.network.neutron [-] [instance: 170e184a-f009-42cf-82dc-eb9696a00a24] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 882.113963] env[62066]: DEBUG nova.compute.manager [req-7548aacc-b6a2-4e2a-a307-1ebeaff9ed4a req-6d53f90f-0de3-4930-86be-0a3b22d9a285 service nova] [instance: 8807ce4d-532f-469c-a302-464c61c7efeb] Received event network-changed-dba295b3-a285-4890-b73b-ed681d77192a {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 882.114330] env[62066]: DEBUG nova.compute.manager [req-7548aacc-b6a2-4e2a-a307-1ebeaff9ed4a req-6d53f90f-0de3-4930-86be-0a3b22d9a285 service nova] [instance: 8807ce4d-532f-469c-a302-464c61c7efeb] Refreshing instance network info cache due to event network-changed-dba295b3-a285-4890-b73b-ed681d77192a. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 882.114445] env[62066]: DEBUG oslo_concurrency.lockutils [req-7548aacc-b6a2-4e2a-a307-1ebeaff9ed4a req-6d53f90f-0de3-4930-86be-0a3b22d9a285 service nova] Acquiring lock "refresh_cache-8807ce4d-532f-469c-a302-464c61c7efeb" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 882.114748] env[62066]: DEBUG oslo_concurrency.lockutils [req-7548aacc-b6a2-4e2a-a307-1ebeaff9ed4a req-6d53f90f-0de3-4930-86be-0a3b22d9a285 service nova] Acquired lock "refresh_cache-8807ce4d-532f-469c-a302-464c61c7efeb" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 882.114951] env[62066]: DEBUG nova.network.neutron [req-7548aacc-b6a2-4e2a-a307-1ebeaff9ed4a req-6d53f90f-0de3-4930-86be-0a3b22d9a285 service nova] [instance: 8807ce4d-532f-469c-a302-464c61c7efeb] Refreshing network info cache for port dba295b3-a285-4890-b73b-ed681d77192a {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 882.184348] env[62066]: INFO nova.compute.manager [None req-4ccd1f86-3a5c-4e35-bdaa-b8e507c9718e tempest-ServersTestFqdnHostnames-1367628746 tempest-ServersTestFqdnHostnames-1367628746-project-member] [instance: afaca059-fc4e-4de2-8fa0-d2f226ea7051] Took 25.85 seconds to build instance. [ 882.225016] env[62066]: DEBUG oslo_vmware.api [None req-1dc5d31e-1ccf-4559-b370-76e0d5f04709 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1341042, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.290712] env[62066]: DEBUG oslo_vmware.api [None req-4a3c2cb7-f1e2-437a-b493-e5de08e55742 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Task: {'id': task-1341039, 'name': RemoveSnapshot_Task, 'duration_secs': 0.718219} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.291030] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-4a3c2cb7-f1e2-437a-b493-e5de08e55742 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] [instance: 8ff6fbb9-c90f-498d-9a85-d220a8c2f794] Deleted Snapshot of the VM instance {{(pid=62066) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 882.291346] env[62066]: DEBUG nova.compute.manager [None req-4a3c2cb7-f1e2-437a-b493-e5de08e55742 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] [instance: 8ff6fbb9-c90f-498d-9a85-d220a8c2f794] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 882.292196] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5eb2aafa-23ad-45db-89b9-28e6ef9beb0d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.329385] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1341043, 'name': CreateVM_Task, 'duration_secs': 0.405941} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.329636] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8807ce4d-532f-469c-a302-464c61c7efeb] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 882.330611] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f0861cc1-6af6-4429-bc89-cb4b7f228490 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/1147d77f-02f3-4d4b-a853-4a0e7d8b0a96" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 882.330904] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f0861cc1-6af6-4429-bc89-cb4b7f228490 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Acquired lock "[datastore2] devstack-image-cache_base/1147d77f-02f3-4d4b-a853-4a0e7d8b0a96" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 882.331424] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f0861cc1-6af6-4429-bc89-cb4b7f228490 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/1147d77f-02f3-4d4b-a853-4a0e7d8b0a96" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 882.332211] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-27f599d6-b0cd-456b-8770-4566ffd58e6c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.341351] env[62066]: DEBUG oslo_vmware.api [None req-f0861cc1-6af6-4429-bc89-cb4b7f228490 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Waiting for the task: (returnval){ [ 882.341351] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]5229c096-90e5-7919-095f-91cf1c4a1031" [ 882.341351] env[62066]: _type = "Task" [ 882.341351] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.358032] env[62066]: DEBUG oslo_vmware.api [None req-f0861cc1-6af6-4429-bc89-cb4b7f228490 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]5229c096-90e5-7919-095f-91cf1c4a1031, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.368067] env[62066]: DEBUG oslo_vmware.api [None req-e9fb52ef-3f9c-460e-85ba-59eb989b863f tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Task: {'id': task-1341044, 'name': PowerOffVM_Task, 'duration_secs': 0.23581} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.368463] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-e9fb52ef-3f9c-460e-85ba-59eb989b863f tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] [instance: 2634fcc0-96bd-4513-8f92-515f83fe23d7] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 882.369592] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41d3c4f2-ebf4-45b9-8c76-b84af2dbc303 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.380229] env[62066]: DEBUG oslo_concurrency.lockutils [None req-535b520b-dc22-40fb-a872-f0522e7e73a7 tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Lock "3120c9a5-0558-4e83-b0fb-994a989398fb" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.830s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 882.400775] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9198bf43-e8d7-4b2c-8972-52367c15034d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.442332] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-e9fb52ef-3f9c-460e-85ba-59eb989b863f tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] [instance: 2634fcc0-96bd-4513-8f92-515f83fe23d7] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 882.442857] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ce021737-60a1-44c3-89cc-00dd253ae1ec {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.452283] env[62066]: DEBUG oslo_vmware.api [None req-e9fb52ef-3f9c-460e-85ba-59eb989b863f tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Waiting for the task: (returnval){ [ 882.452283] env[62066]: value = "task-1341045" [ 882.452283] env[62066]: _type = "Task" [ 882.452283] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.461949] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-e9fb52ef-3f9c-460e-85ba-59eb989b863f tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] [instance: 2634fcc0-96bd-4513-8f92-515f83fe23d7] VM already powered off {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 882.462220] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-e9fb52ef-3f9c-460e-85ba-59eb989b863f tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] [instance: 2634fcc0-96bd-4513-8f92-515f83fe23d7] Processing image 50ff584c-3b50-4395-af07-3e66769bc9f7 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 882.462474] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e9fb52ef-3f9c-460e-85ba-59eb989b863f tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 882.462631] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e9fb52ef-3f9c-460e-85ba-59eb989b863f tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Acquired lock "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 882.462839] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-e9fb52ef-3f9c-460e-85ba-59eb989b863f tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 882.465371] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-312d8ea0-9584-4da4-84ae-dfaa2079edb9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.473738] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-e9fb52ef-3f9c-460e-85ba-59eb989b863f tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 882.474242] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-e9fb52ef-3f9c-460e-85ba-59eb989b863f tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 882.474800] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-beacae01-c981-408b-a2c7-00314d019ba6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.482543] env[62066]: DEBUG oslo_vmware.api [None req-e9fb52ef-3f9c-460e-85ba-59eb989b863f tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Waiting for the task: (returnval){ [ 882.482543] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]52ededde-8289-86b4-a7ef-e7c0b7d8b986" [ 882.482543] env[62066]: _type = "Task" [ 882.482543] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.489804] env[62066]: DEBUG oslo_vmware.api [None req-e9fb52ef-3f9c-460e-85ba-59eb989b863f tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52ededde-8289-86b4-a7ef-e7c0b7d8b986, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.586420] env[62066]: DEBUG oslo_concurrency.lockutils [None req-64742fb3-cc48-46bc-b10b-6ecbf4d3709b tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Acquiring lock "d77e3405-d158-4a8c-9f54-d5fbd26fed48" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 882.586667] env[62066]: DEBUG oslo_concurrency.lockutils [None req-64742fb3-cc48-46bc-b10b-6ecbf4d3709b tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Lock "d77e3405-d158-4a8c-9f54-d5fbd26fed48" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 882.586871] env[62066]: DEBUG oslo_concurrency.lockutils [None req-64742fb3-cc48-46bc-b10b-6ecbf4d3709b tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Acquiring lock "d77e3405-d158-4a8c-9f54-d5fbd26fed48-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 882.587065] env[62066]: DEBUG oslo_concurrency.lockutils [None req-64742fb3-cc48-46bc-b10b-6ecbf4d3709b tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Lock "d77e3405-d158-4a8c-9f54-d5fbd26fed48-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 882.587236] env[62066]: DEBUG oslo_concurrency.lockutils [None req-64742fb3-cc48-46bc-b10b-6ecbf4d3709b tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Lock "d77e3405-d158-4a8c-9f54-d5fbd26fed48-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 882.593257] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdeb8c1f-87c5-4567-80c6-7f4214dcfd32 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.598966] env[62066]: INFO nova.compute.manager [None req-64742fb3-cc48-46bc-b10b-6ecbf4d3709b tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] [instance: d77e3405-d158-4a8c-9f54-d5fbd26fed48] Terminating instance [ 882.604047] env[62066]: DEBUG oslo_concurrency.lockutils [None req-64742fb3-cc48-46bc-b10b-6ecbf4d3709b tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Acquiring lock "refresh_cache-d77e3405-d158-4a8c-9f54-d5fbd26fed48" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 882.604223] env[62066]: DEBUG oslo_concurrency.lockutils [None req-64742fb3-cc48-46bc-b10b-6ecbf4d3709b tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Acquired lock "refresh_cache-d77e3405-d158-4a8c-9f54-d5fbd26fed48" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 882.604390] env[62066]: DEBUG nova.network.neutron [None req-64742fb3-cc48-46bc-b10b-6ecbf4d3709b tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] [instance: d77e3405-d158-4a8c-9f54-d5fbd26fed48] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 882.611720] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09200b37-5d25-4ac3-aad6-0dc45c76a595 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.642631] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1f57124-ecce-4ea2-9c61-1b15c7372b89 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.653785] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4035ab87-171b-470e-a524-5aa62472e173 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.669707] env[62066]: DEBUG nova.compute.provider_tree [None req-7b75bfaa-cfbe-4d54-8646-a77cb9e92e69 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 882.690777] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4ccd1f86-3a5c-4e35-bdaa-b8e507c9718e tempest-ServersTestFqdnHostnames-1367628746 tempest-ServersTestFqdnHostnames-1367628746-project-member] Lock "afaca059-fc4e-4de2-8fa0-d2f226ea7051" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 27.379s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 882.722146] env[62066]: DEBUG oslo_vmware.api [None req-1dc5d31e-1ccf-4559-b370-76e0d5f04709 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1341042, 'name': PowerOnVM_Task, 'duration_secs': 0.635372} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.722343] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-1dc5d31e-1ccf-4559-b370-76e0d5f04709 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 76306c3f-0674-40fe-9864-d82d11e4bb5e] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 882.722534] env[62066]: DEBUG nova.compute.manager [None req-1dc5d31e-1ccf-4559-b370-76e0d5f04709 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 76306c3f-0674-40fe-9864-d82d11e4bb5e] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 882.726084] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e462855d-bbe2-42bd-9754-4896451da927 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.812252] env[62066]: INFO nova.compute.manager [None req-4a3c2cb7-f1e2-437a-b493-e5de08e55742 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] [instance: 8ff6fbb9-c90f-498d-9a85-d220a8c2f794] Shelve offloading [ 882.814090] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-4a3c2cb7-f1e2-437a-b493-e5de08e55742 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] [instance: 8ff6fbb9-c90f-498d-9a85-d220a8c2f794] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 882.814351] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ea05b4de-c1f4-4fa8-a97d-8395d4918de2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.825217] env[62066]: DEBUG oslo_vmware.api [None req-4a3c2cb7-f1e2-437a-b493-e5de08e55742 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Waiting for the task: (returnval){ [ 882.825217] env[62066]: value = "task-1341046" [ 882.825217] env[62066]: _type = "Task" [ 882.825217] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.840786] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-4a3c2cb7-f1e2-437a-b493-e5de08e55742 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] [instance: 8ff6fbb9-c90f-498d-9a85-d220a8c2f794] VM already powered off {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 882.841058] env[62066]: DEBUG nova.compute.manager [None req-4a3c2cb7-f1e2-437a-b493-e5de08e55742 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] [instance: 8ff6fbb9-c90f-498d-9a85-d220a8c2f794] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 882.842428] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf3ad0ff-c8f9-4c5e-ac66-cdbd5fd51d99 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.852129] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4a3c2cb7-f1e2-437a-b493-e5de08e55742 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Acquiring lock "refresh_cache-8ff6fbb9-c90f-498d-9a85-d220a8c2f794" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 882.852351] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4a3c2cb7-f1e2-437a-b493-e5de08e55742 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Acquired lock "refresh_cache-8ff6fbb9-c90f-498d-9a85-d220a8c2f794" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 882.852529] env[62066]: DEBUG nova.network.neutron [None req-4a3c2cb7-f1e2-437a-b493-e5de08e55742 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] [instance: 8ff6fbb9-c90f-498d-9a85-d220a8c2f794] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 882.858227] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f0861cc1-6af6-4429-bc89-cb4b7f228490 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Releasing lock "[datastore2] devstack-image-cache_base/1147d77f-02f3-4d4b-a853-4a0e7d8b0a96" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 882.858488] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-f0861cc1-6af6-4429-bc89-cb4b7f228490 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 8807ce4d-532f-469c-a302-464c61c7efeb] Processing image 1147d77f-02f3-4d4b-a853-4a0e7d8b0a96 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 882.858721] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f0861cc1-6af6-4429-bc89-cb4b7f228490 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/1147d77f-02f3-4d4b-a853-4a0e7d8b0a96/1147d77f-02f3-4d4b-a853-4a0e7d8b0a96.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 882.858867] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f0861cc1-6af6-4429-bc89-cb4b7f228490 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Acquired lock "[datastore2] devstack-image-cache_base/1147d77f-02f3-4d4b-a853-4a0e7d8b0a96/1147d77f-02f3-4d4b-a853-4a0e7d8b0a96.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 882.859053] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-f0861cc1-6af6-4429-bc89-cb4b7f228490 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 882.861193] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ea5d8785-aa4e-45af-afa9-4148aab5cd51 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.870347] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-f0861cc1-6af6-4429-bc89-cb4b7f228490 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 882.870347] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-f0861cc1-6af6-4429-bc89-cb4b7f228490 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 882.871086] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b16e17ec-eaef-4ed9-b279-0a4f8cee3772 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.876470] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-98bb85f9-df22-41ab-a6e4-7c41686713f5 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] [instance: de2e2be7-efdb-45a8-842a-640ab9deb1d9] Volume attach. Driver type: vmdk {{(pid=62066) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 882.876697] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-98bb85f9-df22-41ab-a6e4-7c41686713f5 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] [instance: de2e2be7-efdb-45a8-842a-640ab9deb1d9] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-286087', 'volume_id': '36cdc9f5-eb4f-44b9-9aac-a7cd3fc5d570', 'name': 'volume-36cdc9f5-eb4f-44b9-9aac-a7cd3fc5d570', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'de2e2be7-efdb-45a8-842a-640ab9deb1d9', 'attached_at': '', 'detached_at': '', 'volume_id': '36cdc9f5-eb4f-44b9-9aac-a7cd3fc5d570', 'serial': '36cdc9f5-eb4f-44b9-9aac-a7cd3fc5d570'} {{(pid=62066) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 882.877759] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37229afc-ea93-43f4-8168-010ce542179f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.881017] env[62066]: DEBUG nova.network.neutron [req-7548aacc-b6a2-4e2a-a307-1ebeaff9ed4a req-6d53f90f-0de3-4930-86be-0a3b22d9a285 service nova] [instance: 8807ce4d-532f-469c-a302-464c61c7efeb] Updated VIF entry in instance network info cache for port dba295b3-a285-4890-b73b-ed681d77192a. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 882.885601] env[62066]: DEBUG nova.network.neutron [req-7548aacc-b6a2-4e2a-a307-1ebeaff9ed4a req-6d53f90f-0de3-4930-86be-0a3b22d9a285 service nova] [instance: 8807ce4d-532f-469c-a302-464c61c7efeb] Updating instance_info_cache with network_info: [{"id": "dba295b3-a285-4890-b73b-ed681d77192a", "address": "fa:16:3e:36:70:cf", "network": {"id": "a3852490-7a1f-499c-813a-46cbf52318da", "bridge": "br-int", "label": "tempest-ImagesTestJSON-735840736-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "75f0352c852947369474c9e05766c584", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "50171613-b419-45e3-9ada-fcb6cd921428", "external-id": "nsx-vlan-transportzone-914", "segmentation_id": 914, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdba295b3-a2", "ovs_interfaceid": "dba295b3-a285-4890-b73b-ed681d77192a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 882.901818] env[62066]: DEBUG oslo_vmware.api [None req-f0861cc1-6af6-4429-bc89-cb4b7f228490 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Waiting for the task: (returnval){ [ 882.901818] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]5238b993-566a-1a0a-e514-a36f3d760227" [ 882.901818] env[62066]: _type = "Task" [ 882.901818] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.903016] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e66a3803-db47-4160-a3f6-8f12f7e793f6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.934720] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-f0861cc1-6af6-4429-bc89-cb4b7f228490 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 8807ce4d-532f-469c-a302-464c61c7efeb] Preparing fetch location {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 882.935045] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-f0861cc1-6af6-4429-bc89-cb4b7f228490 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 8807ce4d-532f-469c-a302-464c61c7efeb] Fetch image to [datastore2] OSTACK_IMG_1e3b548c-3176-483b-8156-8e6eb27c2d80/OSTACK_IMG_1e3b548c-3176-483b-8156-8e6eb27c2d80.vmdk {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 882.935265] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-f0861cc1-6af6-4429-bc89-cb4b7f228490 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 8807ce4d-532f-469c-a302-464c61c7efeb] Downloading stream optimized image 1147d77f-02f3-4d4b-a853-4a0e7d8b0a96 to [datastore2] OSTACK_IMG_1e3b548c-3176-483b-8156-8e6eb27c2d80/OSTACK_IMG_1e3b548c-3176-483b-8156-8e6eb27c2d80.vmdk on the data store datastore2 as vApp {{(pid=62066) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 882.935461] env[62066]: DEBUG nova.virt.vmwareapi.images [None req-f0861cc1-6af6-4429-bc89-cb4b7f228490 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 8807ce4d-532f-469c-a302-464c61c7efeb] Downloading image file data 1147d77f-02f3-4d4b-a853-4a0e7d8b0a96 to the ESX as VM named 'OSTACK_IMG_1e3b548c-3176-483b-8156-8e6eb27c2d80' {{(pid=62066) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 882.947393] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-98bb85f9-df22-41ab-a6e4-7c41686713f5 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] [instance: de2e2be7-efdb-45a8-842a-640ab9deb1d9] Reconfiguring VM instance instance-00000048 to attach disk [datastore2] volume-36cdc9f5-eb4f-44b9-9aac-a7cd3fc5d570/volume-36cdc9f5-eb4f-44b9-9aac-a7cd3fc5d570.vmdk or device None with type thin {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 882.949150] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4ecd85a6-6b2c-47f8-bdee-6e5649900aba {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.982861] env[62066]: DEBUG oslo_vmware.api [None req-98bb85f9-df22-41ab-a6e4-7c41686713f5 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Waiting for the task: (returnval){ [ 882.982861] env[62066]: value = "task-1341047" [ 882.982861] env[62066]: _type = "Task" [ 882.982861] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.018944] env[62066]: DEBUG oslo_vmware.api [None req-98bb85f9-df22-41ab-a6e4-7c41686713f5 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Task: {'id': task-1341047, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.019536] env[62066]: DEBUG oslo_vmware.api [None req-e9fb52ef-3f9c-460e-85ba-59eb989b863f tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52ededde-8289-86b4-a7ef-e7c0b7d8b986, 'name': SearchDatastore_Task, 'duration_secs': 0.008909} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.020046] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ffae9302-a2c3-4776-8a32-9c59d778a6d9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.025124] env[62066]: DEBUG oslo_vmware.api [None req-e9fb52ef-3f9c-460e-85ba-59eb989b863f tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Waiting for the task: (returnval){ [ 883.025124] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]520e3593-a5c8-83d8-5460-8e5684d3b5fc" [ 883.025124] env[62066]: _type = "Task" [ 883.025124] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.034560] env[62066]: DEBUG oslo_vmware.api [None req-e9fb52ef-3f9c-460e-85ba-59eb989b863f tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]520e3593-a5c8-83d8-5460-8e5684d3b5fc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.054867] env[62066]: DEBUG oslo_vmware.rw_handles [None req-f0861cc1-6af6-4429-bc89-cb4b7f228490 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 883.054867] env[62066]: value = "resgroup-9" [ 883.054867] env[62066]: _type = "ResourcePool" [ 883.054867] env[62066]: }. {{(pid=62066) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 883.054867] env[62066]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-64c6db3f-4432-4a50-a907-0b4739387d73 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.076111] env[62066]: DEBUG oslo_vmware.rw_handles [None req-f0861cc1-6af6-4429-bc89-cb4b7f228490 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Lease: (returnval){ [ 883.076111] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]522d5280-29f5-0135-dadd-343cfece8900" [ 883.076111] env[62066]: _type = "HttpNfcLease" [ 883.076111] env[62066]: } obtained for vApp import into resource pool (val){ [ 883.076111] env[62066]: value = "resgroup-9" [ 883.076111] env[62066]: _type = "ResourcePool" [ 883.076111] env[62066]: }. {{(pid=62066) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 883.076343] env[62066]: DEBUG oslo_vmware.api [None req-f0861cc1-6af6-4429-bc89-cb4b7f228490 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Waiting for the lease: (returnval){ [ 883.076343] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]522d5280-29f5-0135-dadd-343cfece8900" [ 883.076343] env[62066]: _type = "HttpNfcLease" [ 883.076343] env[62066]: } to be ready. {{(pid=62066) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 883.083744] env[62066]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 883.083744] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]522d5280-29f5-0135-dadd-343cfece8900" [ 883.083744] env[62066]: _type = "HttpNfcLease" [ 883.083744] env[62066]: } is initializing. {{(pid=62066) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 883.125280] env[62066]: DEBUG nova.network.neutron [None req-64742fb3-cc48-46bc-b10b-6ecbf4d3709b tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] [instance: d77e3405-d158-4a8c-9f54-d5fbd26fed48] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 883.175642] env[62066]: DEBUG nova.scheduler.client.report [None req-7b75bfaa-cfbe-4d54-8646-a77cb9e92e69 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 883.196270] env[62066]: DEBUG nova.network.neutron [None req-64742fb3-cc48-46bc-b10b-6ecbf4d3709b tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] [instance: d77e3405-d158-4a8c-9f54-d5fbd26fed48] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 883.241958] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1dc5d31e-1ccf-4559-b370-76e0d5f04709 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 883.384129] env[62066]: DEBUG oslo_concurrency.lockutils [req-7548aacc-b6a2-4e2a-a307-1ebeaff9ed4a req-6d53f90f-0de3-4930-86be-0a3b22d9a285 service nova] Releasing lock "refresh_cache-8807ce4d-532f-469c-a302-464c61c7efeb" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 883.495859] env[62066]: DEBUG oslo_vmware.api [None req-98bb85f9-df22-41ab-a6e4-7c41686713f5 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Task: {'id': task-1341047, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.535338] env[62066]: DEBUG oslo_vmware.api [None req-e9fb52ef-3f9c-460e-85ba-59eb989b863f tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]520e3593-a5c8-83d8-5460-8e5684d3b5fc, 'name': SearchDatastore_Task, 'duration_secs': 0.019236} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.535628] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e9fb52ef-3f9c-460e-85ba-59eb989b863f tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Releasing lock "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 883.535896] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-e9fb52ef-3f9c-460e-85ba-59eb989b863f tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Copying virtual disk from [datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk to [datastore1] 2634fcc0-96bd-4513-8f92-515f83fe23d7/50ff584c-3b50-4395-af07-3e66769bc9f7-rescue.vmdk. {{(pid=62066) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 883.536177] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1c8769df-24c7-49a9-8029-c94c0657e64d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.542912] env[62066]: DEBUG oslo_vmware.api [None req-e9fb52ef-3f9c-460e-85ba-59eb989b863f tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Waiting for the task: (returnval){ [ 883.542912] env[62066]: value = "task-1341049" [ 883.542912] env[62066]: _type = "Task" [ 883.542912] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.551621] env[62066]: DEBUG oslo_vmware.api [None req-e9fb52ef-3f9c-460e-85ba-59eb989b863f tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Task: {'id': task-1341049, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.589333] env[62066]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 883.589333] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]522d5280-29f5-0135-dadd-343cfece8900" [ 883.589333] env[62066]: _type = "HttpNfcLease" [ 883.589333] env[62066]: } is initializing. {{(pid=62066) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 883.681532] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7b75bfaa-cfbe-4d54-8646-a77cb9e92e69 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.848s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 883.683913] env[62066]: DEBUG oslo_concurrency.lockutils [None req-09aaefa9-33e0-4c34-9b0e-8039240cf379 tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.221s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 883.684294] env[62066]: DEBUG nova.objects.instance [None req-09aaefa9-33e0-4c34-9b0e-8039240cf379 tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] Lazy-loading 'resources' on Instance uuid 0bc31a8b-a5fc-4f6e-a4c0-024bf4260755 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 883.698605] env[62066]: DEBUG oslo_concurrency.lockutils [None req-64742fb3-cc48-46bc-b10b-6ecbf4d3709b tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Releasing lock "refresh_cache-d77e3405-d158-4a8c-9f54-d5fbd26fed48" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 883.699027] env[62066]: DEBUG nova.compute.manager [None req-64742fb3-cc48-46bc-b10b-6ecbf4d3709b tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] [instance: d77e3405-d158-4a8c-9f54-d5fbd26fed48] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 883.699250] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-64742fb3-cc48-46bc-b10b-6ecbf4d3709b tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] [instance: d77e3405-d158-4a8c-9f54-d5fbd26fed48] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 883.700602] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f14c7a0a-4113-4ec0-b24b-757044f9ddfd {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.708109] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-64742fb3-cc48-46bc-b10b-6ecbf4d3709b tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] [instance: d77e3405-d158-4a8c-9f54-d5fbd26fed48] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 883.708359] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6199c2cb-dd6b-457f-91c2-4987cd22c634 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.711435] env[62066]: INFO nova.scheduler.client.report [None req-7b75bfaa-cfbe-4d54-8646-a77cb9e92e69 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Deleted allocations for instance 1a4b0637-1a56-41ef-b89b-6b56d24ed206 [ 883.716389] env[62066]: DEBUG oslo_vmware.api [None req-64742fb3-cc48-46bc-b10b-6ecbf4d3709b tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Waiting for the task: (returnval){ [ 883.716389] env[62066]: value = "task-1341050" [ 883.716389] env[62066]: _type = "Task" [ 883.716389] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.719918] env[62066]: DEBUG nova.network.neutron [-] [instance: 170e184a-f009-42cf-82dc-eb9696a00a24] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 883.726252] env[62066]: DEBUG oslo_vmware.api [None req-64742fb3-cc48-46bc-b10b-6ecbf4d3709b tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Task: {'id': task-1341050, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.893539] env[62066]: DEBUG nova.network.neutron [None req-4a3c2cb7-f1e2-437a-b493-e5de08e55742 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] [instance: 8ff6fbb9-c90f-498d-9a85-d220a8c2f794] Updating instance_info_cache with network_info: [{"id": "8c324833-7bf9-4007-85aa-5b16871f63c7", "address": "fa:16:3e:00:85:34", "network": {"id": "0e17b5e2-526e-4d4c-a463-5e34408044cf", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1588316661-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.184", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1f05d8a66d0c4a479e49a947e4b4bbff", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ffcecdaa-a7b8-49fc-9371-dbdb7744688e", "external-id": "nsx-vlan-transportzone-994", "segmentation_id": 994, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8c324833-7b", "ovs_interfaceid": "8c324833-7bf9-4007-85aa-5b16871f63c7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 883.994626] env[62066]: DEBUG oslo_vmware.api [None req-98bb85f9-df22-41ab-a6e4-7c41686713f5 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Task: {'id': task-1341047, 'name': ReconfigVM_Task, 'duration_secs': 0.560331} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.995030] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-98bb85f9-df22-41ab-a6e4-7c41686713f5 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] [instance: de2e2be7-efdb-45a8-842a-640ab9deb1d9] Reconfigured VM instance instance-00000048 to attach disk [datastore2] volume-36cdc9f5-eb4f-44b9-9aac-a7cd3fc5d570/volume-36cdc9f5-eb4f-44b9-9aac-a7cd3fc5d570.vmdk or device None with type thin {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 884.000841] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-95b5ebc2-c46d-4645-8519-996975e8f34c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.017046] env[62066]: DEBUG oslo_vmware.api [None req-98bb85f9-df22-41ab-a6e4-7c41686713f5 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Waiting for the task: (returnval){ [ 884.017046] env[62066]: value = "task-1341051" [ 884.017046] env[62066]: _type = "Task" [ 884.017046] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.027149] env[62066]: DEBUG oslo_vmware.api [None req-98bb85f9-df22-41ab-a6e4-7c41686713f5 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Task: {'id': task-1341051, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.053896] env[62066]: DEBUG oslo_vmware.api [None req-e9fb52ef-3f9c-460e-85ba-59eb989b863f tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Task: {'id': task-1341049, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.085199] env[62066]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 884.085199] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]522d5280-29f5-0135-dadd-343cfece8900" [ 884.085199] env[62066]: _type = "HttpNfcLease" [ 884.085199] env[62066]: } is ready. {{(pid=62066) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 884.085611] env[62066]: DEBUG oslo_vmware.rw_handles [None req-f0861cc1-6af6-4429-bc89-cb4b7f228490 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 884.085611] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]522d5280-29f5-0135-dadd-343cfece8900" [ 884.085611] env[62066]: _type = "HttpNfcLease" [ 884.085611] env[62066]: }. {{(pid=62066) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 884.087090] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66ece746-766d-4746-8b5b-443af7e136b3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.096282] env[62066]: DEBUG oslo_vmware.rw_handles [None req-f0861cc1-6af6-4429-bc89-cb4b7f228490 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/523770ef-9b20-2a18-c98f-52042f3a2529/disk-0.vmdk from lease info. {{(pid=62066) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 884.096551] env[62066]: DEBUG oslo_vmware.rw_handles [None req-f0861cc1-6af6-4429-bc89-cb4b7f228490 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Creating HTTP connection to write to file with size = 21334016 and URL = https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/523770ef-9b20-2a18-c98f-52042f3a2529/disk-0.vmdk. {{(pid=62066) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 884.166364] env[62066]: DEBUG nova.compute.manager [req-b17f0189-f469-4847-929a-b5503930f6fa req-6ce39cee-116a-472a-bb9c-24cba98972fb service nova] [instance: 170e184a-f009-42cf-82dc-eb9696a00a24] Received event network-vif-deleted-e8281131-9972-486f-b3be-2452487f99a3 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 884.166602] env[62066]: DEBUG nova.compute.manager [req-b17f0189-f469-4847-929a-b5503930f6fa req-6ce39cee-116a-472a-bb9c-24cba98972fb service nova] [instance: afaca059-fc4e-4de2-8fa0-d2f226ea7051] Received event network-changed-725b6138-b39f-41c3-bf84-1a13a3e30fa9 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 884.166749] env[62066]: DEBUG nova.compute.manager [req-b17f0189-f469-4847-929a-b5503930f6fa req-6ce39cee-116a-472a-bb9c-24cba98972fb service nova] [instance: afaca059-fc4e-4de2-8fa0-d2f226ea7051] Refreshing instance network info cache due to event network-changed-725b6138-b39f-41c3-bf84-1a13a3e30fa9. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 884.166990] env[62066]: DEBUG oslo_concurrency.lockutils [req-b17f0189-f469-4847-929a-b5503930f6fa req-6ce39cee-116a-472a-bb9c-24cba98972fb service nova] Acquiring lock "refresh_cache-afaca059-fc4e-4de2-8fa0-d2f226ea7051" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 884.167102] env[62066]: DEBUG oslo_concurrency.lockutils [req-b17f0189-f469-4847-929a-b5503930f6fa req-6ce39cee-116a-472a-bb9c-24cba98972fb service nova] Acquired lock "refresh_cache-afaca059-fc4e-4de2-8fa0-d2f226ea7051" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 884.167273] env[62066]: DEBUG nova.network.neutron [req-b17f0189-f469-4847-929a-b5503930f6fa req-6ce39cee-116a-472a-bb9c-24cba98972fb service nova] [instance: afaca059-fc4e-4de2-8fa0-d2f226ea7051] Refreshing network info cache for port 725b6138-b39f-41c3-bf84-1a13a3e30fa9 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 884.173970] env[62066]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-899964d3-78f0-461e-b369-6ca7e72c30f2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.224928] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7b75bfaa-cfbe-4d54-8646-a77cb9e92e69 tempest-ListImageFiltersTestJSON-425425723 tempest-ListImageFiltersTestJSON-425425723-project-member] Lock "1a4b0637-1a56-41ef-b89b-6b56d24ed206" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.452s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 884.226230] env[62066]: INFO nova.compute.manager [-] [instance: 170e184a-f009-42cf-82dc-eb9696a00a24] Took 2.19 seconds to deallocate network for instance. [ 884.236471] env[62066]: DEBUG oslo_vmware.api [None req-64742fb3-cc48-46bc-b10b-6ecbf4d3709b tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Task: {'id': task-1341050, 'name': PowerOffVM_Task, 'duration_secs': 0.201234} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.236967] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-64742fb3-cc48-46bc-b10b-6ecbf4d3709b tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] [instance: d77e3405-d158-4a8c-9f54-d5fbd26fed48] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 884.237173] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-64742fb3-cc48-46bc-b10b-6ecbf4d3709b tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] [instance: d77e3405-d158-4a8c-9f54-d5fbd26fed48] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 884.237432] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e9a7de8b-6328-4507-9bdd-e18a023b766d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.262485] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-64742fb3-cc48-46bc-b10b-6ecbf4d3709b tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] [instance: d77e3405-d158-4a8c-9f54-d5fbd26fed48] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 884.264020] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-64742fb3-cc48-46bc-b10b-6ecbf4d3709b tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] [instance: d77e3405-d158-4a8c-9f54-d5fbd26fed48] Deleting contents of the VM from datastore datastore1 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 884.264020] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-64742fb3-cc48-46bc-b10b-6ecbf4d3709b tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Deleting the datastore file [datastore1] d77e3405-d158-4a8c-9f54-d5fbd26fed48 {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 884.264020] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-32747103-1d4f-47a4-8f14-f492a19bedcf {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.275804] env[62066]: DEBUG oslo_vmware.api [None req-64742fb3-cc48-46bc-b10b-6ecbf4d3709b tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Waiting for the task: (returnval){ [ 884.275804] env[62066]: value = "task-1341053" [ 884.275804] env[62066]: _type = "Task" [ 884.275804] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.285157] env[62066]: DEBUG oslo_vmware.api [None req-64742fb3-cc48-46bc-b10b-6ecbf4d3709b tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Task: {'id': task-1341053, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.400023] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4a3c2cb7-f1e2-437a-b493-e5de08e55742 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Releasing lock "refresh_cache-8ff6fbb9-c90f-498d-9a85-d220a8c2f794" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 884.433094] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ffb6f90-74e6-413e-a095-ecf32d2fe0cf {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.444802] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f8b0650-25ee-464e-bd41-47af4f28b041 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.481313] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05bf4a8a-9443-49d1-9434-677d3fda0bc0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.500257] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-faff5d62-dbab-4cd9-a3e6-3eea7695a395 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.516284] env[62066]: DEBUG nova.compute.provider_tree [None req-09aaefa9-33e0-4c34-9b0e-8039240cf379 tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 884.529175] env[62066]: DEBUG oslo_vmware.api [None req-98bb85f9-df22-41ab-a6e4-7c41686713f5 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Task: {'id': task-1341051, 'name': ReconfigVM_Task, 'duration_secs': 0.402697} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.532531] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-98bb85f9-df22-41ab-a6e4-7c41686713f5 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] [instance: de2e2be7-efdb-45a8-842a-640ab9deb1d9] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-286087', 'volume_id': '36cdc9f5-eb4f-44b9-9aac-a7cd3fc5d570', 'name': 'volume-36cdc9f5-eb4f-44b9-9aac-a7cd3fc5d570', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'de2e2be7-efdb-45a8-842a-640ab9deb1d9', 'attached_at': '', 'detached_at': '', 'volume_id': '36cdc9f5-eb4f-44b9-9aac-a7cd3fc5d570', 'serial': '36cdc9f5-eb4f-44b9-9aac-a7cd3fc5d570'} {{(pid=62066) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 884.560943] env[62066]: DEBUG oslo_vmware.api [None req-e9fb52ef-3f9c-460e-85ba-59eb989b863f tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Task: {'id': task-1341049, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.602182} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.560943] env[62066]: INFO nova.virt.vmwareapi.ds_util [None req-e9fb52ef-3f9c-460e-85ba-59eb989b863f tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Copied virtual disk from [datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk to [datastore1] 2634fcc0-96bd-4513-8f92-515f83fe23d7/50ff584c-3b50-4395-af07-3e66769bc9f7-rescue.vmdk. [ 884.561248] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a456fa4c-8b05-4d9f-be51-d7d7df69928f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.588769] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-e9fb52ef-3f9c-460e-85ba-59eb989b863f tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] [instance: 2634fcc0-96bd-4513-8f92-515f83fe23d7] Reconfiguring VM instance instance-00000051 to attach disk [datastore1] 2634fcc0-96bd-4513-8f92-515f83fe23d7/50ff584c-3b50-4395-af07-3e66769bc9f7-rescue.vmdk or device None with type thin {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 884.593573] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a91827e5-5cf8-40eb-9a4e-a157e3ffef11 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.616623] env[62066]: DEBUG oslo_vmware.api [None req-e9fb52ef-3f9c-460e-85ba-59eb989b863f tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Waiting for the task: (returnval){ [ 884.616623] env[62066]: value = "task-1341054" [ 884.616623] env[62066]: _type = "Task" [ 884.616623] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.630845] env[62066]: DEBUG oslo_vmware.api [None req-e9fb52ef-3f9c-460e-85ba-59eb989b863f tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Task: {'id': task-1341054, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.738279] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ebe8ad74-3036-4fe0-ab5d-5d9cbfc986ca tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 884.787843] env[62066]: DEBUG oslo_vmware.api [None req-64742fb3-cc48-46bc-b10b-6ecbf4d3709b tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Task: {'id': task-1341053, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.121116} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.788124] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-64742fb3-cc48-46bc-b10b-6ecbf4d3709b tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 884.788318] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-64742fb3-cc48-46bc-b10b-6ecbf4d3709b tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] [instance: d77e3405-d158-4a8c-9f54-d5fbd26fed48] Deleted contents of the VM from datastore datastore1 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 884.788516] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-64742fb3-cc48-46bc-b10b-6ecbf4d3709b tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] [instance: d77e3405-d158-4a8c-9f54-d5fbd26fed48] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 884.788699] env[62066]: INFO nova.compute.manager [None req-64742fb3-cc48-46bc-b10b-6ecbf4d3709b tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] [instance: d77e3405-d158-4a8c-9f54-d5fbd26fed48] Took 1.09 seconds to destroy the instance on the hypervisor. [ 884.788949] env[62066]: DEBUG oslo.service.loopingcall [None req-64742fb3-cc48-46bc-b10b-6ecbf4d3709b tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 884.789174] env[62066]: DEBUG nova.compute.manager [-] [instance: d77e3405-d158-4a8c-9f54-d5fbd26fed48] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 884.789317] env[62066]: DEBUG nova.network.neutron [-] [instance: d77e3405-d158-4a8c-9f54-d5fbd26fed48] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 884.832390] env[62066]: DEBUG nova.network.neutron [-] [instance: d77e3405-d158-4a8c-9f54-d5fbd26fed48] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 884.994490] env[62066]: DEBUG oslo_concurrency.lockutils [None req-61031532-cb43-40c8-81a2-438112a5a962 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Acquiring lock "76306c3f-0674-40fe-9864-d82d11e4bb5e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 884.994854] env[62066]: DEBUG oslo_concurrency.lockutils [None req-61031532-cb43-40c8-81a2-438112a5a962 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Lock "76306c3f-0674-40fe-9864-d82d11e4bb5e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 884.995126] env[62066]: DEBUG oslo_concurrency.lockutils [None req-61031532-cb43-40c8-81a2-438112a5a962 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Acquiring lock "76306c3f-0674-40fe-9864-d82d11e4bb5e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 884.995362] env[62066]: DEBUG oslo_concurrency.lockutils [None req-61031532-cb43-40c8-81a2-438112a5a962 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Lock "76306c3f-0674-40fe-9864-d82d11e4bb5e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 884.995576] env[62066]: DEBUG oslo_concurrency.lockutils [None req-61031532-cb43-40c8-81a2-438112a5a962 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Lock "76306c3f-0674-40fe-9864-d82d11e4bb5e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 884.998312] env[62066]: INFO nova.compute.manager [None req-61031532-cb43-40c8-81a2-438112a5a962 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 76306c3f-0674-40fe-9864-d82d11e4bb5e] Terminating instance [ 885.002447] env[62066]: DEBUG nova.compute.manager [None req-61031532-cb43-40c8-81a2-438112a5a962 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 76306c3f-0674-40fe-9864-d82d11e4bb5e] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 885.002741] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-61031532-cb43-40c8-81a2-438112a5a962 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 76306c3f-0674-40fe-9864-d82d11e4bb5e] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 885.004087] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55f4a2ad-338c-4cd5-a15c-8337eca4c71e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.009231] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-4a3c2cb7-f1e2-437a-b493-e5de08e55742 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] [instance: 8ff6fbb9-c90f-498d-9a85-d220a8c2f794] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 885.012975] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5005775d-5cfe-4264-9974-4a5ecb3d861c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.014382] env[62066]: DEBUG oslo_vmware.rw_handles [None req-f0861cc1-6af6-4429-bc89-cb4b7f228490 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Completed reading data from the image iterator. {{(pid=62066) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 885.014564] env[62066]: DEBUG oslo_vmware.rw_handles [None req-f0861cc1-6af6-4429-bc89-cb4b7f228490 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/523770ef-9b20-2a18-c98f-52042f3a2529/disk-0.vmdk. {{(pid=62066) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 885.016624] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9dd5d217-665c-41a6-b41e-b6a42661c460 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.023454] env[62066]: DEBUG nova.scheduler.client.report [None req-09aaefa9-33e0-4c34-9b0e-8039240cf379 tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 885.027602] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-61031532-cb43-40c8-81a2-438112a5a962 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 76306c3f-0674-40fe-9864-d82d11e4bb5e] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 885.030756] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-97f6c8eb-94c0-485d-95de-594726b7a208 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.032595] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-4a3c2cb7-f1e2-437a-b493-e5de08e55742 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] [instance: 8ff6fbb9-c90f-498d-9a85-d220a8c2f794] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 885.034267] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-983cb3df-9742-423a-be2c-a6ab915197d2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.035921] env[62066]: DEBUG oslo_vmware.rw_handles [None req-f0861cc1-6af6-4429-bc89-cb4b7f228490 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/523770ef-9b20-2a18-c98f-52042f3a2529/disk-0.vmdk is in state: ready. {{(pid=62066) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 885.036212] env[62066]: DEBUG oslo_vmware.rw_handles [None req-f0861cc1-6af6-4429-bc89-cb4b7f228490 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Releasing lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/523770ef-9b20-2a18-c98f-52042f3a2529/disk-0.vmdk. {{(pid=62066) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 885.038542] env[62066]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-0b6fe997-08c1-40f3-a3e4-f8b12c3781bb {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.041730] env[62066]: DEBUG oslo_vmware.api [None req-61031532-cb43-40c8-81a2-438112a5a962 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Waiting for the task: (returnval){ [ 885.041730] env[62066]: value = "task-1341055" [ 885.041730] env[62066]: _type = "Task" [ 885.041730] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.052483] env[62066]: DEBUG oslo_vmware.api [None req-61031532-cb43-40c8-81a2-438112a5a962 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1341055, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.112112] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-4a3c2cb7-f1e2-437a-b493-e5de08e55742 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] [instance: 8ff6fbb9-c90f-498d-9a85-d220a8c2f794] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 885.112112] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-4a3c2cb7-f1e2-437a-b493-e5de08e55742 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] [instance: 8ff6fbb9-c90f-498d-9a85-d220a8c2f794] Deleting contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 885.112112] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-4a3c2cb7-f1e2-437a-b493-e5de08e55742 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Deleting the datastore file [datastore2] 8ff6fbb9-c90f-498d-9a85-d220a8c2f794 {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 885.112112] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-cdaa3471-950c-457e-adfe-c72bdc1e6710 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.126594] env[62066]: DEBUG nova.network.neutron [req-b17f0189-f469-4847-929a-b5503930f6fa req-6ce39cee-116a-472a-bb9c-24cba98972fb service nova] [instance: afaca059-fc4e-4de2-8fa0-d2f226ea7051] Updated VIF entry in instance network info cache for port 725b6138-b39f-41c3-bf84-1a13a3e30fa9. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 885.127013] env[62066]: DEBUG nova.network.neutron [req-b17f0189-f469-4847-929a-b5503930f6fa req-6ce39cee-116a-472a-bb9c-24cba98972fb service nova] [instance: afaca059-fc4e-4de2-8fa0-d2f226ea7051] Updating instance_info_cache with network_info: [{"id": "725b6138-b39f-41c3-bf84-1a13a3e30fa9", "address": "fa:16:3e:a3:87:a3", "network": {"id": "2d8ba809-853f-4d1b-b4b5-6d255c8a53f3", "bridge": "br-int", "label": "tempest-ServersTestFqdnHostnames-762924253-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.227", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "511b5cd53c7745eb8c3e9589b1de01c6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5fdd0624-2edb-4733-8284-225815c07f73", "external-id": "nsx-vlan-transportzone-330", "segmentation_id": 330, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap725b6138-b3", "ovs_interfaceid": "725b6138-b39f-41c3-bf84-1a13a3e30fa9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 885.131040] env[62066]: DEBUG oslo_vmware.api [None req-4a3c2cb7-f1e2-437a-b493-e5de08e55742 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Waiting for the task: (returnval){ [ 885.131040] env[62066]: value = "task-1341057" [ 885.131040] env[62066]: _type = "Task" [ 885.131040] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.136327] env[62066]: DEBUG oslo_vmware.api [None req-e9fb52ef-3f9c-460e-85ba-59eb989b863f tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Task: {'id': task-1341054, 'name': ReconfigVM_Task, 'duration_secs': 0.402389} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.136731] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-e9fb52ef-3f9c-460e-85ba-59eb989b863f tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] [instance: 2634fcc0-96bd-4513-8f92-515f83fe23d7] Reconfigured VM instance instance-00000051 to attach disk [datastore1] 2634fcc0-96bd-4513-8f92-515f83fe23d7/50ff584c-3b50-4395-af07-3e66769bc9f7-rescue.vmdk or device None with type thin {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 885.137923] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2f2e275-f8ff-4128-ad45-866581b02b75 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.143413] env[62066]: DEBUG oslo_vmware.api [None req-4a3c2cb7-f1e2-437a-b493-e5de08e55742 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Task: {'id': task-1341057, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.170669] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7b1f3ab8-298d-48f2-b024-d20dde6f31a7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.187131] env[62066]: DEBUG oslo_vmware.api [None req-e9fb52ef-3f9c-460e-85ba-59eb989b863f tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Waiting for the task: (returnval){ [ 885.187131] env[62066]: value = "task-1341058" [ 885.187131] env[62066]: _type = "Task" [ 885.187131] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.196129] env[62066]: DEBUG oslo_vmware.api [None req-e9fb52ef-3f9c-460e-85ba-59eb989b863f tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Task: {'id': task-1341058, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.256452] env[62066]: DEBUG oslo_vmware.rw_handles [None req-f0861cc1-6af6-4429-bc89-cb4b7f228490 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Closed VMDK write handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/523770ef-9b20-2a18-c98f-52042f3a2529/disk-0.vmdk. {{(pid=62066) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 885.256452] env[62066]: INFO nova.virt.vmwareapi.images [None req-f0861cc1-6af6-4429-bc89-cb4b7f228490 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 8807ce4d-532f-469c-a302-464c61c7efeb] Downloaded image file data 1147d77f-02f3-4d4b-a853-4a0e7d8b0a96 [ 885.256452] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a203c69f-0b90-4d57-b55b-b0d924447aac {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.272917] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2c8429d0-d1b5-4b6d-8361-08cb34d88850 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.318269] env[62066]: INFO nova.virt.vmwareapi.images [None req-f0861cc1-6af6-4429-bc89-cb4b7f228490 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 8807ce4d-532f-469c-a302-464c61c7efeb] The imported VM was unregistered [ 885.322016] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-f0861cc1-6af6-4429-bc89-cb4b7f228490 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 8807ce4d-532f-469c-a302-464c61c7efeb] Caching image {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 885.322016] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-f0861cc1-6af6-4429-bc89-cb4b7f228490 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Creating directory with path [datastore2] devstack-image-cache_base/1147d77f-02f3-4d4b-a853-4a0e7d8b0a96 {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 885.322016] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9e55e9a7-2e3a-4e20-bf0d-39019d74c1bf {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.334665] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-f0861cc1-6af6-4429-bc89-cb4b7f228490 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Created directory with path [datastore2] devstack-image-cache_base/1147d77f-02f3-4d4b-a853-4a0e7d8b0a96 {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 885.334877] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-f0861cc1-6af6-4429-bc89-cb4b7f228490 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_1e3b548c-3176-483b-8156-8e6eb27c2d80/OSTACK_IMG_1e3b548c-3176-483b-8156-8e6eb27c2d80.vmdk to [datastore2] devstack-image-cache_base/1147d77f-02f3-4d4b-a853-4a0e7d8b0a96/1147d77f-02f3-4d4b-a853-4a0e7d8b0a96.vmdk. {{(pid=62066) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 885.335162] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-a15ea34d-5d17-400a-8dc2-c3232919ac42 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.337398] env[62066]: DEBUG nova.network.neutron [-] [instance: d77e3405-d158-4a8c-9f54-d5fbd26fed48] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 885.342759] env[62066]: DEBUG oslo_vmware.api [None req-f0861cc1-6af6-4429-bc89-cb4b7f228490 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Waiting for the task: (returnval){ [ 885.342759] env[62066]: value = "task-1341060" [ 885.342759] env[62066]: _type = "Task" [ 885.342759] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.352959] env[62066]: DEBUG oslo_vmware.api [None req-f0861cc1-6af6-4429-bc89-cb4b7f228490 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': task-1341060, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.528987] env[62066]: DEBUG oslo_concurrency.lockutils [None req-09aaefa9-33e0-4c34-9b0e-8039240cf379 tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.845s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 885.532230] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c06200c8-d207-47fe-8d97-8cb6c611ec30 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.796s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 885.533388] env[62066]: INFO nova.compute.claims [None req-c06200c8-d207-47fe-8d97-8cb6c611ec30 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: b718b09b-a7fa-445b-8be3-5b3ebca210a2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 885.548162] env[62066]: INFO nova.scheduler.client.report [None req-09aaefa9-33e0-4c34-9b0e-8039240cf379 tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] Deleted allocations for instance 0bc31a8b-a5fc-4f6e-a4c0-024bf4260755 [ 885.555355] env[62066]: DEBUG oslo_vmware.api [None req-61031532-cb43-40c8-81a2-438112a5a962 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1341055, 'name': PowerOffVM_Task, 'duration_secs': 0.252088} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.555630] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-61031532-cb43-40c8-81a2-438112a5a962 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 76306c3f-0674-40fe-9864-d82d11e4bb5e] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 885.555808] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-61031532-cb43-40c8-81a2-438112a5a962 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 76306c3f-0674-40fe-9864-d82d11e4bb5e] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 885.556066] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d8b30de4-ecd9-4e92-ade6-71f21b338b45 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.591980] env[62066]: DEBUG nova.objects.instance [None req-98bb85f9-df22-41ab-a6e4-7c41686713f5 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Lazy-loading 'flavor' on Instance uuid de2e2be7-efdb-45a8-842a-640ab9deb1d9 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 885.631613] env[62066]: DEBUG oslo_concurrency.lockutils [req-b17f0189-f469-4847-929a-b5503930f6fa req-6ce39cee-116a-472a-bb9c-24cba98972fb service nova] Releasing lock "refresh_cache-afaca059-fc4e-4de2-8fa0-d2f226ea7051" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 885.641699] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-61031532-cb43-40c8-81a2-438112a5a962 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 76306c3f-0674-40fe-9864-d82d11e4bb5e] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 885.641935] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-61031532-cb43-40c8-81a2-438112a5a962 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 76306c3f-0674-40fe-9864-d82d11e4bb5e] Deleting contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 885.643303] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-61031532-cb43-40c8-81a2-438112a5a962 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Deleting the datastore file [datastore2] 76306c3f-0674-40fe-9864-d82d11e4bb5e {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 885.643303] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-bb80725d-153c-4e64-9729-e8d3d69ee4e1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.648916] env[62066]: DEBUG oslo_vmware.api [None req-4a3c2cb7-f1e2-437a-b493-e5de08e55742 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Task: {'id': task-1341057, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.213266} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.649534] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-4a3c2cb7-f1e2-437a-b493-e5de08e55742 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 885.649733] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-4a3c2cb7-f1e2-437a-b493-e5de08e55742 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] [instance: 8ff6fbb9-c90f-498d-9a85-d220a8c2f794] Deleted contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 885.649944] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-4a3c2cb7-f1e2-437a-b493-e5de08e55742 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] [instance: 8ff6fbb9-c90f-498d-9a85-d220a8c2f794] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 885.653911] env[62066]: DEBUG oslo_vmware.api [None req-61031532-cb43-40c8-81a2-438112a5a962 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Waiting for the task: (returnval){ [ 885.653911] env[62066]: value = "task-1341062" [ 885.653911] env[62066]: _type = "Task" [ 885.653911] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.663367] env[62066]: DEBUG oslo_vmware.api [None req-61031532-cb43-40c8-81a2-438112a5a962 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1341062, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.677592] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cbf7a2c2-7fb7-405b-bf5c-0776ba26eb99 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Acquiring lock "9c989152-ec69-478e-a0b0-62983852576c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 885.677592] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cbf7a2c2-7fb7-405b-bf5c-0776ba26eb99 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Lock "9c989152-ec69-478e-a0b0-62983852576c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 885.679980] env[62066]: INFO nova.scheduler.client.report [None req-4a3c2cb7-f1e2-437a-b493-e5de08e55742 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Deleted allocations for instance 8ff6fbb9-c90f-498d-9a85-d220a8c2f794 [ 885.696889] env[62066]: DEBUG oslo_vmware.api [None req-e9fb52ef-3f9c-460e-85ba-59eb989b863f tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Task: {'id': task-1341058, 'name': ReconfigVM_Task, 'duration_secs': 0.199428} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.697709] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-e9fb52ef-3f9c-460e-85ba-59eb989b863f tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] [instance: 2634fcc0-96bd-4513-8f92-515f83fe23d7] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 885.698506] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5f654973-b2ba-4a11-818f-16f9608d9fb9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.707278] env[62066]: DEBUG oslo_vmware.api [None req-e9fb52ef-3f9c-460e-85ba-59eb989b863f tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Waiting for the task: (returnval){ [ 885.707278] env[62066]: value = "task-1341063" [ 885.707278] env[62066]: _type = "Task" [ 885.707278] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.714404] env[62066]: DEBUG oslo_vmware.api [None req-e9fb52ef-3f9c-460e-85ba-59eb989b863f tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Task: {'id': task-1341063, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.840496] env[62066]: INFO nova.compute.manager [-] [instance: d77e3405-d158-4a8c-9f54-d5fbd26fed48] Took 1.05 seconds to deallocate network for instance. [ 885.860505] env[62066]: DEBUG oslo_vmware.api [None req-f0861cc1-6af6-4429-bc89-cb4b7f228490 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': task-1341060, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.056490] env[62066]: DEBUG oslo_concurrency.lockutils [None req-09aaefa9-33e0-4c34-9b0e-8039240cf379 tempest-ServerRescueTestJSONUnderV235-1277469920 tempest-ServerRescueTestJSONUnderV235-1277469920-project-member] Lock "0bc31a8b-a5fc-4f6e-a4c0-024bf4260755" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.105s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 886.096697] env[62066]: DEBUG oslo_concurrency.lockutils [None req-98bb85f9-df22-41ab-a6e4-7c41686713f5 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Lock "de2e2be7-efdb-45a8-842a-640ab9deb1d9" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 8.371s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 886.173361] env[62066]: DEBUG oslo_vmware.api [None req-61031532-cb43-40c8-81a2-438112a5a962 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1341062, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.146791} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.173361] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-61031532-cb43-40c8-81a2-438112a5a962 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 886.173361] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-61031532-cb43-40c8-81a2-438112a5a962 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 76306c3f-0674-40fe-9864-d82d11e4bb5e] Deleted contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 886.173361] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-61031532-cb43-40c8-81a2-438112a5a962 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 76306c3f-0674-40fe-9864-d82d11e4bb5e] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 886.173361] env[62066]: INFO nova.compute.manager [None req-61031532-cb43-40c8-81a2-438112a5a962 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 76306c3f-0674-40fe-9864-d82d11e4bb5e] Took 1.17 seconds to destroy the instance on the hypervisor. [ 886.173361] env[62066]: DEBUG oslo.service.loopingcall [None req-61031532-cb43-40c8-81a2-438112a5a962 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 886.173361] env[62066]: DEBUG nova.compute.manager [-] [instance: 76306c3f-0674-40fe-9864-d82d11e4bb5e] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 886.173361] env[62066]: DEBUG nova.network.neutron [-] [instance: 76306c3f-0674-40fe-9864-d82d11e4bb5e] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 886.185159] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4a3c2cb7-f1e2-437a-b493-e5de08e55742 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 886.185546] env[62066]: DEBUG nova.compute.manager [None req-cbf7a2c2-7fb7-405b-bf5c-0776ba26eb99 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 9c989152-ec69-478e-a0b0-62983852576c] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 886.219277] env[62066]: DEBUG oslo_vmware.api [None req-e9fb52ef-3f9c-460e-85ba-59eb989b863f tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Task: {'id': task-1341063, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.356046] env[62066]: DEBUG oslo_concurrency.lockutils [None req-64742fb3-cc48-46bc-b10b-6ecbf4d3709b tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 886.361808] env[62066]: DEBUG oslo_vmware.api [None req-f0861cc1-6af6-4429-bc89-cb4b7f228490 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': task-1341060, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.507252] env[62066]: DEBUG oslo_concurrency.lockutils [None req-6013cc06-a721-40d2-9804-908f29b48f28 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Acquiring lock "de2e2be7-efdb-45a8-842a-640ab9deb1d9" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 886.507535] env[62066]: DEBUG oslo_concurrency.lockutils [None req-6013cc06-a721-40d2-9804-908f29b48f28 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Lock "de2e2be7-efdb-45a8-842a-640ab9deb1d9" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 886.562436] env[62066]: DEBUG nova.compute.manager [req-a4ac481f-d177-4c32-83c5-0d8ce4e43a01 req-e5cee386-c8e5-4146-83a5-2518eeedcab1 service nova] [instance: 8ff6fbb9-c90f-498d-9a85-d220a8c2f794] Received event network-vif-unplugged-8c324833-7bf9-4007-85aa-5b16871f63c7 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 886.562594] env[62066]: DEBUG oslo_concurrency.lockutils [req-a4ac481f-d177-4c32-83c5-0d8ce4e43a01 req-e5cee386-c8e5-4146-83a5-2518eeedcab1 service nova] Acquiring lock "8ff6fbb9-c90f-498d-9a85-d220a8c2f794-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 886.564091] env[62066]: DEBUG oslo_concurrency.lockutils [req-a4ac481f-d177-4c32-83c5-0d8ce4e43a01 req-e5cee386-c8e5-4146-83a5-2518eeedcab1 service nova] Lock "8ff6fbb9-c90f-498d-9a85-d220a8c2f794-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 886.564466] env[62066]: DEBUG oslo_concurrency.lockutils [req-a4ac481f-d177-4c32-83c5-0d8ce4e43a01 req-e5cee386-c8e5-4146-83a5-2518eeedcab1 service nova] Lock "8ff6fbb9-c90f-498d-9a85-d220a8c2f794-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 886.566119] env[62066]: DEBUG nova.compute.manager [req-a4ac481f-d177-4c32-83c5-0d8ce4e43a01 req-e5cee386-c8e5-4146-83a5-2518eeedcab1 service nova] [instance: 8ff6fbb9-c90f-498d-9a85-d220a8c2f794] No waiting events found dispatching network-vif-unplugged-8c324833-7bf9-4007-85aa-5b16871f63c7 {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 886.566119] env[62066]: WARNING nova.compute.manager [req-a4ac481f-d177-4c32-83c5-0d8ce4e43a01 req-e5cee386-c8e5-4146-83a5-2518eeedcab1 service nova] [instance: 8ff6fbb9-c90f-498d-9a85-d220a8c2f794] Received unexpected event network-vif-unplugged-8c324833-7bf9-4007-85aa-5b16871f63c7 for instance with vm_state shelved_offloaded and task_state None. [ 886.569085] env[62066]: DEBUG nova.compute.manager [req-a4ac481f-d177-4c32-83c5-0d8ce4e43a01 req-e5cee386-c8e5-4146-83a5-2518eeedcab1 service nova] [instance: 8ff6fbb9-c90f-498d-9a85-d220a8c2f794] Received event network-changed-8c324833-7bf9-4007-85aa-5b16871f63c7 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 886.569085] env[62066]: DEBUG nova.compute.manager [req-a4ac481f-d177-4c32-83c5-0d8ce4e43a01 req-e5cee386-c8e5-4146-83a5-2518eeedcab1 service nova] [instance: 8ff6fbb9-c90f-498d-9a85-d220a8c2f794] Refreshing instance network info cache due to event network-changed-8c324833-7bf9-4007-85aa-5b16871f63c7. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 886.569085] env[62066]: DEBUG oslo_concurrency.lockutils [req-a4ac481f-d177-4c32-83c5-0d8ce4e43a01 req-e5cee386-c8e5-4146-83a5-2518eeedcab1 service nova] Acquiring lock "refresh_cache-8ff6fbb9-c90f-498d-9a85-d220a8c2f794" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 886.569085] env[62066]: DEBUG oslo_concurrency.lockutils [req-a4ac481f-d177-4c32-83c5-0d8ce4e43a01 req-e5cee386-c8e5-4146-83a5-2518eeedcab1 service nova] Acquired lock "refresh_cache-8ff6fbb9-c90f-498d-9a85-d220a8c2f794" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 886.569085] env[62066]: DEBUG nova.network.neutron [req-a4ac481f-d177-4c32-83c5-0d8ce4e43a01 req-e5cee386-c8e5-4146-83a5-2518eeedcab1 service nova] [instance: 8ff6fbb9-c90f-498d-9a85-d220a8c2f794] Refreshing network info cache for port 8c324833-7bf9-4007-85aa-5b16871f63c7 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 886.720547] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cbf7a2c2-7fb7-405b-bf5c-0776ba26eb99 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 886.730382] env[62066]: DEBUG oslo_vmware.api [None req-e9fb52ef-3f9c-460e-85ba-59eb989b863f tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Task: {'id': task-1341063, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.816182] env[62066]: DEBUG nova.compute.manager [req-efd3f73d-6ae0-4e4b-8675-275636e753a6 req-85ad4a45-895e-4ff0-9a9d-acb9f044fd96 service nova] [instance: 76306c3f-0674-40fe-9864-d82d11e4bb5e] Received event network-vif-deleted-28f3b624-430a-474d-a24b-9e54db92e9e2 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 886.816649] env[62066]: INFO nova.compute.manager [req-efd3f73d-6ae0-4e4b-8675-275636e753a6 req-85ad4a45-895e-4ff0-9a9d-acb9f044fd96 service nova] [instance: 76306c3f-0674-40fe-9864-d82d11e4bb5e] Neutron deleted interface 28f3b624-430a-474d-a24b-9e54db92e9e2; detaching it from the instance and deleting it from the info cache [ 886.816649] env[62066]: DEBUG nova.network.neutron [req-efd3f73d-6ae0-4e4b-8675-275636e753a6 req-85ad4a45-895e-4ff0-9a9d-acb9f044fd96 service nova] [instance: 76306c3f-0674-40fe-9864-d82d11e4bb5e] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 886.823591] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1e6cafb-bb39-4b9a-b055-24e4741deb1f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.836752] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76ed77e0-983d-4c38-b2ae-9488062a6c52 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.885629] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cf0f315-d2e7-4076-b9f1-28b48d85ac07 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.901512] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1cbf3999-2f2e-46de-b676-183ecbb4718a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.907089] env[62066]: DEBUG oslo_vmware.api [None req-f0861cc1-6af6-4429-bc89-cb4b7f228490 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': task-1341060, 'name': MoveVirtualDisk_Task} progress is 21%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.922191] env[62066]: DEBUG nova.compute.provider_tree [None req-c06200c8-d207-47fe-8d97-8cb6c611ec30 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 887.012905] env[62066]: INFO nova.compute.manager [None req-6013cc06-a721-40d2-9804-908f29b48f28 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] [instance: de2e2be7-efdb-45a8-842a-640ab9deb1d9] Detaching volume 36cdc9f5-eb4f-44b9-9aac-a7cd3fc5d570 [ 887.060249] env[62066]: INFO nova.virt.block_device [None req-6013cc06-a721-40d2-9804-908f29b48f28 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] [instance: de2e2be7-efdb-45a8-842a-640ab9deb1d9] Attempting to driver detach volume 36cdc9f5-eb4f-44b9-9aac-a7cd3fc5d570 from mountpoint /dev/sdb [ 887.060563] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-6013cc06-a721-40d2-9804-908f29b48f28 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] [instance: de2e2be7-efdb-45a8-842a-640ab9deb1d9] Volume detach. Driver type: vmdk {{(pid=62066) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 887.060748] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-6013cc06-a721-40d2-9804-908f29b48f28 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] [instance: de2e2be7-efdb-45a8-842a-640ab9deb1d9] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-286087', 'volume_id': '36cdc9f5-eb4f-44b9-9aac-a7cd3fc5d570', 'name': 'volume-36cdc9f5-eb4f-44b9-9aac-a7cd3fc5d570', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'de2e2be7-efdb-45a8-842a-640ab9deb1d9', 'attached_at': '', 'detached_at': '', 'volume_id': '36cdc9f5-eb4f-44b9-9aac-a7cd3fc5d570', 'serial': '36cdc9f5-eb4f-44b9-9aac-a7cd3fc5d570'} {{(pid=62066) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 887.061701] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc506592-757f-48dc-a20c-c88f002ce6fa {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.095669] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97d78fcc-3206-4d21-9d6d-9d74670633b9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.108534] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1017e9f3-0a13-43b5-a43b-38b661422890 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.139285] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-101f74b9-85a5-4f67-93c8-551979346ea4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.161807] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-6013cc06-a721-40d2-9804-908f29b48f28 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] The volume has not been displaced from its original location: [datastore2] volume-36cdc9f5-eb4f-44b9-9aac-a7cd3fc5d570/volume-36cdc9f5-eb4f-44b9-9aac-a7cd3fc5d570.vmdk. No consolidation needed. {{(pid=62066) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 887.167582] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-6013cc06-a721-40d2-9804-908f29b48f28 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] [instance: de2e2be7-efdb-45a8-842a-640ab9deb1d9] Reconfiguring VM instance instance-00000048 to detach disk 2001 {{(pid=62066) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 887.168075] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cd2bf843-96c8-4215-b394-150c95e1e577 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.193298] env[62066]: DEBUG oslo_vmware.api [None req-6013cc06-a721-40d2-9804-908f29b48f28 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Waiting for the task: (returnval){ [ 887.193298] env[62066]: value = "task-1341064" [ 887.193298] env[62066]: _type = "Task" [ 887.193298] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.206187] env[62066]: DEBUG oslo_vmware.api [None req-6013cc06-a721-40d2-9804-908f29b48f28 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Task: {'id': task-1341064, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.219370] env[62066]: DEBUG oslo_vmware.api [None req-e9fb52ef-3f9c-460e-85ba-59eb989b863f tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Task: {'id': task-1341063, 'name': PowerOnVM_Task, 'duration_secs': 1.134047} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.221853] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-e9fb52ef-3f9c-460e-85ba-59eb989b863f tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] [instance: 2634fcc0-96bd-4513-8f92-515f83fe23d7] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 887.226312] env[62066]: DEBUG nova.compute.manager [None req-e9fb52ef-3f9c-460e-85ba-59eb989b863f tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] [instance: 2634fcc0-96bd-4513-8f92-515f83fe23d7] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 887.227524] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27cf8e24-2b99-4557-8108-e4ff11448f4c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.269981] env[62066]: DEBUG nova.network.neutron [-] [instance: 76306c3f-0674-40fe-9864-d82d11e4bb5e] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 887.329677] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c7c7129f-0db9-47ed-8071-c7778bbf1a0e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.345893] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f491e36-b08b-4070-bf60-6fa0d9ff7024 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.377031] env[62066]: DEBUG nova.compute.manager [req-efd3f73d-6ae0-4e4b-8675-275636e753a6 req-85ad4a45-895e-4ff0-9a9d-acb9f044fd96 service nova] [instance: 76306c3f-0674-40fe-9864-d82d11e4bb5e] Detach interface failed, port_id=28f3b624-430a-474d-a24b-9e54db92e9e2, reason: Instance 76306c3f-0674-40fe-9864-d82d11e4bb5e could not be found. {{(pid=62066) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 887.398409] env[62066]: DEBUG oslo_vmware.api [None req-f0861cc1-6af6-4429-bc89-cb4b7f228490 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': task-1341060, 'name': MoveVirtualDisk_Task} progress is 40%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.425614] env[62066]: DEBUG nova.scheduler.client.report [None req-c06200c8-d207-47fe-8d97-8cb6c611ec30 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 887.664367] env[62066]: DEBUG nova.network.neutron [req-a4ac481f-d177-4c32-83c5-0d8ce4e43a01 req-e5cee386-c8e5-4146-83a5-2518eeedcab1 service nova] [instance: 8ff6fbb9-c90f-498d-9a85-d220a8c2f794] Updated VIF entry in instance network info cache for port 8c324833-7bf9-4007-85aa-5b16871f63c7. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 887.664841] env[62066]: DEBUG nova.network.neutron [req-a4ac481f-d177-4c32-83c5-0d8ce4e43a01 req-e5cee386-c8e5-4146-83a5-2518eeedcab1 service nova] [instance: 8ff6fbb9-c90f-498d-9a85-d220a8c2f794] Updating instance_info_cache with network_info: [{"id": "8c324833-7bf9-4007-85aa-5b16871f63c7", "address": "fa:16:3e:00:85:34", "network": {"id": "0e17b5e2-526e-4d4c-a463-5e34408044cf", "bridge": null, "label": "tempest-AttachVolumeShelveTestJSON-1588316661-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.184", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1f05d8a66d0c4a479e49a947e4b4bbff", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap8c324833-7b", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 887.714017] env[62066]: DEBUG oslo_vmware.api [None req-6013cc06-a721-40d2-9804-908f29b48f28 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Task: {'id': task-1341064, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.775676] env[62066]: INFO nova.compute.manager [-] [instance: 76306c3f-0674-40fe-9864-d82d11e4bb5e] Took 1.60 seconds to deallocate network for instance. [ 887.894092] env[62066]: DEBUG oslo_vmware.api [None req-f0861cc1-6af6-4429-bc89-cb4b7f228490 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': task-1341060, 'name': MoveVirtualDisk_Task} progress is 63%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.932925] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c06200c8-d207-47fe-8d97-8cb6c611ec30 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.401s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 887.933516] env[62066]: DEBUG nova.compute.manager [None req-c06200c8-d207-47fe-8d97-8cb6c611ec30 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: b718b09b-a7fa-445b-8be3-5b3ebca210a2] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 887.938066] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1dc5d31e-1ccf-4559-b370-76e0d5f04709 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 4.696s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 887.938066] env[62066]: DEBUG nova.objects.instance [None req-1dc5d31e-1ccf-4559-b370-76e0d5f04709 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 76306c3f-0674-40fe-9864-d82d11e4bb5e] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62066) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 888.168413] env[62066]: DEBUG oslo_concurrency.lockutils [req-a4ac481f-d177-4c32-83c5-0d8ce4e43a01 req-e5cee386-c8e5-4146-83a5-2518eeedcab1 service nova] Releasing lock "refresh_cache-8ff6fbb9-c90f-498d-9a85-d220a8c2f794" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 888.212617] env[62066]: DEBUG oslo_vmware.api [None req-6013cc06-a721-40d2-9804-908f29b48f28 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Task: {'id': task-1341064, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.279897] env[62066]: DEBUG oslo_concurrency.lockutils [None req-61031532-cb43-40c8-81a2-438112a5a962 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 888.399311] env[62066]: DEBUG oslo_vmware.api [None req-f0861cc1-6af6-4429-bc89-cb4b7f228490 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': task-1341060, 'name': MoveVirtualDisk_Task} progress is 83%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.443415] env[62066]: DEBUG nova.compute.utils [None req-c06200c8-d207-47fe-8d97-8cb6c611ec30 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 888.451201] env[62066]: DEBUG nova.compute.manager [None req-c06200c8-d207-47fe-8d97-8cb6c611ec30 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: b718b09b-a7fa-445b-8be3-5b3ebca210a2] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 888.451201] env[62066]: DEBUG nova.network.neutron [None req-c06200c8-d207-47fe-8d97-8cb6c611ec30 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: b718b09b-a7fa-445b-8be3-5b3ebca210a2] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 888.559369] env[62066]: DEBUG nova.policy [None req-c06200c8-d207-47fe-8d97-8cb6c611ec30 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '747026d272d4467bbf14a0c30589c3b7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0d4e33a99c7741fb8cdd97f4ec5dbbd8', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 888.710845] env[62066]: DEBUG oslo_vmware.api [None req-6013cc06-a721-40d2-9804-908f29b48f28 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Task: {'id': task-1341064, 'name': ReconfigVM_Task, 'duration_secs': 1.387389} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.711290] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-6013cc06-a721-40d2-9804-908f29b48f28 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] [instance: de2e2be7-efdb-45a8-842a-640ab9deb1d9] Reconfigured VM instance instance-00000048 to detach disk 2001 {{(pid=62066) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 888.716281] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-162b246f-6e86-4be7-acc4-c9e4eb3bedf3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.734432] env[62066]: DEBUG oslo_vmware.api [None req-6013cc06-a721-40d2-9804-908f29b48f28 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Waiting for the task: (returnval){ [ 888.734432] env[62066]: value = "task-1341065" [ 888.734432] env[62066]: _type = "Task" [ 888.734432] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.747762] env[62066]: DEBUG oslo_vmware.api [None req-6013cc06-a721-40d2-9804-908f29b48f28 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Task: {'id': task-1341065, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.790538] env[62066]: INFO nova.compute.manager [None req-7882bbc3-f566-458c-8feb-7b619f1c8a07 tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] [instance: 2634fcc0-96bd-4513-8f92-515f83fe23d7] Unrescuing [ 888.790847] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7882bbc3-f566-458c-8feb-7b619f1c8a07 tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Acquiring lock "refresh_cache-2634fcc0-96bd-4513-8f92-515f83fe23d7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 888.791092] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7882bbc3-f566-458c-8feb-7b619f1c8a07 tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Acquired lock "refresh_cache-2634fcc0-96bd-4513-8f92-515f83fe23d7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 888.791279] env[62066]: DEBUG nova.network.neutron [None req-7882bbc3-f566-458c-8feb-7b619f1c8a07 tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] [instance: 2634fcc0-96bd-4513-8f92-515f83fe23d7] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 888.903943] env[62066]: DEBUG oslo_vmware.api [None req-f0861cc1-6af6-4429-bc89-cb4b7f228490 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': task-1341060, 'name': MoveVirtualDisk_Task, 'duration_secs': 3.548724} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.905158] env[62066]: INFO nova.virt.vmwareapi.ds_util [None req-f0861cc1-6af6-4429-bc89-cb4b7f228490 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_1e3b548c-3176-483b-8156-8e6eb27c2d80/OSTACK_IMG_1e3b548c-3176-483b-8156-8e6eb27c2d80.vmdk to [datastore2] devstack-image-cache_base/1147d77f-02f3-4d4b-a853-4a0e7d8b0a96/1147d77f-02f3-4d4b-a853-4a0e7d8b0a96.vmdk. [ 888.905521] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-f0861cc1-6af6-4429-bc89-cb4b7f228490 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 8807ce4d-532f-469c-a302-464c61c7efeb] Cleaning up location [datastore2] OSTACK_IMG_1e3b548c-3176-483b-8156-8e6eb27c2d80 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 888.905854] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-f0861cc1-6af6-4429-bc89-cb4b7f228490 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_1e3b548c-3176-483b-8156-8e6eb27c2d80 {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 888.906305] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-dcd0fa30-5aec-4510-9057-19ddc1f09d1b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.914537] env[62066]: DEBUG oslo_vmware.api [None req-f0861cc1-6af6-4429-bc89-cb4b7f228490 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Waiting for the task: (returnval){ [ 888.914537] env[62066]: value = "task-1341066" [ 888.914537] env[62066]: _type = "Task" [ 888.914537] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.926879] env[62066]: DEBUG oslo_vmware.api [None req-f0861cc1-6af6-4429-bc89-cb4b7f228490 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': task-1341066, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.927777] env[62066]: DEBUG nova.network.neutron [None req-c06200c8-d207-47fe-8d97-8cb6c611ec30 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: b718b09b-a7fa-445b-8be3-5b3ebca210a2] Successfully created port: 3d01745e-0d9e-4bae-ba4d-242380edbab6 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 888.952612] env[62066]: DEBUG nova.compute.manager [None req-c06200c8-d207-47fe-8d97-8cb6c611ec30 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: b718b09b-a7fa-445b-8be3-5b3ebca210a2] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 888.956653] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1dc5d31e-1ccf-4559-b370-76e0d5f04709 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.019s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 888.958139] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ebe8ad74-3036-4fe0-ab5d-5d9cbfc986ca tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.220s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 888.958374] env[62066]: DEBUG nova.objects.instance [None req-ebe8ad74-3036-4fe0-ab5d-5d9cbfc986ca tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Lazy-loading 'resources' on Instance uuid 170e184a-f009-42cf-82dc-eb9696a00a24 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 889.249325] env[62066]: DEBUG oslo_vmware.api [None req-6013cc06-a721-40d2-9804-908f29b48f28 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Task: {'id': task-1341065, 'name': ReconfigVM_Task, 'duration_secs': 0.185956} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.249325] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-6013cc06-a721-40d2-9804-908f29b48f28 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] [instance: de2e2be7-efdb-45a8-842a-640ab9deb1d9] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-286087', 'volume_id': '36cdc9f5-eb4f-44b9-9aac-a7cd3fc5d570', 'name': 'volume-36cdc9f5-eb4f-44b9-9aac-a7cd3fc5d570', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'de2e2be7-efdb-45a8-842a-640ab9deb1d9', 'attached_at': '', 'detached_at': '', 'volume_id': '36cdc9f5-eb4f-44b9-9aac-a7cd3fc5d570', 'serial': '36cdc9f5-eb4f-44b9-9aac-a7cd3fc5d570'} {{(pid=62066) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 889.316532] env[62066]: DEBUG oslo_concurrency.lockutils [None req-81c18bf6-1c8d-45a0-a66c-dcda0761f479 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Acquiring lock "8ff6fbb9-c90f-498d-9a85-d220a8c2f794" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 889.424130] env[62066]: DEBUG oslo_vmware.api [None req-f0861cc1-6af6-4429-bc89-cb4b7f228490 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': task-1341066, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.052365} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.424411] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-f0861cc1-6af6-4429-bc89-cb4b7f228490 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 889.424584] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f0861cc1-6af6-4429-bc89-cb4b7f228490 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Releasing lock "[datastore2] devstack-image-cache_base/1147d77f-02f3-4d4b-a853-4a0e7d8b0a96/1147d77f-02f3-4d4b-a853-4a0e7d8b0a96.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 889.424832] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-f0861cc1-6af6-4429-bc89-cb4b7f228490 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/1147d77f-02f3-4d4b-a853-4a0e7d8b0a96/1147d77f-02f3-4d4b-a853-4a0e7d8b0a96.vmdk to [datastore2] 8807ce4d-532f-469c-a302-464c61c7efeb/8807ce4d-532f-469c-a302-464c61c7efeb.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 889.425120] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-aba9635c-3312-4528-85c7-b7405bf46583 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.433514] env[62066]: DEBUG oslo_vmware.api [None req-f0861cc1-6af6-4429-bc89-cb4b7f228490 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Waiting for the task: (returnval){ [ 889.433514] env[62066]: value = "task-1341067" [ 889.433514] env[62066]: _type = "Task" [ 889.433514] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.441455] env[62066]: DEBUG oslo_vmware.api [None req-f0861cc1-6af6-4429-bc89-cb4b7f228490 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': task-1341067, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.607385] env[62066]: DEBUG nova.network.neutron [None req-7882bbc3-f566-458c-8feb-7b619f1c8a07 tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] [instance: 2634fcc0-96bd-4513-8f92-515f83fe23d7] Updating instance_info_cache with network_info: [{"id": "cf42ffe4-6621-4c29-9a6d-687ade676346", "address": "fa:16:3e:4c:ae:2f", "network": {"id": "fac1b86c-746c-4b6e-8d7f-21fe6ed404a3", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1827773279-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "59855b1cf91d43e3ba0482f414c4e546", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "88651df2-0506-4f6c-b868-dd30a81f2b1c", "external-id": "nsx-vlan-transportzone-366", "segmentation_id": 366, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcf42ffe4-66", "ovs_interfaceid": "cf42ffe4-6621-4c29-9a6d-687ade676346", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 889.666255] env[62066]: DEBUG oslo_concurrency.lockutils [None req-93fb3d16-c0a4-4bf0-b776-e2995ef54e48 tempest-ServerTagsTestJSON-2015212290 tempest-ServerTagsTestJSON-2015212290-project-member] Acquiring lock "f4a84152-f222-46a2-9a35-8a0a0078fc12" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 889.666496] env[62066]: DEBUG oslo_concurrency.lockutils [None req-93fb3d16-c0a4-4bf0-b776-e2995ef54e48 tempest-ServerTagsTestJSON-2015212290 tempest-ServerTagsTestJSON-2015212290-project-member] Lock "f4a84152-f222-46a2-9a35-8a0a0078fc12" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 889.670850] env[62066]: DEBUG oslo_concurrency.lockutils [None req-adc52c21-6d1b-4740-84d8-9cdb3fbc5d28 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Acquiring lock "90c33434-c127-450b-9ff0-75181b4ac385" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 889.671162] env[62066]: DEBUG oslo_concurrency.lockutils [None req-adc52c21-6d1b-4740-84d8-9cdb3fbc5d28 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Lock "90c33434-c127-450b-9ff0-75181b4ac385" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 889.737449] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c196453d-5adf-4d59-b2df-ebc0c58f91d7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.746046] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2459e147-bd1a-4b22-a85c-bd15033e8c03 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.781527] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f92fc94-4573-46f1-b633-c168592b5da3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.790428] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbc32751-76e6-49b5-b0d4-f38f27b6ea8f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.804109] env[62066]: DEBUG nova.compute.provider_tree [None req-ebe8ad74-3036-4fe0-ab5d-5d9cbfc986ca tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 889.807092] env[62066]: DEBUG nova.objects.instance [None req-6013cc06-a721-40d2-9804-908f29b48f28 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Lazy-loading 'flavor' on Instance uuid de2e2be7-efdb-45a8-842a-640ab9deb1d9 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 889.945176] env[62066]: DEBUG oslo_vmware.api [None req-f0861cc1-6af6-4429-bc89-cb4b7f228490 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': task-1341067, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.969203] env[62066]: DEBUG nova.compute.manager [None req-c06200c8-d207-47fe-8d97-8cb6c611ec30 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: b718b09b-a7fa-445b-8be3-5b3ebca210a2] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 889.995666] env[62066]: DEBUG nova.virt.hardware [None req-c06200c8-d207-47fe-8d97-8cb6c611ec30 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-23T13:40:41Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-23T13:40:23Z,direct_url=,disk_format='vmdk',id=50ff584c-3b50-4395-af07-3e66769bc9f7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='eb52f7069a374c61ae946f052007c6d9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-23T13:40:24Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 889.996008] env[62066]: DEBUG nova.virt.hardware [None req-c06200c8-d207-47fe-8d97-8cb6c611ec30 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 889.996259] env[62066]: DEBUG nova.virt.hardware [None req-c06200c8-d207-47fe-8d97-8cb6c611ec30 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 889.996524] env[62066]: DEBUG nova.virt.hardware [None req-c06200c8-d207-47fe-8d97-8cb6c611ec30 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 889.996739] env[62066]: DEBUG nova.virt.hardware [None req-c06200c8-d207-47fe-8d97-8cb6c611ec30 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 889.996918] env[62066]: DEBUG nova.virt.hardware [None req-c06200c8-d207-47fe-8d97-8cb6c611ec30 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 889.997153] env[62066]: DEBUG nova.virt.hardware [None req-c06200c8-d207-47fe-8d97-8cb6c611ec30 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 889.997456] env[62066]: DEBUG nova.virt.hardware [None req-c06200c8-d207-47fe-8d97-8cb6c611ec30 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 889.997644] env[62066]: DEBUG nova.virt.hardware [None req-c06200c8-d207-47fe-8d97-8cb6c611ec30 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 889.997923] env[62066]: DEBUG nova.virt.hardware [None req-c06200c8-d207-47fe-8d97-8cb6c611ec30 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 889.997977] env[62066]: DEBUG nova.virt.hardware [None req-c06200c8-d207-47fe-8d97-8cb6c611ec30 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 889.998909] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f271ecf-02ad-4888-9d04-e89a53c9bd3f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.007280] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbf00503-a176-41c1-aa1e-24b93c3ba8e6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.111832] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7882bbc3-f566-458c-8feb-7b619f1c8a07 tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Releasing lock "refresh_cache-2634fcc0-96bd-4513-8f92-515f83fe23d7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 890.112349] env[62066]: DEBUG nova.objects.instance [None req-7882bbc3-f566-458c-8feb-7b619f1c8a07 tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Lazy-loading 'flavor' on Instance uuid 2634fcc0-96bd-4513-8f92-515f83fe23d7 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 890.171567] env[62066]: DEBUG nova.compute.manager [None req-93fb3d16-c0a4-4bf0-b776-e2995ef54e48 tempest-ServerTagsTestJSON-2015212290 tempest-ServerTagsTestJSON-2015212290-project-member] [instance: f4a84152-f222-46a2-9a35-8a0a0078fc12] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 890.174636] env[62066]: DEBUG nova.compute.manager [None req-adc52c21-6d1b-4740-84d8-9cdb3fbc5d28 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: 90c33434-c127-450b-9ff0-75181b4ac385] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 890.307103] env[62066]: DEBUG nova.scheduler.client.report [None req-ebe8ad74-3036-4fe0-ab5d-5d9cbfc986ca tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 890.430409] env[62066]: DEBUG nova.compute.manager [req-e50cd375-d214-4790-bf1e-8629e8d203d4 req-e06fa75c-603a-4601-aad8-2af7999a2d4c service nova] [instance: b718b09b-a7fa-445b-8be3-5b3ebca210a2] Received event network-vif-plugged-3d01745e-0d9e-4bae-ba4d-242380edbab6 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 890.430409] env[62066]: DEBUG oslo_concurrency.lockutils [req-e50cd375-d214-4790-bf1e-8629e8d203d4 req-e06fa75c-603a-4601-aad8-2af7999a2d4c service nova] Acquiring lock "b718b09b-a7fa-445b-8be3-5b3ebca210a2-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 890.430409] env[62066]: DEBUG oslo_concurrency.lockutils [req-e50cd375-d214-4790-bf1e-8629e8d203d4 req-e06fa75c-603a-4601-aad8-2af7999a2d4c service nova] Lock "b718b09b-a7fa-445b-8be3-5b3ebca210a2-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 890.430409] env[62066]: DEBUG oslo_concurrency.lockutils [req-e50cd375-d214-4790-bf1e-8629e8d203d4 req-e06fa75c-603a-4601-aad8-2af7999a2d4c service nova] Lock "b718b09b-a7fa-445b-8be3-5b3ebca210a2-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 890.430409] env[62066]: DEBUG nova.compute.manager [req-e50cd375-d214-4790-bf1e-8629e8d203d4 req-e06fa75c-603a-4601-aad8-2af7999a2d4c service nova] [instance: b718b09b-a7fa-445b-8be3-5b3ebca210a2] No waiting events found dispatching network-vif-plugged-3d01745e-0d9e-4bae-ba4d-242380edbab6 {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 890.430409] env[62066]: WARNING nova.compute.manager [req-e50cd375-d214-4790-bf1e-8629e8d203d4 req-e06fa75c-603a-4601-aad8-2af7999a2d4c service nova] [instance: b718b09b-a7fa-445b-8be3-5b3ebca210a2] Received unexpected event network-vif-plugged-3d01745e-0d9e-4bae-ba4d-242380edbab6 for instance with vm_state building and task_state spawning. [ 890.443577] env[62066]: DEBUG oslo_vmware.api [None req-f0861cc1-6af6-4429-bc89-cb4b7f228490 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': task-1341067, 'name': CopyVirtualDisk_Task} progress is 9%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.536060] env[62066]: DEBUG nova.network.neutron [None req-c06200c8-d207-47fe-8d97-8cb6c611ec30 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: b718b09b-a7fa-445b-8be3-5b3ebca210a2] Successfully updated port: 3d01745e-0d9e-4bae-ba4d-242380edbab6 {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 890.620426] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14b1645b-ded3-4100-a849-bd1d8d59554a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.648270] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-7882bbc3-f566-458c-8feb-7b619f1c8a07 tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] [instance: 2634fcc0-96bd-4513-8f92-515f83fe23d7] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 890.650633] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4683ca90-fd71-4c51-bfab-d66d78aa0a1d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.664391] env[62066]: DEBUG oslo_vmware.api [None req-7882bbc3-f566-458c-8feb-7b619f1c8a07 tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Waiting for the task: (returnval){ [ 890.664391] env[62066]: value = "task-1341068" [ 890.664391] env[62066]: _type = "Task" [ 890.664391] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.676300] env[62066]: DEBUG oslo_vmware.api [None req-7882bbc3-f566-458c-8feb-7b619f1c8a07 tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Task: {'id': task-1341068, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.700122] env[62066]: DEBUG oslo_concurrency.lockutils [None req-adc52c21-6d1b-4740-84d8-9cdb3fbc5d28 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 890.701093] env[62066]: DEBUG oslo_concurrency.lockutils [None req-93fb3d16-c0a4-4bf0-b776-e2995ef54e48 tempest-ServerTagsTestJSON-2015212290 tempest-ServerTagsTestJSON-2015212290-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 890.813840] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ebe8ad74-3036-4fe0-ab5d-5d9cbfc986ca tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.856s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 890.823509] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4a3c2cb7-f1e2-437a-b493-e5de08e55742 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.638s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 890.824176] env[62066]: DEBUG nova.objects.instance [None req-4a3c2cb7-f1e2-437a-b493-e5de08e55742 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Lazy-loading 'resources' on Instance uuid 8ff6fbb9-c90f-498d-9a85-d220a8c2f794 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 890.825251] env[62066]: DEBUG oslo_concurrency.lockutils [None req-6013cc06-a721-40d2-9804-908f29b48f28 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Lock "de2e2be7-efdb-45a8-842a-640ab9deb1d9" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 4.318s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 890.850315] env[62066]: INFO nova.scheduler.client.report [None req-ebe8ad74-3036-4fe0-ab5d-5d9cbfc986ca tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Deleted allocations for instance 170e184a-f009-42cf-82dc-eb9696a00a24 [ 890.950629] env[62066]: DEBUG oslo_vmware.api [None req-f0861cc1-6af6-4429-bc89-cb4b7f228490 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': task-1341067, 'name': CopyVirtualDisk_Task} progress is 26%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.036475] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c06200c8-d207-47fe-8d97-8cb6c611ec30 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Acquiring lock "refresh_cache-b718b09b-a7fa-445b-8be3-5b3ebca210a2" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 891.036646] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c06200c8-d207-47fe-8d97-8cb6c611ec30 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Acquired lock "refresh_cache-b718b09b-a7fa-445b-8be3-5b3ebca210a2" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 891.036803] env[62066]: DEBUG nova.network.neutron [None req-c06200c8-d207-47fe-8d97-8cb6c611ec30 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: b718b09b-a7fa-445b-8be3-5b3ebca210a2] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 891.157037] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a96d87c9-14b1-48f6-a5e4-af778c649411 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Acquiring lock "de2e2be7-efdb-45a8-842a-640ab9deb1d9" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 891.157254] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a96d87c9-14b1-48f6-a5e4-af778c649411 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Lock "de2e2be7-efdb-45a8-842a-640ab9deb1d9" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 891.157598] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a96d87c9-14b1-48f6-a5e4-af778c649411 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Acquiring lock "de2e2be7-efdb-45a8-842a-640ab9deb1d9-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 891.157859] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a96d87c9-14b1-48f6-a5e4-af778c649411 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Lock "de2e2be7-efdb-45a8-842a-640ab9deb1d9-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 891.158130] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a96d87c9-14b1-48f6-a5e4-af778c649411 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Lock "de2e2be7-efdb-45a8-842a-640ab9deb1d9-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 891.160911] env[62066]: INFO nova.compute.manager [None req-a96d87c9-14b1-48f6-a5e4-af778c649411 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] [instance: de2e2be7-efdb-45a8-842a-640ab9deb1d9] Terminating instance [ 891.163477] env[62066]: DEBUG nova.compute.manager [None req-a96d87c9-14b1-48f6-a5e4-af778c649411 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] [instance: de2e2be7-efdb-45a8-842a-640ab9deb1d9] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 891.163679] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-a96d87c9-14b1-48f6-a5e4-af778c649411 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] [instance: de2e2be7-efdb-45a8-842a-640ab9deb1d9] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 891.164781] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1818223-3e39-4e7b-aaef-a62fa9f8833f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.183501] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-a96d87c9-14b1-48f6-a5e4-af778c649411 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] [instance: de2e2be7-efdb-45a8-842a-640ab9deb1d9] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 891.183949] env[62066]: DEBUG oslo_vmware.api [None req-7882bbc3-f566-458c-8feb-7b619f1c8a07 tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Task: {'id': task-1341068, 'name': PowerOffVM_Task, 'duration_secs': 0.227907} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.184226] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-00e9a5a8-afbb-4099-b621-7c71601cb783 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.186286] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-7882bbc3-f566-458c-8feb-7b619f1c8a07 tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] [instance: 2634fcc0-96bd-4513-8f92-515f83fe23d7] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 891.193513] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-7882bbc3-f566-458c-8feb-7b619f1c8a07 tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] [instance: 2634fcc0-96bd-4513-8f92-515f83fe23d7] Reconfiguring VM instance instance-00000051 to detach disk 2001 {{(pid=62066) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 891.193843] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2210a22c-7949-4c84-a35a-e456f145cca7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.215206] env[62066]: DEBUG oslo_vmware.api [None req-a96d87c9-14b1-48f6-a5e4-af778c649411 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Waiting for the task: (returnval){ [ 891.215206] env[62066]: value = "task-1341069" [ 891.215206] env[62066]: _type = "Task" [ 891.215206] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.215513] env[62066]: DEBUG oslo_vmware.api [None req-7882bbc3-f566-458c-8feb-7b619f1c8a07 tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Waiting for the task: (returnval){ [ 891.215513] env[62066]: value = "task-1341070" [ 891.215513] env[62066]: _type = "Task" [ 891.215513] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.230152] env[62066]: DEBUG oslo_vmware.api [None req-7882bbc3-f566-458c-8feb-7b619f1c8a07 tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Task: {'id': task-1341070, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.233336] env[62066]: DEBUG oslo_vmware.api [None req-a96d87c9-14b1-48f6-a5e4-af778c649411 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Task: {'id': task-1341069, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.328112] env[62066]: DEBUG nova.objects.instance [None req-4a3c2cb7-f1e2-437a-b493-e5de08e55742 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Lazy-loading 'numa_topology' on Instance uuid 8ff6fbb9-c90f-498d-9a85-d220a8c2f794 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 891.360124] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ebe8ad74-3036-4fe0-ab5d-5d9cbfc986ca tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Lock "170e184a-f009-42cf-82dc-eb9696a00a24" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.463s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 891.449141] env[62066]: DEBUG oslo_vmware.api [None req-f0861cc1-6af6-4429-bc89-cb4b7f228490 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': task-1341067, 'name': CopyVirtualDisk_Task} progress is 49%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.572730] env[62066]: DEBUG nova.network.neutron [None req-c06200c8-d207-47fe-8d97-8cb6c611ec30 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: b718b09b-a7fa-445b-8be3-5b3ebca210a2] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 891.725453] env[62066]: DEBUG nova.network.neutron [None req-c06200c8-d207-47fe-8d97-8cb6c611ec30 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: b718b09b-a7fa-445b-8be3-5b3ebca210a2] Updating instance_info_cache with network_info: [{"id": "3d01745e-0d9e-4bae-ba4d-242380edbab6", "address": "fa:16:3e:ed:53:4c", "network": {"id": "449dfe4e-bb49-43b6-9ba7-b57af74ebfb3", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-282029676-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0d4e33a99c7741fb8cdd97f4ec5dbbd8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db68bd64-5b56-49af-a075-13dcf85cb2e0", "external-id": "nsx-vlan-transportzone-590", "segmentation_id": 590, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3d01745e-0d", "ovs_interfaceid": "3d01745e-0d9e-4bae-ba4d-242380edbab6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 891.734729] env[62066]: DEBUG oslo_vmware.api [None req-7882bbc3-f566-458c-8feb-7b619f1c8a07 tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Task: {'id': task-1341070, 'name': ReconfigVM_Task, 'duration_secs': 0.302781} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.738303] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-7882bbc3-f566-458c-8feb-7b619f1c8a07 tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] [instance: 2634fcc0-96bd-4513-8f92-515f83fe23d7] Reconfigured VM instance instance-00000051 to detach disk 2001 {{(pid=62066) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 891.738522] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-7882bbc3-f566-458c-8feb-7b619f1c8a07 tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] [instance: 2634fcc0-96bd-4513-8f92-515f83fe23d7] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 891.738812] env[62066]: DEBUG oslo_vmware.api [None req-a96d87c9-14b1-48f6-a5e4-af778c649411 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Task: {'id': task-1341069, 'name': PowerOffVM_Task, 'duration_secs': 0.210716} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.739075] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-afdc3493-aa91-485c-816d-c3c78462b364 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.740879] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-a96d87c9-14b1-48f6-a5e4-af778c649411 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] [instance: de2e2be7-efdb-45a8-842a-640ab9deb1d9] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 891.741094] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-a96d87c9-14b1-48f6-a5e4-af778c649411 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] [instance: de2e2be7-efdb-45a8-842a-640ab9deb1d9] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 891.741980] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-47a9a9f6-190b-4e3e-ba54-db0b514cc15a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.752598] env[62066]: DEBUG oslo_vmware.api [None req-7882bbc3-f566-458c-8feb-7b619f1c8a07 tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Waiting for the task: (returnval){ [ 891.752598] env[62066]: value = "task-1341071" [ 891.752598] env[62066]: _type = "Task" [ 891.752598] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.763185] env[62066]: DEBUG oslo_vmware.api [None req-7882bbc3-f566-458c-8feb-7b619f1c8a07 tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Task: {'id': task-1341071, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.810535] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-a96d87c9-14b1-48f6-a5e4-af778c649411 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] [instance: de2e2be7-efdb-45a8-842a-640ab9deb1d9] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 891.810816] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-a96d87c9-14b1-48f6-a5e4-af778c649411 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] [instance: de2e2be7-efdb-45a8-842a-640ab9deb1d9] Deleting contents of the VM from datastore datastore1 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 891.811017] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-a96d87c9-14b1-48f6-a5e4-af778c649411 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Deleting the datastore file [datastore1] de2e2be7-efdb-45a8-842a-640ab9deb1d9 {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 891.811307] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e56aca7f-27f7-41e8-80cc-e057b0d351c7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.820387] env[62066]: DEBUG oslo_vmware.api [None req-a96d87c9-14b1-48f6-a5e4-af778c649411 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Waiting for the task: (returnval){ [ 891.820387] env[62066]: value = "task-1341073" [ 891.820387] env[62066]: _type = "Task" [ 891.820387] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.831394] env[62066]: DEBUG nova.objects.base [None req-4a3c2cb7-f1e2-437a-b493-e5de08e55742 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Object Instance<8ff6fbb9-c90f-498d-9a85-d220a8c2f794> lazy-loaded attributes: resources,numa_topology {{(pid=62066) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 891.833881] env[62066]: DEBUG oslo_vmware.api [None req-a96d87c9-14b1-48f6-a5e4-af778c649411 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Task: {'id': task-1341073, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.949090] env[62066]: DEBUG oslo_vmware.api [None req-f0861cc1-6af6-4429-bc89-cb4b7f228490 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': task-1341067, 'name': CopyVirtualDisk_Task} progress is 69%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.052177] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddc45362-130c-4b8d-ae08-1f8c18e73e57 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.062601] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f623bbf-4ab5-4232-9537-488d6d72f1e6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.096842] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ce4cd17-71ec-4b65-ae33-0eb8b0dc470c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.107951] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae8f7bed-6329-48e4-adc4-b1aaafd5a594 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.124146] env[62066]: DEBUG nova.compute.provider_tree [None req-4a3c2cb7-f1e2-437a-b493-e5de08e55742 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 892.229168] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c06200c8-d207-47fe-8d97-8cb6c611ec30 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Releasing lock "refresh_cache-b718b09b-a7fa-445b-8be3-5b3ebca210a2" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 892.229453] env[62066]: DEBUG nova.compute.manager [None req-c06200c8-d207-47fe-8d97-8cb6c611ec30 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: b718b09b-a7fa-445b-8be3-5b3ebca210a2] Instance network_info: |[{"id": "3d01745e-0d9e-4bae-ba4d-242380edbab6", "address": "fa:16:3e:ed:53:4c", "network": {"id": "449dfe4e-bb49-43b6-9ba7-b57af74ebfb3", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-282029676-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0d4e33a99c7741fb8cdd97f4ec5dbbd8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db68bd64-5b56-49af-a075-13dcf85cb2e0", "external-id": "nsx-vlan-transportzone-590", "segmentation_id": 590, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3d01745e-0d", "ovs_interfaceid": "3d01745e-0d9e-4bae-ba4d-242380edbab6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 892.229979] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-c06200c8-d207-47fe-8d97-8cb6c611ec30 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: b718b09b-a7fa-445b-8be3-5b3ebca210a2] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ed:53:4c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'db68bd64-5b56-49af-a075-13dcf85cb2e0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3d01745e-0d9e-4bae-ba4d-242380edbab6', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 892.237674] env[62066]: DEBUG oslo.service.loopingcall [None req-c06200c8-d207-47fe-8d97-8cb6c611ec30 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 892.237970] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b718b09b-a7fa-445b-8be3-5b3ebca210a2] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 892.238237] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-fe7e84fe-ae87-4be2-bece-27915e0a90fe {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.265008] env[62066]: DEBUG oslo_vmware.api [None req-7882bbc3-f566-458c-8feb-7b619f1c8a07 tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Task: {'id': task-1341071, 'name': PowerOnVM_Task, 'duration_secs': 0.470747} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.266531] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-7882bbc3-f566-458c-8feb-7b619f1c8a07 tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] [instance: 2634fcc0-96bd-4513-8f92-515f83fe23d7] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 892.266793] env[62066]: DEBUG nova.compute.manager [None req-7882bbc3-f566-458c-8feb-7b619f1c8a07 tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] [instance: 2634fcc0-96bd-4513-8f92-515f83fe23d7] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 892.267120] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 892.267120] env[62066]: value = "task-1341074" [ 892.267120] env[62066]: _type = "Task" [ 892.267120] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.267941] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b32296d4-03d9-4a6f-82c7-8f405be7fd07 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.282017] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1341074, 'name': CreateVM_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.333125] env[62066]: DEBUG oslo_vmware.api [None req-a96d87c9-14b1-48f6-a5e4-af778c649411 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Task: {'id': task-1341073, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.282811} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.333854] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-a96d87c9-14b1-48f6-a5e4-af778c649411 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 892.333905] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-a96d87c9-14b1-48f6-a5e4-af778c649411 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] [instance: de2e2be7-efdb-45a8-842a-640ab9deb1d9] Deleted contents of the VM from datastore datastore1 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 892.334084] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-a96d87c9-14b1-48f6-a5e4-af778c649411 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] [instance: de2e2be7-efdb-45a8-842a-640ab9deb1d9] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 892.334254] env[62066]: INFO nova.compute.manager [None req-a96d87c9-14b1-48f6-a5e4-af778c649411 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] [instance: de2e2be7-efdb-45a8-842a-640ab9deb1d9] Took 1.17 seconds to destroy the instance on the hypervisor. [ 892.334501] env[62066]: DEBUG oslo.service.loopingcall [None req-a96d87c9-14b1-48f6-a5e4-af778c649411 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 892.334717] env[62066]: DEBUG nova.compute.manager [-] [instance: de2e2be7-efdb-45a8-842a-640ab9deb1d9] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 892.334799] env[62066]: DEBUG nova.network.neutron [-] [instance: de2e2be7-efdb-45a8-842a-640ab9deb1d9] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 892.451246] env[62066]: DEBUG oslo_vmware.api [None req-f0861cc1-6af6-4429-bc89-cb4b7f228490 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': task-1341067, 'name': CopyVirtualDisk_Task} progress is 91%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.461497] env[62066]: DEBUG nova.compute.manager [req-e383aa35-0b73-409c-8a50-da5e1b056316 req-fe1cb515-38b6-4d23-b419-7c6e3c94667e service nova] [instance: b718b09b-a7fa-445b-8be3-5b3ebca210a2] Received event network-changed-3d01745e-0d9e-4bae-ba4d-242380edbab6 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 892.461497] env[62066]: DEBUG nova.compute.manager [req-e383aa35-0b73-409c-8a50-da5e1b056316 req-fe1cb515-38b6-4d23-b419-7c6e3c94667e service nova] [instance: b718b09b-a7fa-445b-8be3-5b3ebca210a2] Refreshing instance network info cache due to event network-changed-3d01745e-0d9e-4bae-ba4d-242380edbab6. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 892.461497] env[62066]: DEBUG oslo_concurrency.lockutils [req-e383aa35-0b73-409c-8a50-da5e1b056316 req-fe1cb515-38b6-4d23-b419-7c6e3c94667e service nova] Acquiring lock "refresh_cache-b718b09b-a7fa-445b-8be3-5b3ebca210a2" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 892.461497] env[62066]: DEBUG oslo_concurrency.lockutils [req-e383aa35-0b73-409c-8a50-da5e1b056316 req-fe1cb515-38b6-4d23-b419-7c6e3c94667e service nova] Acquired lock "refresh_cache-b718b09b-a7fa-445b-8be3-5b3ebca210a2" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 892.461497] env[62066]: DEBUG nova.network.neutron [req-e383aa35-0b73-409c-8a50-da5e1b056316 req-fe1cb515-38b6-4d23-b419-7c6e3c94667e service nova] [instance: b718b09b-a7fa-445b-8be3-5b3ebca210a2] Refreshing network info cache for port 3d01745e-0d9e-4bae-ba4d-242380edbab6 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 892.490996] env[62066]: DEBUG oslo_concurrency.lockutils [None req-36f2458b-83e5-4859-9178-b3e847334e66 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Acquiring lock "4e39b9b0-4ab0-4f48-8a3f-6bc1895597eb" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 892.492050] env[62066]: DEBUG oslo_concurrency.lockutils [None req-36f2458b-83e5-4859-9178-b3e847334e66 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Lock "4e39b9b0-4ab0-4f48-8a3f-6bc1895597eb" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 892.627798] env[62066]: DEBUG nova.scheduler.client.report [None req-4a3c2cb7-f1e2-437a-b493-e5de08e55742 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 892.781366] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1341074, 'name': CreateVM_Task} progress is 25%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.952448] env[62066]: DEBUG oslo_vmware.api [None req-f0861cc1-6af6-4429-bc89-cb4b7f228490 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': task-1341067, 'name': CopyVirtualDisk_Task, 'duration_secs': 3.269844} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.952732] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-f0861cc1-6af6-4429-bc89-cb4b7f228490 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/1147d77f-02f3-4d4b-a853-4a0e7d8b0a96/1147d77f-02f3-4d4b-a853-4a0e7d8b0a96.vmdk to [datastore2] 8807ce4d-532f-469c-a302-464c61c7efeb/8807ce4d-532f-469c-a302-464c61c7efeb.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 892.953520] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b39eb1e-23ef-4aa0-b47b-957c1ff23f25 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.977852] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-f0861cc1-6af6-4429-bc89-cb4b7f228490 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 8807ce4d-532f-469c-a302-464c61c7efeb] Reconfiguring VM instance instance-00000054 to attach disk [datastore2] 8807ce4d-532f-469c-a302-464c61c7efeb/8807ce4d-532f-469c-a302-464c61c7efeb.vmdk or device None with type streamOptimized {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 892.978069] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0af02493-1735-451a-9eb6-caceeeac3a96 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.994354] env[62066]: DEBUG nova.compute.manager [None req-36f2458b-83e5-4859-9178-b3e847334e66 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 4e39b9b0-4ab0-4f48-8a3f-6bc1895597eb] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 892.997925] env[62066]: DEBUG oslo_vmware.api [None req-f0861cc1-6af6-4429-bc89-cb4b7f228490 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Waiting for the task: (returnval){ [ 892.997925] env[62066]: value = "task-1341075" [ 892.997925] env[62066]: _type = "Task" [ 892.997925] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.005886] env[62066]: DEBUG oslo_vmware.api [None req-f0861cc1-6af6-4429-bc89-cb4b7f228490 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': task-1341075, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.135384] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4a3c2cb7-f1e2-437a-b493-e5de08e55742 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.312s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 893.137759] env[62066]: DEBUG oslo_concurrency.lockutils [None req-64742fb3-cc48-46bc-b10b-6ecbf4d3709b tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.784s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 893.138087] env[62066]: DEBUG nova.objects.instance [None req-64742fb3-cc48-46bc-b10b-6ecbf4d3709b tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Lazy-loading 'resources' on Instance uuid d77e3405-d158-4a8c-9f54-d5fbd26fed48 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 893.205902] env[62066]: DEBUG nova.network.neutron [req-e383aa35-0b73-409c-8a50-da5e1b056316 req-fe1cb515-38b6-4d23-b419-7c6e3c94667e service nova] [instance: b718b09b-a7fa-445b-8be3-5b3ebca210a2] Updated VIF entry in instance network info cache for port 3d01745e-0d9e-4bae-ba4d-242380edbab6. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 893.206251] env[62066]: DEBUG nova.network.neutron [req-e383aa35-0b73-409c-8a50-da5e1b056316 req-fe1cb515-38b6-4d23-b419-7c6e3c94667e service nova] [instance: b718b09b-a7fa-445b-8be3-5b3ebca210a2] Updating instance_info_cache with network_info: [{"id": "3d01745e-0d9e-4bae-ba4d-242380edbab6", "address": "fa:16:3e:ed:53:4c", "network": {"id": "449dfe4e-bb49-43b6-9ba7-b57af74ebfb3", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-282029676-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0d4e33a99c7741fb8cdd97f4ec5dbbd8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db68bd64-5b56-49af-a075-13dcf85cb2e0", "external-id": "nsx-vlan-transportzone-590", "segmentation_id": 590, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3d01745e-0d", "ovs_interfaceid": "3d01745e-0d9e-4bae-ba4d-242380edbab6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 893.281819] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1341074, 'name': CreateVM_Task} progress is 25%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.294992] env[62066]: DEBUG nova.network.neutron [-] [instance: de2e2be7-efdb-45a8-842a-640ab9deb1d9] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 893.512400] env[62066]: DEBUG oslo_vmware.api [None req-f0861cc1-6af6-4429-bc89-cb4b7f228490 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': task-1341075, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.520346] env[62066]: DEBUG oslo_concurrency.lockutils [None req-36f2458b-83e5-4859-9178-b3e847334e66 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 893.649200] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4a3c2cb7-f1e2-437a-b493-e5de08e55742 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Lock "8ff6fbb9-c90f-498d-9a85-d220a8c2f794" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 29.956s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 893.649822] env[62066]: DEBUG oslo_concurrency.lockutils [None req-81c18bf6-1c8d-45a0-a66c-dcda0761f479 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Lock "8ff6fbb9-c90f-498d-9a85-d220a8c2f794" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 4.333s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 893.650107] env[62066]: INFO nova.compute.manager [None req-81c18bf6-1c8d-45a0-a66c-dcda0761f479 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] [instance: 8ff6fbb9-c90f-498d-9a85-d220a8c2f794] Unshelving [ 893.709191] env[62066]: DEBUG oslo_concurrency.lockutils [req-e383aa35-0b73-409c-8a50-da5e1b056316 req-fe1cb515-38b6-4d23-b419-7c6e3c94667e service nova] Releasing lock "refresh_cache-b718b09b-a7fa-445b-8be3-5b3ebca210a2" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 893.709807] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cb83f312-6ced-4dc9-b42a-28a22511dbae tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Acquiring lock "2634fcc0-96bd-4513-8f92-515f83fe23d7" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 893.709985] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cb83f312-6ced-4dc9-b42a-28a22511dbae tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Lock "2634fcc0-96bd-4513-8f92-515f83fe23d7" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 893.710229] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cb83f312-6ced-4dc9-b42a-28a22511dbae tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Acquiring lock "2634fcc0-96bd-4513-8f92-515f83fe23d7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 893.710425] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cb83f312-6ced-4dc9-b42a-28a22511dbae tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Lock "2634fcc0-96bd-4513-8f92-515f83fe23d7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 893.710656] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cb83f312-6ced-4dc9-b42a-28a22511dbae tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Lock "2634fcc0-96bd-4513-8f92-515f83fe23d7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 893.712743] env[62066]: INFO nova.compute.manager [None req-cb83f312-6ced-4dc9-b42a-28a22511dbae tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] [instance: 2634fcc0-96bd-4513-8f92-515f83fe23d7] Terminating instance [ 893.717241] env[62066]: DEBUG nova.compute.manager [None req-cb83f312-6ced-4dc9-b42a-28a22511dbae tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] [instance: 2634fcc0-96bd-4513-8f92-515f83fe23d7] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 893.717241] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-cb83f312-6ced-4dc9-b42a-28a22511dbae tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] [instance: 2634fcc0-96bd-4513-8f92-515f83fe23d7] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 893.718314] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5422621-9de1-4434-b5fd-927f35296e14 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.725510] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-cb83f312-6ced-4dc9-b42a-28a22511dbae tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] [instance: 2634fcc0-96bd-4513-8f92-515f83fe23d7] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 893.727854] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-52d4b54f-05de-4b60-8538-0d807d21322e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.733615] env[62066]: DEBUG oslo_vmware.api [None req-cb83f312-6ced-4dc9-b42a-28a22511dbae tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Waiting for the task: (returnval){ [ 893.733615] env[62066]: value = "task-1341076" [ 893.733615] env[62066]: _type = "Task" [ 893.733615] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.742360] env[62066]: DEBUG oslo_vmware.api [None req-cb83f312-6ced-4dc9-b42a-28a22511dbae tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Task: {'id': task-1341076, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.780962] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1341074, 'name': CreateVM_Task, 'duration_secs': 1.304885} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.783169] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b718b09b-a7fa-445b-8be3-5b3ebca210a2] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 893.783994] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c06200c8-d207-47fe-8d97-8cb6c611ec30 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 893.784235] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c06200c8-d207-47fe-8d97-8cb6c611ec30 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Acquired lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 893.784568] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c06200c8-d207-47fe-8d97-8cb6c611ec30 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 893.784821] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ea717a2f-15da-4c1e-aa6b-372a2b54802c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.788959] env[62066]: DEBUG oslo_vmware.api [None req-c06200c8-d207-47fe-8d97-8cb6c611ec30 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Waiting for the task: (returnval){ [ 893.788959] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]525127f8-f62f-e835-9548-e0439f05be73" [ 893.788959] env[62066]: _type = "Task" [ 893.788959] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.799815] env[62066]: INFO nova.compute.manager [-] [instance: de2e2be7-efdb-45a8-842a-640ab9deb1d9] Took 1.46 seconds to deallocate network for instance. [ 893.800127] env[62066]: DEBUG oslo_vmware.api [None req-c06200c8-d207-47fe-8d97-8cb6c611ec30 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]525127f8-f62f-e835-9548-e0439f05be73, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.854875] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a633f1ae-117b-4e91-b3ad-af7bb7672b37 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.864227] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c30b25f-3f31-46ad-8898-fd9c8dccbffb {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.897262] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9eae9fb-f3ef-4df7-ac02-44368a7b763c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.905144] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17ebd518-c5a6-43ee-afdb-22bfbc505833 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.918778] env[62066]: DEBUG nova.compute.provider_tree [None req-64742fb3-cc48-46bc-b10b-6ecbf4d3709b tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 894.012545] env[62066]: DEBUG oslo_vmware.api [None req-f0861cc1-6af6-4429-bc89-cb4b7f228490 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': task-1341075, 'name': ReconfigVM_Task, 'duration_secs': 0.861414} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.012817] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-f0861cc1-6af6-4429-bc89-cb4b7f228490 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 8807ce4d-532f-469c-a302-464c61c7efeb] Reconfigured VM instance instance-00000054 to attach disk [datastore2] 8807ce4d-532f-469c-a302-464c61c7efeb/8807ce4d-532f-469c-a302-464c61c7efeb.vmdk or device None with type streamOptimized {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 894.013480] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4c5420a9-d92b-41f2-b0d9-d8b940574a9d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.020435] env[62066]: DEBUG oslo_vmware.api [None req-f0861cc1-6af6-4429-bc89-cb4b7f228490 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Waiting for the task: (returnval){ [ 894.020435] env[62066]: value = "task-1341077" [ 894.020435] env[62066]: _type = "Task" [ 894.020435] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.028720] env[62066]: DEBUG oslo_vmware.api [None req-f0861cc1-6af6-4429-bc89-cb4b7f228490 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': task-1341077, 'name': Rename_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.242842] env[62066]: DEBUG oslo_vmware.api [None req-cb83f312-6ced-4dc9-b42a-28a22511dbae tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Task: {'id': task-1341076, 'name': PowerOffVM_Task, 'duration_secs': 0.181008} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.243150] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-cb83f312-6ced-4dc9-b42a-28a22511dbae tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] [instance: 2634fcc0-96bd-4513-8f92-515f83fe23d7] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 894.243442] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-cb83f312-6ced-4dc9-b42a-28a22511dbae tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] [instance: 2634fcc0-96bd-4513-8f92-515f83fe23d7] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 894.243612] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-dbaec3a6-a42c-4c63-90f2-375ad01ecd39 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.300525] env[62066]: DEBUG oslo_vmware.api [None req-c06200c8-d207-47fe-8d97-8cb6c611ec30 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]525127f8-f62f-e835-9548-e0439f05be73, 'name': SearchDatastore_Task, 'duration_secs': 0.009279} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.300791] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c06200c8-d207-47fe-8d97-8cb6c611ec30 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Releasing lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 894.301107] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-c06200c8-d207-47fe-8d97-8cb6c611ec30 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: b718b09b-a7fa-445b-8be3-5b3ebca210a2] Processing image 50ff584c-3b50-4395-af07-3e66769bc9f7 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 894.301368] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c06200c8-d207-47fe-8d97-8cb6c611ec30 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 894.301521] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c06200c8-d207-47fe-8d97-8cb6c611ec30 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Acquired lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 894.301703] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-c06200c8-d207-47fe-8d97-8cb6c611ec30 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 894.301963] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3393401d-75ad-4e84-be69-d2d0e4f2d29e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.308787] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a96d87c9-14b1-48f6-a5e4-af778c649411 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 894.311555] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-cb83f312-6ced-4dc9-b42a-28a22511dbae tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] [instance: 2634fcc0-96bd-4513-8f92-515f83fe23d7] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 894.311787] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-cb83f312-6ced-4dc9-b42a-28a22511dbae tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] [instance: 2634fcc0-96bd-4513-8f92-515f83fe23d7] Deleting contents of the VM from datastore datastore1 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 894.311973] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-cb83f312-6ced-4dc9-b42a-28a22511dbae tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Deleting the datastore file [datastore1] 2634fcc0-96bd-4513-8f92-515f83fe23d7 {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 894.312979] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0aa6c6e8-6ce9-482f-a277-c18f549e0ded {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.315035] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-c06200c8-d207-47fe-8d97-8cb6c611ec30 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 894.315035] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-c06200c8-d207-47fe-8d97-8cb6c611ec30 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 894.315635] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b40d4bc3-e577-4954-8697-f395e605c46d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.321830] env[62066]: DEBUG oslo_vmware.api [None req-c06200c8-d207-47fe-8d97-8cb6c611ec30 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Waiting for the task: (returnval){ [ 894.321830] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]521add74-694a-e790-62c6-02f08bbb1b63" [ 894.321830] env[62066]: _type = "Task" [ 894.321830] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.323067] env[62066]: DEBUG oslo_vmware.api [None req-cb83f312-6ced-4dc9-b42a-28a22511dbae tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Waiting for the task: (returnval){ [ 894.323067] env[62066]: value = "task-1341079" [ 894.323067] env[62066]: _type = "Task" [ 894.323067] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.333565] env[62066]: DEBUG oslo_vmware.api [None req-c06200c8-d207-47fe-8d97-8cb6c611ec30 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]521add74-694a-e790-62c6-02f08bbb1b63, 'name': SearchDatastore_Task, 'duration_secs': 0.009332} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.336843] env[62066]: DEBUG oslo_vmware.api [None req-cb83f312-6ced-4dc9-b42a-28a22511dbae tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Task: {'id': task-1341079, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.337075] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f31adcdc-f43f-45cf-8651-bd1a4c3d7b8b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.342333] env[62066]: DEBUG oslo_vmware.api [None req-c06200c8-d207-47fe-8d97-8cb6c611ec30 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Waiting for the task: (returnval){ [ 894.342333] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]5238e108-1cb0-3f09-0acc-56e44e642339" [ 894.342333] env[62066]: _type = "Task" [ 894.342333] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.349830] env[62066]: DEBUG oslo_vmware.api [None req-c06200c8-d207-47fe-8d97-8cb6c611ec30 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]5238e108-1cb0-3f09-0acc-56e44e642339, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.422122] env[62066]: DEBUG nova.scheduler.client.report [None req-64742fb3-cc48-46bc-b10b-6ecbf4d3709b tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 894.483668] env[62066]: DEBUG nova.compute.manager [req-43a7b333-a879-44ee-a474-6185dd5c9c67 req-a1ee0f16-7d39-48da-b3eb-b0c9cd960074 service nova] [instance: de2e2be7-efdb-45a8-842a-640ab9deb1d9] Received event network-vif-deleted-c833a993-4af0-4bdb-90d2-cc82a796a560 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 894.530606] env[62066]: DEBUG oslo_vmware.api [None req-f0861cc1-6af6-4429-bc89-cb4b7f228490 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': task-1341077, 'name': Rename_Task, 'duration_secs': 0.181062} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.530869] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-f0861cc1-6af6-4429-bc89-cb4b7f228490 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 8807ce4d-532f-469c-a302-464c61c7efeb] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 894.531018] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d58acd96-13e3-430e-89ac-da0b575d04c4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.536862] env[62066]: DEBUG oslo_vmware.api [None req-f0861cc1-6af6-4429-bc89-cb4b7f228490 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Waiting for the task: (returnval){ [ 894.536862] env[62066]: value = "task-1341080" [ 894.536862] env[62066]: _type = "Task" [ 894.536862] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.545853] env[62066]: DEBUG oslo_vmware.api [None req-f0861cc1-6af6-4429-bc89-cb4b7f228490 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': task-1341080, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.671507] env[62066]: DEBUG oslo_concurrency.lockutils [None req-81c18bf6-1c8d-45a0-a66c-dcda0761f479 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 894.838034] env[62066]: DEBUG oslo_vmware.api [None req-cb83f312-6ced-4dc9-b42a-28a22511dbae tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Task: {'id': task-1341079, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.186451} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.838034] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-cb83f312-6ced-4dc9-b42a-28a22511dbae tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 894.838034] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-cb83f312-6ced-4dc9-b42a-28a22511dbae tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] [instance: 2634fcc0-96bd-4513-8f92-515f83fe23d7] Deleted contents of the VM from datastore datastore1 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 894.838034] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-cb83f312-6ced-4dc9-b42a-28a22511dbae tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] [instance: 2634fcc0-96bd-4513-8f92-515f83fe23d7] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 894.838344] env[62066]: INFO nova.compute.manager [None req-cb83f312-6ced-4dc9-b42a-28a22511dbae tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] [instance: 2634fcc0-96bd-4513-8f92-515f83fe23d7] Took 1.12 seconds to destroy the instance on the hypervisor. [ 894.838441] env[62066]: DEBUG oslo.service.loopingcall [None req-cb83f312-6ced-4dc9-b42a-28a22511dbae tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 894.838659] env[62066]: DEBUG nova.compute.manager [-] [instance: 2634fcc0-96bd-4513-8f92-515f83fe23d7] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 894.838755] env[62066]: DEBUG nova.network.neutron [-] [instance: 2634fcc0-96bd-4513-8f92-515f83fe23d7] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 894.851508] env[62066]: DEBUG oslo_vmware.api [None req-c06200c8-d207-47fe-8d97-8cb6c611ec30 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]5238e108-1cb0-3f09-0acc-56e44e642339, 'name': SearchDatastore_Task, 'duration_secs': 0.010301} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.851508] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c06200c8-d207-47fe-8d97-8cb6c611ec30 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Releasing lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 894.851698] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-c06200c8-d207-47fe-8d97-8cb6c611ec30 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk to [datastore2] b718b09b-a7fa-445b-8be3-5b3ebca210a2/b718b09b-a7fa-445b-8be3-5b3ebca210a2.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 894.851996] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-299280cd-8b3e-4964-ad7c-0ed594324fd3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.858178] env[62066]: DEBUG oslo_vmware.api [None req-c06200c8-d207-47fe-8d97-8cb6c611ec30 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Waiting for the task: (returnval){ [ 894.858178] env[62066]: value = "task-1341081" [ 894.858178] env[62066]: _type = "Task" [ 894.858178] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.865532] env[62066]: DEBUG oslo_vmware.api [None req-c06200c8-d207-47fe-8d97-8cb6c611ec30 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': task-1341081, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.927613] env[62066]: DEBUG oslo_concurrency.lockutils [None req-64742fb3-cc48-46bc-b10b-6ecbf4d3709b tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.790s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 894.929794] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cbf7a2c2-7fb7-405b-bf5c-0776ba26eb99 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.210s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 894.931377] env[62066]: INFO nova.compute.claims [None req-cbf7a2c2-7fb7-405b-bf5c-0776ba26eb99 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 9c989152-ec69-478e-a0b0-62983852576c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 894.946685] env[62066]: INFO nova.scheduler.client.report [None req-64742fb3-cc48-46bc-b10b-6ecbf4d3709b tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Deleted allocations for instance d77e3405-d158-4a8c-9f54-d5fbd26fed48 [ 895.046811] env[62066]: DEBUG oslo_vmware.api [None req-f0861cc1-6af6-4429-bc89-cb4b7f228490 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': task-1341080, 'name': PowerOnVM_Task, 'duration_secs': 0.467559} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.047008] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-f0861cc1-6af6-4429-bc89-cb4b7f228490 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 8807ce4d-532f-469c-a302-464c61c7efeb] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 895.047444] env[62066]: INFO nova.compute.manager [None req-f0861cc1-6af6-4429-bc89-cb4b7f228490 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 8807ce4d-532f-469c-a302-464c61c7efeb] Took 15.97 seconds to spawn the instance on the hypervisor. [ 895.047444] env[62066]: DEBUG nova.compute.manager [None req-f0861cc1-6af6-4429-bc89-cb4b7f228490 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 8807ce4d-532f-469c-a302-464c61c7efeb] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 895.048206] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7882308-d051-45d9-af03-474e723b85f2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.369089] env[62066]: DEBUG oslo_vmware.api [None req-c06200c8-d207-47fe-8d97-8cb6c611ec30 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': task-1341081, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.456424] env[62066]: DEBUG oslo_concurrency.lockutils [None req-64742fb3-cc48-46bc-b10b-6ecbf4d3709b tempest-ServerShowV247Test-1178657913 tempest-ServerShowV247Test-1178657913-project-member] Lock "d77e3405-d158-4a8c-9f54-d5fbd26fed48" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 12.870s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 895.567044] env[62066]: INFO nova.compute.manager [None req-f0861cc1-6af6-4429-bc89-cb4b7f228490 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 8807ce4d-532f-469c-a302-464c61c7efeb] Took 28.48 seconds to build instance. [ 895.581650] env[62066]: DEBUG nova.network.neutron [-] [instance: 2634fcc0-96bd-4513-8f92-515f83fe23d7] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 895.869580] env[62066]: DEBUG oslo_vmware.api [None req-c06200c8-d207-47fe-8d97-8cb6c611ec30 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': task-1341081, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.609291} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.869869] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-c06200c8-d207-47fe-8d97-8cb6c611ec30 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk to [datastore2] b718b09b-a7fa-445b-8be3-5b3ebca210a2/b718b09b-a7fa-445b-8be3-5b3ebca210a2.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 895.870170] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-c06200c8-d207-47fe-8d97-8cb6c611ec30 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: b718b09b-a7fa-445b-8be3-5b3ebca210a2] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 895.870458] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-90a8a9a1-52d5-46bf-99a0-b618e699131a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.877196] env[62066]: DEBUG oslo_vmware.api [None req-c06200c8-d207-47fe-8d97-8cb6c611ec30 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Waiting for the task: (returnval){ [ 895.877196] env[62066]: value = "task-1341082" [ 895.877196] env[62066]: _type = "Task" [ 895.877196] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.885100] env[62066]: DEBUG oslo_vmware.api [None req-c06200c8-d207-47fe-8d97-8cb6c611ec30 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': task-1341082, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.069672] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f0861cc1-6af6-4429-bc89-cb4b7f228490 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Lock "8807ce4d-532f-469c-a302-464c61c7efeb" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 30.013s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 896.084177] env[62066]: INFO nova.compute.manager [-] [instance: 2634fcc0-96bd-4513-8f92-515f83fe23d7] Took 1.25 seconds to deallocate network for instance. [ 896.162036] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de2edc03-b892-4c8a-ab16-017a3468812b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.173456] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53cd49c0-6486-41eb-a571-ec2451f13fbc {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.203785] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abbb3b57-9c1e-4cb6-abce-b65bad159d3e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.211408] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eeeeedcb-4bad-46c1-87f4-497f5900fe12 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.226314] env[62066]: DEBUG nova.compute.provider_tree [None req-cbf7a2c2-7fb7-405b-bf5c-0776ba26eb99 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 896.387433] env[62066]: DEBUG oslo_vmware.api [None req-c06200c8-d207-47fe-8d97-8cb6c611ec30 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': task-1341082, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.222522} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.387877] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-c06200c8-d207-47fe-8d97-8cb6c611ec30 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: b718b09b-a7fa-445b-8be3-5b3ebca210a2] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 896.388680] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-294b3468-ecd8-473f-9b33-2617cc3cf484 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.411072] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-c06200c8-d207-47fe-8d97-8cb6c611ec30 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: b718b09b-a7fa-445b-8be3-5b3ebca210a2] Reconfiguring VM instance instance-00000055 to attach disk [datastore2] b718b09b-a7fa-445b-8be3-5b3ebca210a2/b718b09b-a7fa-445b-8be3-5b3ebca210a2.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 896.411630] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6bf542d5-156e-4224-977b-88079f5d50fa {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.430586] env[62066]: DEBUG oslo_vmware.api [None req-c06200c8-d207-47fe-8d97-8cb6c611ec30 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Waiting for the task: (returnval){ [ 896.430586] env[62066]: value = "task-1341083" [ 896.430586] env[62066]: _type = "Task" [ 896.430586] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.438107] env[62066]: DEBUG oslo_vmware.api [None req-c06200c8-d207-47fe-8d97-8cb6c611ec30 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': task-1341083, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.507582] env[62066]: DEBUG nova.compute.manager [req-3b46081a-80e2-4dba-8ce9-4a1a18acea05 req-58ce4ed9-2c6e-4e7c-816a-f8a3b4507d3c service nova] [instance: 2634fcc0-96bd-4513-8f92-515f83fe23d7] Received event network-vif-deleted-cf42ffe4-6621-4c29-9a6d-687ade676346 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 896.593029] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cb83f312-6ced-4dc9-b42a-28a22511dbae tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 896.729110] env[62066]: DEBUG nova.scheduler.client.report [None req-cbf7a2c2-7fb7-405b-bf5c-0776ba26eb99 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 896.766985] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3ba3ebd2-6915-4f7b-8349-9384d9a8c223 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Acquiring lock "8807ce4d-532f-469c-a302-464c61c7efeb" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 896.767293] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3ba3ebd2-6915-4f7b-8349-9384d9a8c223 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Lock "8807ce4d-532f-469c-a302-464c61c7efeb" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 896.767514] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3ba3ebd2-6915-4f7b-8349-9384d9a8c223 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Acquiring lock "8807ce4d-532f-469c-a302-464c61c7efeb-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 896.768200] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3ba3ebd2-6915-4f7b-8349-9384d9a8c223 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Lock "8807ce4d-532f-469c-a302-464c61c7efeb-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 896.768336] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3ba3ebd2-6915-4f7b-8349-9384d9a8c223 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Lock "8807ce4d-532f-469c-a302-464c61c7efeb-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 896.770468] env[62066]: INFO nova.compute.manager [None req-3ba3ebd2-6915-4f7b-8349-9384d9a8c223 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 8807ce4d-532f-469c-a302-464c61c7efeb] Terminating instance [ 896.772775] env[62066]: DEBUG nova.compute.manager [None req-3ba3ebd2-6915-4f7b-8349-9384d9a8c223 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 8807ce4d-532f-469c-a302-464c61c7efeb] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 896.772979] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-3ba3ebd2-6915-4f7b-8349-9384d9a8c223 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 8807ce4d-532f-469c-a302-464c61c7efeb] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 896.773841] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d15087e8-c6e6-476a-af4f-fd84cf94e6d0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.781975] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-3ba3ebd2-6915-4f7b-8349-9384d9a8c223 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 8807ce4d-532f-469c-a302-464c61c7efeb] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 896.781975] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6ffed713-565b-43ce-8b1e-3a763fabaf19 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.789393] env[62066]: DEBUG oslo_vmware.api [None req-3ba3ebd2-6915-4f7b-8349-9384d9a8c223 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Waiting for the task: (returnval){ [ 896.789393] env[62066]: value = "task-1341084" [ 896.789393] env[62066]: _type = "Task" [ 896.789393] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.796725] env[62066]: DEBUG oslo_vmware.api [None req-3ba3ebd2-6915-4f7b-8349-9384d9a8c223 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': task-1341084, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.941080] env[62066]: DEBUG oslo_vmware.api [None req-c06200c8-d207-47fe-8d97-8cb6c611ec30 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': task-1341083, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.207406] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ad6adfa4-e5e7-49bc-8c71-f4f4f133daea tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Acquiring lock "15afaea0-65de-4f32-851c-365003e5498f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 897.207660] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ad6adfa4-e5e7-49bc-8c71-f4f4f133daea tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Lock "15afaea0-65de-4f32-851c-365003e5498f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 897.234277] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cbf7a2c2-7fb7-405b-bf5c-0776ba26eb99 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.304s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 897.234754] env[62066]: DEBUG nova.compute.manager [None req-cbf7a2c2-7fb7-405b-bf5c-0776ba26eb99 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 9c989152-ec69-478e-a0b0-62983852576c] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 897.237345] env[62066]: DEBUG oslo_concurrency.lockutils [None req-61031532-cb43-40c8-81a2-438112a5a962 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 8.958s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 897.237568] env[62066]: DEBUG nova.objects.instance [None req-61031532-cb43-40c8-81a2-438112a5a962 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Lazy-loading 'resources' on Instance uuid 76306c3f-0674-40fe-9864-d82d11e4bb5e {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 897.298434] env[62066]: DEBUG oslo_vmware.api [None req-3ba3ebd2-6915-4f7b-8349-9384d9a8c223 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': task-1341084, 'name': PowerOffVM_Task, 'duration_secs': 0.190028} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.298689] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-3ba3ebd2-6915-4f7b-8349-9384d9a8c223 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 8807ce4d-532f-469c-a302-464c61c7efeb] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 897.298865] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-3ba3ebd2-6915-4f7b-8349-9384d9a8c223 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 8807ce4d-532f-469c-a302-464c61c7efeb] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 897.299206] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-cd490d61-6d5a-43ee-8d7b-cf1f386928c8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.358073] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-3ba3ebd2-6915-4f7b-8349-9384d9a8c223 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 8807ce4d-532f-469c-a302-464c61c7efeb] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 897.359859] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-3ba3ebd2-6915-4f7b-8349-9384d9a8c223 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 8807ce4d-532f-469c-a302-464c61c7efeb] Deleting contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 897.359859] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-3ba3ebd2-6915-4f7b-8349-9384d9a8c223 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Deleting the datastore file [datastore2] 8807ce4d-532f-469c-a302-464c61c7efeb {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 897.359859] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9b825ace-b73d-44bc-b3ea-7a9c77ab09d9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.367381] env[62066]: DEBUG oslo_vmware.api [None req-3ba3ebd2-6915-4f7b-8349-9384d9a8c223 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Waiting for the task: (returnval){ [ 897.367381] env[62066]: value = "task-1341086" [ 897.367381] env[62066]: _type = "Task" [ 897.367381] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.374891] env[62066]: DEBUG oslo_vmware.api [None req-3ba3ebd2-6915-4f7b-8349-9384d9a8c223 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': task-1341086, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.440831] env[62066]: DEBUG oslo_vmware.api [None req-c06200c8-d207-47fe-8d97-8cb6c611ec30 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': task-1341083, 'name': ReconfigVM_Task, 'duration_secs': 0.518384} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.441170] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-c06200c8-d207-47fe-8d97-8cb6c611ec30 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: b718b09b-a7fa-445b-8be3-5b3ebca210a2] Reconfigured VM instance instance-00000055 to attach disk [datastore2] b718b09b-a7fa-445b-8be3-5b3ebca210a2/b718b09b-a7fa-445b-8be3-5b3ebca210a2.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 897.441808] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6ba4d19c-c029-412a-9163-053c1c8ae920 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.447493] env[62066]: DEBUG oslo_vmware.api [None req-c06200c8-d207-47fe-8d97-8cb6c611ec30 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Waiting for the task: (returnval){ [ 897.447493] env[62066]: value = "task-1341087" [ 897.447493] env[62066]: _type = "Task" [ 897.447493] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.456136] env[62066]: DEBUG oslo_vmware.api [None req-c06200c8-d207-47fe-8d97-8cb6c611ec30 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': task-1341087, 'name': Rename_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.710433] env[62066]: DEBUG nova.compute.manager [None req-ad6adfa4-e5e7-49bc-8c71-f4f4f133daea tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] [instance: 15afaea0-65de-4f32-851c-365003e5498f] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 897.745498] env[62066]: DEBUG nova.compute.utils [None req-cbf7a2c2-7fb7-405b-bf5c-0776ba26eb99 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 897.747410] env[62066]: DEBUG nova.compute.manager [None req-cbf7a2c2-7fb7-405b-bf5c-0776ba26eb99 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 9c989152-ec69-478e-a0b0-62983852576c] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 897.747410] env[62066]: DEBUG nova.network.neutron [None req-cbf7a2c2-7fb7-405b-bf5c-0776ba26eb99 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 9c989152-ec69-478e-a0b0-62983852576c] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 897.803166] env[62066]: DEBUG nova.policy [None req-cbf7a2c2-7fb7-405b-bf5c-0776ba26eb99 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2ab20cbd0bab4ae7ba46d9135605a509', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd609babc987a42e2a8ddb4bfb9c3b103', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 897.879738] env[62066]: DEBUG oslo_vmware.api [None req-3ba3ebd2-6915-4f7b-8349-9384d9a8c223 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': task-1341086, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.131884} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.879738] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-3ba3ebd2-6915-4f7b-8349-9384d9a8c223 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 897.879892] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-3ba3ebd2-6915-4f7b-8349-9384d9a8c223 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 8807ce4d-532f-469c-a302-464c61c7efeb] Deleted contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 897.880033] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-3ba3ebd2-6915-4f7b-8349-9384d9a8c223 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 8807ce4d-532f-469c-a302-464c61c7efeb] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 897.880257] env[62066]: INFO nova.compute.manager [None req-3ba3ebd2-6915-4f7b-8349-9384d9a8c223 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 8807ce4d-532f-469c-a302-464c61c7efeb] Took 1.11 seconds to destroy the instance on the hypervisor. [ 897.880548] env[62066]: DEBUG oslo.service.loopingcall [None req-3ba3ebd2-6915-4f7b-8349-9384d9a8c223 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 897.880836] env[62066]: DEBUG nova.compute.manager [-] [instance: 8807ce4d-532f-469c-a302-464c61c7efeb] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 897.880954] env[62066]: DEBUG nova.network.neutron [-] [instance: 8807ce4d-532f-469c-a302-464c61c7efeb] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 897.959091] env[62066]: DEBUG oslo_vmware.api [None req-c06200c8-d207-47fe-8d97-8cb6c611ec30 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': task-1341087, 'name': Rename_Task, 'duration_secs': 0.141925} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.959922] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-c06200c8-d207-47fe-8d97-8cb6c611ec30 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: b718b09b-a7fa-445b-8be3-5b3ebca210a2] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 897.960746] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1336af8f-0534-4f61-a801-6d4922f1645c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.963045] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a3cff8e3-f34b-4c7e-bcf8-be5857328ec9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.969176] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d13144fc-53b4-4fca-8e18-0d340d103703 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.972967] env[62066]: DEBUG oslo_vmware.api [None req-c06200c8-d207-47fe-8d97-8cb6c611ec30 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Waiting for the task: (returnval){ [ 897.972967] env[62066]: value = "task-1341088" [ 897.972967] env[62066]: _type = "Task" [ 897.972967] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.980570] env[62066]: DEBUG oslo_vmware.api [None req-c06200c8-d207-47fe-8d97-8cb6c611ec30 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': task-1341088, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.006380] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2da286b1-60c6-4acd-9088-f84b9fa869ea {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.014220] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d16061a-077b-4559-8c3e-72410d29d577 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.029160] env[62066]: DEBUG nova.compute.provider_tree [None req-61031532-cb43-40c8-81a2-438112a5a962 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 898.061712] env[62066]: DEBUG nova.network.neutron [None req-cbf7a2c2-7fb7-405b-bf5c-0776ba26eb99 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 9c989152-ec69-478e-a0b0-62983852576c] Successfully created port: 1c0d0075-7f22-441e-a283-21e9b5c5aecc {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 898.128035] env[62066]: DEBUG nova.compute.manager [req-af3f4c99-f4db-4fb9-a4c1-4c9d8c4e9743 req-505d0394-8fa0-4039-ac66-ce194a59ed5f service nova] [instance: 8807ce4d-532f-469c-a302-464c61c7efeb] Received event network-vif-deleted-dba295b3-a285-4890-b73b-ed681d77192a {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 898.128329] env[62066]: INFO nova.compute.manager [req-af3f4c99-f4db-4fb9-a4c1-4c9d8c4e9743 req-505d0394-8fa0-4039-ac66-ce194a59ed5f service nova] [instance: 8807ce4d-532f-469c-a302-464c61c7efeb] Neutron deleted interface dba295b3-a285-4890-b73b-ed681d77192a; detaching it from the instance and deleting it from the info cache [ 898.128441] env[62066]: DEBUG nova.network.neutron [req-af3f4c99-f4db-4fb9-a4c1-4c9d8c4e9743 req-505d0394-8fa0-4039-ac66-ce194a59ed5f service nova] [instance: 8807ce4d-532f-469c-a302-464c61c7efeb] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 898.229032] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ad6adfa4-e5e7-49bc-8c71-f4f4f133daea tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 898.251990] env[62066]: DEBUG nova.compute.manager [None req-cbf7a2c2-7fb7-405b-bf5c-0776ba26eb99 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 9c989152-ec69-478e-a0b0-62983852576c] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 898.483931] env[62066]: DEBUG oslo_vmware.api [None req-c06200c8-d207-47fe-8d97-8cb6c611ec30 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': task-1341088, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.532280] env[62066]: DEBUG nova.scheduler.client.report [None req-61031532-cb43-40c8-81a2-438112a5a962 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 898.604086] env[62066]: DEBUG nova.network.neutron [-] [instance: 8807ce4d-532f-469c-a302-464c61c7efeb] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 898.633303] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ac43ef1d-6a07-40c7-9c01-4ecebcf92e2e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.641986] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7649dbc-bbdd-4055-8511-1145e3e193e6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.670088] env[62066]: DEBUG nova.compute.manager [req-af3f4c99-f4db-4fb9-a4c1-4c9d8c4e9743 req-505d0394-8fa0-4039-ac66-ce194a59ed5f service nova] [instance: 8807ce4d-532f-469c-a302-464c61c7efeb] Detach interface failed, port_id=dba295b3-a285-4890-b73b-ed681d77192a, reason: Instance 8807ce4d-532f-469c-a302-464c61c7efeb could not be found. {{(pid=62066) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 898.984712] env[62066]: DEBUG oslo_vmware.api [None req-c06200c8-d207-47fe-8d97-8cb6c611ec30 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': task-1341088, 'name': PowerOnVM_Task, 'duration_secs': 0.697288} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.985064] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-c06200c8-d207-47fe-8d97-8cb6c611ec30 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: b718b09b-a7fa-445b-8be3-5b3ebca210a2] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 898.985226] env[62066]: INFO nova.compute.manager [None req-c06200c8-d207-47fe-8d97-8cb6c611ec30 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: b718b09b-a7fa-445b-8be3-5b3ebca210a2] Took 9.02 seconds to spawn the instance on the hypervisor. [ 898.985421] env[62066]: DEBUG nova.compute.manager [None req-c06200c8-d207-47fe-8d97-8cb6c611ec30 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: b718b09b-a7fa-445b-8be3-5b3ebca210a2] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 898.986220] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a98fa359-bd0e-4654-b296-9618162d4bd0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.036989] env[62066]: DEBUG oslo_concurrency.lockutils [None req-61031532-cb43-40c8-81a2-438112a5a962 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.800s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 899.039230] env[62066]: DEBUG oslo_concurrency.lockutils [None req-adc52c21-6d1b-4740-84d8-9cdb3fbc5d28 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.339s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 899.040964] env[62066]: INFO nova.compute.claims [None req-adc52c21-6d1b-4740-84d8-9cdb3fbc5d28 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: 90c33434-c127-450b-9ff0-75181b4ac385] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 899.059258] env[62066]: INFO nova.scheduler.client.report [None req-61031532-cb43-40c8-81a2-438112a5a962 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Deleted allocations for instance 76306c3f-0674-40fe-9864-d82d11e4bb5e [ 899.107246] env[62066]: INFO nova.compute.manager [-] [instance: 8807ce4d-532f-469c-a302-464c61c7efeb] Took 1.23 seconds to deallocate network for instance. [ 899.261355] env[62066]: DEBUG nova.compute.manager [None req-cbf7a2c2-7fb7-405b-bf5c-0776ba26eb99 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 9c989152-ec69-478e-a0b0-62983852576c] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 899.285711] env[62066]: DEBUG nova.virt.hardware [None req-cbf7a2c2-7fb7-405b-bf5c-0776ba26eb99 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-23T13:40:41Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-23T13:40:23Z,direct_url=,disk_format='vmdk',id=50ff584c-3b50-4395-af07-3e66769bc9f7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='eb52f7069a374c61ae946f052007c6d9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-23T13:40:24Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 899.285972] env[62066]: DEBUG nova.virt.hardware [None req-cbf7a2c2-7fb7-405b-bf5c-0776ba26eb99 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 899.286157] env[62066]: DEBUG nova.virt.hardware [None req-cbf7a2c2-7fb7-405b-bf5c-0776ba26eb99 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 899.286349] env[62066]: DEBUG nova.virt.hardware [None req-cbf7a2c2-7fb7-405b-bf5c-0776ba26eb99 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 899.286500] env[62066]: DEBUG nova.virt.hardware [None req-cbf7a2c2-7fb7-405b-bf5c-0776ba26eb99 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 899.286649] env[62066]: DEBUG nova.virt.hardware [None req-cbf7a2c2-7fb7-405b-bf5c-0776ba26eb99 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 899.286858] env[62066]: DEBUG nova.virt.hardware [None req-cbf7a2c2-7fb7-405b-bf5c-0776ba26eb99 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 899.287032] env[62066]: DEBUG nova.virt.hardware [None req-cbf7a2c2-7fb7-405b-bf5c-0776ba26eb99 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 899.287212] env[62066]: DEBUG nova.virt.hardware [None req-cbf7a2c2-7fb7-405b-bf5c-0776ba26eb99 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 899.287377] env[62066]: DEBUG nova.virt.hardware [None req-cbf7a2c2-7fb7-405b-bf5c-0776ba26eb99 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 899.287550] env[62066]: DEBUG nova.virt.hardware [None req-cbf7a2c2-7fb7-405b-bf5c-0776ba26eb99 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 899.288415] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ad095e5-c631-489d-980c-d21c053272ba {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.296328] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de545021-fbb7-4f63-a9b5-18271afd23e3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.456745] env[62066]: DEBUG nova.compute.manager [req-47038d09-7892-4962-82e0-e65188601a44 req-2b258769-dac9-4202-a64b-af227fc9ba4f service nova] [instance: 9c989152-ec69-478e-a0b0-62983852576c] Received event network-vif-plugged-1c0d0075-7f22-441e-a283-21e9b5c5aecc {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 899.456957] env[62066]: DEBUG oslo_concurrency.lockutils [req-47038d09-7892-4962-82e0-e65188601a44 req-2b258769-dac9-4202-a64b-af227fc9ba4f service nova] Acquiring lock "9c989152-ec69-478e-a0b0-62983852576c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 899.457192] env[62066]: DEBUG oslo_concurrency.lockutils [req-47038d09-7892-4962-82e0-e65188601a44 req-2b258769-dac9-4202-a64b-af227fc9ba4f service nova] Lock "9c989152-ec69-478e-a0b0-62983852576c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 899.457368] env[62066]: DEBUG oslo_concurrency.lockutils [req-47038d09-7892-4962-82e0-e65188601a44 req-2b258769-dac9-4202-a64b-af227fc9ba4f service nova] Lock "9c989152-ec69-478e-a0b0-62983852576c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 899.457544] env[62066]: DEBUG nova.compute.manager [req-47038d09-7892-4962-82e0-e65188601a44 req-2b258769-dac9-4202-a64b-af227fc9ba4f service nova] [instance: 9c989152-ec69-478e-a0b0-62983852576c] No waiting events found dispatching network-vif-plugged-1c0d0075-7f22-441e-a283-21e9b5c5aecc {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 899.457713] env[62066]: WARNING nova.compute.manager [req-47038d09-7892-4962-82e0-e65188601a44 req-2b258769-dac9-4202-a64b-af227fc9ba4f service nova] [instance: 9c989152-ec69-478e-a0b0-62983852576c] Received unexpected event network-vif-plugged-1c0d0075-7f22-441e-a283-21e9b5c5aecc for instance with vm_state building and task_state spawning. [ 899.507592] env[62066]: INFO nova.compute.manager [None req-c06200c8-d207-47fe-8d97-8cb6c611ec30 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: b718b09b-a7fa-445b-8be3-5b3ebca210a2] Took 18.79 seconds to build instance. [ 899.543753] env[62066]: DEBUG nova.network.neutron [None req-cbf7a2c2-7fb7-405b-bf5c-0776ba26eb99 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 9c989152-ec69-478e-a0b0-62983852576c] Successfully updated port: 1c0d0075-7f22-441e-a283-21e9b5c5aecc {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 899.567303] env[62066]: DEBUG oslo_concurrency.lockutils [None req-61031532-cb43-40c8-81a2-438112a5a962 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Lock "76306c3f-0674-40fe-9864-d82d11e4bb5e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 14.571s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 899.615093] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3ba3ebd2-6915-4f7b-8349-9384d9a8c223 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 900.009708] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c06200c8-d207-47fe-8d97-8cb6c611ec30 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Lock "b718b09b-a7fa-445b-8be3-5b3ebca210a2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.299s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 900.051691] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cbf7a2c2-7fb7-405b-bf5c-0776ba26eb99 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Acquiring lock "refresh_cache-9c989152-ec69-478e-a0b0-62983852576c" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 900.051896] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cbf7a2c2-7fb7-405b-bf5c-0776ba26eb99 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Acquired lock "refresh_cache-9c989152-ec69-478e-a0b0-62983852576c" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 900.052100] env[62066]: DEBUG nova.network.neutron [None req-cbf7a2c2-7fb7-405b-bf5c-0776ba26eb99 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 9c989152-ec69-478e-a0b0-62983852576c] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 900.173079] env[62066]: DEBUG nova.compute.manager [req-c9c84ac9-d773-47ae-8202-e9d022cef8ff req-e14120d6-ac55-44cc-8858-b0fdcf03985c service nova] [instance: b718b09b-a7fa-445b-8be3-5b3ebca210a2] Received event network-changed-3d01745e-0d9e-4bae-ba4d-242380edbab6 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 900.173294] env[62066]: DEBUG nova.compute.manager [req-c9c84ac9-d773-47ae-8202-e9d022cef8ff req-e14120d6-ac55-44cc-8858-b0fdcf03985c service nova] [instance: b718b09b-a7fa-445b-8be3-5b3ebca210a2] Refreshing instance network info cache due to event network-changed-3d01745e-0d9e-4bae-ba4d-242380edbab6. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 900.173509] env[62066]: DEBUG oslo_concurrency.lockutils [req-c9c84ac9-d773-47ae-8202-e9d022cef8ff req-e14120d6-ac55-44cc-8858-b0fdcf03985c service nova] Acquiring lock "refresh_cache-b718b09b-a7fa-445b-8be3-5b3ebca210a2" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 900.173658] env[62066]: DEBUG oslo_concurrency.lockutils [req-c9c84ac9-d773-47ae-8202-e9d022cef8ff req-e14120d6-ac55-44cc-8858-b0fdcf03985c service nova] Acquired lock "refresh_cache-b718b09b-a7fa-445b-8be3-5b3ebca210a2" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 900.173827] env[62066]: DEBUG nova.network.neutron [req-c9c84ac9-d773-47ae-8202-e9d022cef8ff req-e14120d6-ac55-44cc-8858-b0fdcf03985c service nova] [instance: b718b09b-a7fa-445b-8be3-5b3ebca210a2] Refreshing network info cache for port 3d01745e-0d9e-4bae-ba4d-242380edbab6 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 900.243037] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90d4d4cb-2e9c-4d66-98d8-4a25dd5cb890 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.250670] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11e28e06-3097-4b5c-9907-4c6726d5b816 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.280140] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d1abf7d-53f1-4db6-8f65-fe5b887baf24 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.287653] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f07f83e-b78a-4f4f-a7ec-b9b9f02e6eb4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.300505] env[62066]: DEBUG nova.compute.provider_tree [None req-adc52c21-6d1b-4740-84d8-9cdb3fbc5d28 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 900.585869] env[62066]: DEBUG nova.network.neutron [None req-cbf7a2c2-7fb7-405b-bf5c-0776ba26eb99 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 9c989152-ec69-478e-a0b0-62983852576c] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 900.707701] env[62066]: DEBUG nova.network.neutron [None req-cbf7a2c2-7fb7-405b-bf5c-0776ba26eb99 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 9c989152-ec69-478e-a0b0-62983852576c] Updating instance_info_cache with network_info: [{"id": "1c0d0075-7f22-441e-a283-21e9b5c5aecc", "address": "fa:16:3e:c5:c3:07", "network": {"id": "1c6c51e4-96f8-4175-b7f5-96b8a42fcb13", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-2006089809-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d609babc987a42e2a8ddb4bfb9c3b103", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a965790c-2d2f-4c2a-9ee7-745f4d53039b", "external-id": "nsx-vlan-transportzone-708", "segmentation_id": 708, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1c0d0075-7f", "ovs_interfaceid": "1c0d0075-7f22-441e-a283-21e9b5c5aecc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 900.803815] env[62066]: DEBUG nova.scheduler.client.report [None req-adc52c21-6d1b-4740-84d8-9cdb3fbc5d28 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 900.897344] env[62066]: DEBUG nova.network.neutron [req-c9c84ac9-d773-47ae-8202-e9d022cef8ff req-e14120d6-ac55-44cc-8858-b0fdcf03985c service nova] [instance: b718b09b-a7fa-445b-8be3-5b3ebca210a2] Updated VIF entry in instance network info cache for port 3d01745e-0d9e-4bae-ba4d-242380edbab6. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 900.897708] env[62066]: DEBUG nova.network.neutron [req-c9c84ac9-d773-47ae-8202-e9d022cef8ff req-e14120d6-ac55-44cc-8858-b0fdcf03985c service nova] [instance: b718b09b-a7fa-445b-8be3-5b3ebca210a2] Updating instance_info_cache with network_info: [{"id": "3d01745e-0d9e-4bae-ba4d-242380edbab6", "address": "fa:16:3e:ed:53:4c", "network": {"id": "449dfe4e-bb49-43b6-9ba7-b57af74ebfb3", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-282029676-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.224", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0d4e33a99c7741fb8cdd97f4ec5dbbd8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db68bd64-5b56-49af-a075-13dcf85cb2e0", "external-id": "nsx-vlan-transportzone-590", "segmentation_id": 590, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3d01745e-0d", "ovs_interfaceid": "3d01745e-0d9e-4bae-ba4d-242380edbab6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 901.211180] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cbf7a2c2-7fb7-405b-bf5c-0776ba26eb99 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Releasing lock "refresh_cache-9c989152-ec69-478e-a0b0-62983852576c" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 901.211668] env[62066]: DEBUG nova.compute.manager [None req-cbf7a2c2-7fb7-405b-bf5c-0776ba26eb99 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 9c989152-ec69-478e-a0b0-62983852576c] Instance network_info: |[{"id": "1c0d0075-7f22-441e-a283-21e9b5c5aecc", "address": "fa:16:3e:c5:c3:07", "network": {"id": "1c6c51e4-96f8-4175-b7f5-96b8a42fcb13", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-2006089809-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d609babc987a42e2a8ddb4bfb9c3b103", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a965790c-2d2f-4c2a-9ee7-745f4d53039b", "external-id": "nsx-vlan-transportzone-708", "segmentation_id": 708, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1c0d0075-7f", "ovs_interfaceid": "1c0d0075-7f22-441e-a283-21e9b5c5aecc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 901.212278] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-cbf7a2c2-7fb7-405b-bf5c-0776ba26eb99 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 9c989152-ec69-478e-a0b0-62983852576c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c5:c3:07', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a965790c-2d2f-4c2a-9ee7-745f4d53039b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1c0d0075-7f22-441e-a283-21e9b5c5aecc', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 901.222621] env[62066]: DEBUG oslo.service.loopingcall [None req-cbf7a2c2-7fb7-405b-bf5c-0776ba26eb99 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 901.222918] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9c989152-ec69-478e-a0b0-62983852576c] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 901.223241] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c028b68d-3d30-4f2d-be3a-8d594106cc3d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.245387] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 901.245387] env[62066]: value = "task-1341089" [ 901.245387] env[62066]: _type = "Task" [ 901.245387] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.254315] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1341089, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.309203] env[62066]: DEBUG oslo_concurrency.lockutils [None req-adc52c21-6d1b-4740-84d8-9cdb3fbc5d28 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.270s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 901.309774] env[62066]: DEBUG nova.compute.manager [None req-adc52c21-6d1b-4740-84d8-9cdb3fbc5d28 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: 90c33434-c127-450b-9ff0-75181b4ac385] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 901.312859] env[62066]: DEBUG oslo_concurrency.lockutils [None req-93fb3d16-c0a4-4bf0-b776-e2995ef54e48 tempest-ServerTagsTestJSON-2015212290 tempest-ServerTagsTestJSON-2015212290-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.612s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 901.314426] env[62066]: INFO nova.compute.claims [None req-93fb3d16-c0a4-4bf0-b776-e2995ef54e48 tempest-ServerTagsTestJSON-2015212290 tempest-ServerTagsTestJSON-2015212290-project-member] [instance: f4a84152-f222-46a2-9a35-8a0a0078fc12] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 901.400885] env[62066]: DEBUG oslo_concurrency.lockutils [req-c9c84ac9-d773-47ae-8202-e9d022cef8ff req-e14120d6-ac55-44cc-8858-b0fdcf03985c service nova] Releasing lock "refresh_cache-b718b09b-a7fa-445b-8be3-5b3ebca210a2" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 901.481751] env[62066]: DEBUG nova.compute.manager [req-37514658-a9a9-4fc1-8cb8-c8920b2e192c req-4e1a4acd-c60c-4d7c-9f2e-e862a11614a5 service nova] [instance: 9c989152-ec69-478e-a0b0-62983852576c] Received event network-changed-1c0d0075-7f22-441e-a283-21e9b5c5aecc {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 901.481923] env[62066]: DEBUG nova.compute.manager [req-37514658-a9a9-4fc1-8cb8-c8920b2e192c req-4e1a4acd-c60c-4d7c-9f2e-e862a11614a5 service nova] [instance: 9c989152-ec69-478e-a0b0-62983852576c] Refreshing instance network info cache due to event network-changed-1c0d0075-7f22-441e-a283-21e9b5c5aecc. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 901.482654] env[62066]: DEBUG oslo_concurrency.lockutils [req-37514658-a9a9-4fc1-8cb8-c8920b2e192c req-4e1a4acd-c60c-4d7c-9f2e-e862a11614a5 service nova] Acquiring lock "refresh_cache-9c989152-ec69-478e-a0b0-62983852576c" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 901.482834] env[62066]: DEBUG oslo_concurrency.lockutils [req-37514658-a9a9-4fc1-8cb8-c8920b2e192c req-4e1a4acd-c60c-4d7c-9f2e-e862a11614a5 service nova] Acquired lock "refresh_cache-9c989152-ec69-478e-a0b0-62983852576c" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 901.483016] env[62066]: DEBUG nova.network.neutron [req-37514658-a9a9-4fc1-8cb8-c8920b2e192c req-4e1a4acd-c60c-4d7c-9f2e-e862a11614a5 service nova] [instance: 9c989152-ec69-478e-a0b0-62983852576c] Refreshing network info cache for port 1c0d0075-7f22-441e-a283-21e9b5c5aecc {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 901.755126] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1341089, 'name': CreateVM_Task, 'duration_secs': 0.291151} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.755303] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9c989152-ec69-478e-a0b0-62983852576c] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 901.755959] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cbf7a2c2-7fb7-405b-bf5c-0776ba26eb99 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 901.756147] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cbf7a2c2-7fb7-405b-bf5c-0776ba26eb99 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Acquired lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 901.756477] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cbf7a2c2-7fb7-405b-bf5c-0776ba26eb99 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 901.756722] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f571b2eb-b9a7-4c4d-b593-268bd55e6119 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.760960] env[62066]: DEBUG oslo_vmware.api [None req-cbf7a2c2-7fb7-405b-bf5c-0776ba26eb99 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Waiting for the task: (returnval){ [ 901.760960] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]52af833c-695e-9925-6c0f-34fb23efe521" [ 901.760960] env[62066]: _type = "Task" [ 901.760960] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.768371] env[62066]: DEBUG oslo_vmware.api [None req-cbf7a2c2-7fb7-405b-bf5c-0776ba26eb99 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52af833c-695e-9925-6c0f-34fb23efe521, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.818915] env[62066]: DEBUG nova.compute.utils [None req-adc52c21-6d1b-4740-84d8-9cdb3fbc5d28 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 901.822245] env[62066]: DEBUG nova.compute.manager [None req-adc52c21-6d1b-4740-84d8-9cdb3fbc5d28 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: 90c33434-c127-450b-9ff0-75181b4ac385] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 901.822417] env[62066]: DEBUG nova.network.neutron [None req-adc52c21-6d1b-4740-84d8-9cdb3fbc5d28 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: 90c33434-c127-450b-9ff0-75181b4ac385] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 901.862896] env[62066]: DEBUG nova.policy [None req-adc52c21-6d1b-4740-84d8-9cdb3fbc5d28 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '588e53cee85f4ab484b76e7a59fcbe78', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8475ad5a900548cba568360999c846ea', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 902.155417] env[62066]: DEBUG nova.network.neutron [None req-adc52c21-6d1b-4740-84d8-9cdb3fbc5d28 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: 90c33434-c127-450b-9ff0-75181b4ac385] Successfully created port: 1b4bbccc-7650-4977-9133-da7bda63ceeb {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 902.204444] env[62066]: DEBUG nova.network.neutron [req-37514658-a9a9-4fc1-8cb8-c8920b2e192c req-4e1a4acd-c60c-4d7c-9f2e-e862a11614a5 service nova] [instance: 9c989152-ec69-478e-a0b0-62983852576c] Updated VIF entry in instance network info cache for port 1c0d0075-7f22-441e-a283-21e9b5c5aecc. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 902.204801] env[62066]: DEBUG nova.network.neutron [req-37514658-a9a9-4fc1-8cb8-c8920b2e192c req-4e1a4acd-c60c-4d7c-9f2e-e862a11614a5 service nova] [instance: 9c989152-ec69-478e-a0b0-62983852576c] Updating instance_info_cache with network_info: [{"id": "1c0d0075-7f22-441e-a283-21e9b5c5aecc", "address": "fa:16:3e:c5:c3:07", "network": {"id": "1c6c51e4-96f8-4175-b7f5-96b8a42fcb13", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-2006089809-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d609babc987a42e2a8ddb4bfb9c3b103", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a965790c-2d2f-4c2a-9ee7-745f4d53039b", "external-id": "nsx-vlan-transportzone-708", "segmentation_id": 708, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1c0d0075-7f", "ovs_interfaceid": "1c0d0075-7f22-441e-a283-21e9b5c5aecc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 902.271401] env[62066]: DEBUG oslo_vmware.api [None req-cbf7a2c2-7fb7-405b-bf5c-0776ba26eb99 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52af833c-695e-9925-6c0f-34fb23efe521, 'name': SearchDatastore_Task, 'duration_secs': 0.01476} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.271914] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cbf7a2c2-7fb7-405b-bf5c-0776ba26eb99 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Releasing lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 902.272085] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-cbf7a2c2-7fb7-405b-bf5c-0776ba26eb99 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 9c989152-ec69-478e-a0b0-62983852576c] Processing image 50ff584c-3b50-4395-af07-3e66769bc9f7 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 902.272208] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cbf7a2c2-7fb7-405b-bf5c-0776ba26eb99 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 902.272363] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cbf7a2c2-7fb7-405b-bf5c-0776ba26eb99 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Acquired lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 902.272547] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-cbf7a2c2-7fb7-405b-bf5c-0776ba26eb99 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 902.273534] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-bd2abf19-c164-4d4c-826e-aef641ada528 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.281347] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-cbf7a2c2-7fb7-405b-bf5c-0776ba26eb99 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 902.281347] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-cbf7a2c2-7fb7-405b-bf5c-0776ba26eb99 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 902.282063] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7e92f93c-dce0-4541-93fb-406a42cc633b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.287060] env[62066]: DEBUG oslo_vmware.api [None req-cbf7a2c2-7fb7-405b-bf5c-0776ba26eb99 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Waiting for the task: (returnval){ [ 902.287060] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]5223ec28-c882-ea36-eb48-84d227145294" [ 902.287060] env[62066]: _type = "Task" [ 902.287060] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.294328] env[62066]: DEBUG oslo_vmware.api [None req-cbf7a2c2-7fb7-405b-bf5c-0776ba26eb99 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]5223ec28-c882-ea36-eb48-84d227145294, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.323061] env[62066]: DEBUG nova.compute.manager [None req-adc52c21-6d1b-4740-84d8-9cdb3fbc5d28 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: 90c33434-c127-450b-9ff0-75181b4ac385] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 902.531972] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d331a49-453f-4e6a-b59e-320a06b283b4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.539277] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccc5d6c0-0878-479a-a5ae-c716b9b36596 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.569922] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8707ea3-5955-4d83-a224-7692684d8c2c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.577235] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-809aaa65-78c4-4f00-8a1b-0d89845a261e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.590311] env[62066]: DEBUG nova.compute.provider_tree [None req-93fb3d16-c0a4-4bf0-b776-e2995ef54e48 tempest-ServerTagsTestJSON-2015212290 tempest-ServerTagsTestJSON-2015212290-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 902.707506] env[62066]: DEBUG oslo_concurrency.lockutils [req-37514658-a9a9-4fc1-8cb8-c8920b2e192c req-4e1a4acd-c60c-4d7c-9f2e-e862a11614a5 service nova] Releasing lock "refresh_cache-9c989152-ec69-478e-a0b0-62983852576c" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 902.796898] env[62066]: DEBUG oslo_vmware.api [None req-cbf7a2c2-7fb7-405b-bf5c-0776ba26eb99 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]5223ec28-c882-ea36-eb48-84d227145294, 'name': SearchDatastore_Task, 'duration_secs': 0.011436} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.797686] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e6802957-117c-4f6c-9eb8-b29a5d71659b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.802569] env[62066]: DEBUG oslo_vmware.api [None req-cbf7a2c2-7fb7-405b-bf5c-0776ba26eb99 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Waiting for the task: (returnval){ [ 902.802569] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]52b107b2-7ba2-e469-2e06-3343d0bb68ae" [ 902.802569] env[62066]: _type = "Task" [ 902.802569] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.810276] env[62066]: DEBUG oslo_vmware.api [None req-cbf7a2c2-7fb7-405b-bf5c-0776ba26eb99 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52b107b2-7ba2-e469-2e06-3343d0bb68ae, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.094226] env[62066]: DEBUG nova.scheduler.client.report [None req-93fb3d16-c0a4-4bf0-b776-e2995ef54e48 tempest-ServerTagsTestJSON-2015212290 tempest-ServerTagsTestJSON-2015212290-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 903.137345] env[62066]: DEBUG oslo_service.periodic_task [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 903.137561] env[62066]: DEBUG oslo_service.periodic_task [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 903.137706] env[62066]: DEBUG nova.compute.manager [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Starting heal instance info cache {{(pid=62066) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 903.312270] env[62066]: DEBUG oslo_vmware.api [None req-cbf7a2c2-7fb7-405b-bf5c-0776ba26eb99 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52b107b2-7ba2-e469-2e06-3343d0bb68ae, 'name': SearchDatastore_Task, 'duration_secs': 0.009022} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.312518] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cbf7a2c2-7fb7-405b-bf5c-0776ba26eb99 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Releasing lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 903.312768] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-cbf7a2c2-7fb7-405b-bf5c-0776ba26eb99 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk to [datastore2] 9c989152-ec69-478e-a0b0-62983852576c/9c989152-ec69-478e-a0b0-62983852576c.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 903.313041] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9b4cb1dd-df67-4729-bfe8-60ffa756cdfd {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.319081] env[62066]: DEBUG oslo_vmware.api [None req-cbf7a2c2-7fb7-405b-bf5c-0776ba26eb99 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Waiting for the task: (returnval){ [ 903.319081] env[62066]: value = "task-1341090" [ 903.319081] env[62066]: _type = "Task" [ 903.319081] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.326762] env[62066]: DEBUG oslo_vmware.api [None req-cbf7a2c2-7fb7-405b-bf5c-0776ba26eb99 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1341090, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.333898] env[62066]: DEBUG nova.compute.manager [None req-adc52c21-6d1b-4740-84d8-9cdb3fbc5d28 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: 90c33434-c127-450b-9ff0-75181b4ac385] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 903.368237] env[62066]: DEBUG nova.virt.hardware [None req-adc52c21-6d1b-4740-84d8-9cdb3fbc5d28 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-23T13:40:41Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-23T13:40:23Z,direct_url=,disk_format='vmdk',id=50ff584c-3b50-4395-af07-3e66769bc9f7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='eb52f7069a374c61ae946f052007c6d9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-23T13:40:24Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 903.368493] env[62066]: DEBUG nova.virt.hardware [None req-adc52c21-6d1b-4740-84d8-9cdb3fbc5d28 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 903.368684] env[62066]: DEBUG nova.virt.hardware [None req-adc52c21-6d1b-4740-84d8-9cdb3fbc5d28 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 903.368874] env[62066]: DEBUG nova.virt.hardware [None req-adc52c21-6d1b-4740-84d8-9cdb3fbc5d28 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 903.369040] env[62066]: DEBUG nova.virt.hardware [None req-adc52c21-6d1b-4740-84d8-9cdb3fbc5d28 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 903.369206] env[62066]: DEBUG nova.virt.hardware [None req-adc52c21-6d1b-4740-84d8-9cdb3fbc5d28 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 903.369422] env[62066]: DEBUG nova.virt.hardware [None req-adc52c21-6d1b-4740-84d8-9cdb3fbc5d28 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 903.369586] env[62066]: DEBUG nova.virt.hardware [None req-adc52c21-6d1b-4740-84d8-9cdb3fbc5d28 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 903.369759] env[62066]: DEBUG nova.virt.hardware [None req-adc52c21-6d1b-4740-84d8-9cdb3fbc5d28 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 903.369982] env[62066]: DEBUG nova.virt.hardware [None req-adc52c21-6d1b-4740-84d8-9cdb3fbc5d28 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 903.370182] env[62066]: DEBUG nova.virt.hardware [None req-adc52c21-6d1b-4740-84d8-9cdb3fbc5d28 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 903.371055] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5e4e50d-4c83-445f-9b28-97aa4df996d6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.378688] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a896246d-104e-4549-a45f-993b3b2210e7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.599458] env[62066]: DEBUG oslo_concurrency.lockutils [None req-93fb3d16-c0a4-4bf0-b776-e2995ef54e48 tempest-ServerTagsTestJSON-2015212290 tempest-ServerTagsTestJSON-2015212290-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.286s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 903.600222] env[62066]: DEBUG nova.compute.manager [None req-93fb3d16-c0a4-4bf0-b776-e2995ef54e48 tempest-ServerTagsTestJSON-2015212290 tempest-ServerTagsTestJSON-2015212290-project-member] [instance: f4a84152-f222-46a2-9a35-8a0a0078fc12] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 903.604375] env[62066]: DEBUG oslo_concurrency.lockutils [None req-36f2458b-83e5-4859-9178-b3e847334e66 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.084s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 903.606195] env[62066]: INFO nova.compute.claims [None req-36f2458b-83e5-4859-9178-b3e847334e66 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 4e39b9b0-4ab0-4f48-8a3f-6bc1895597eb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 903.777967] env[62066]: DEBUG nova.compute.manager [req-51092456-d4c2-4234-8caa-bcdebfa603ac req-9b36e184-261f-4955-8b87-cab520ea50ff service nova] [instance: 90c33434-c127-450b-9ff0-75181b4ac385] Received event network-vif-plugged-1b4bbccc-7650-4977-9133-da7bda63ceeb {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 903.778348] env[62066]: DEBUG oslo_concurrency.lockutils [req-51092456-d4c2-4234-8caa-bcdebfa603ac req-9b36e184-261f-4955-8b87-cab520ea50ff service nova] Acquiring lock "90c33434-c127-450b-9ff0-75181b4ac385-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 903.778698] env[62066]: DEBUG oslo_concurrency.lockutils [req-51092456-d4c2-4234-8caa-bcdebfa603ac req-9b36e184-261f-4955-8b87-cab520ea50ff service nova] Lock "90c33434-c127-450b-9ff0-75181b4ac385-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 903.778993] env[62066]: DEBUG oslo_concurrency.lockutils [req-51092456-d4c2-4234-8caa-bcdebfa603ac req-9b36e184-261f-4955-8b87-cab520ea50ff service nova] Lock "90c33434-c127-450b-9ff0-75181b4ac385-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 903.779684] env[62066]: DEBUG nova.compute.manager [req-51092456-d4c2-4234-8caa-bcdebfa603ac req-9b36e184-261f-4955-8b87-cab520ea50ff service nova] [instance: 90c33434-c127-450b-9ff0-75181b4ac385] No waiting events found dispatching network-vif-plugged-1b4bbccc-7650-4977-9133-da7bda63ceeb {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 903.780029] env[62066]: WARNING nova.compute.manager [req-51092456-d4c2-4234-8caa-bcdebfa603ac req-9b36e184-261f-4955-8b87-cab520ea50ff service nova] [instance: 90c33434-c127-450b-9ff0-75181b4ac385] Received unexpected event network-vif-plugged-1b4bbccc-7650-4977-9133-da7bda63ceeb for instance with vm_state building and task_state spawning. [ 903.828472] env[62066]: DEBUG oslo_vmware.api [None req-cbf7a2c2-7fb7-405b-bf5c-0776ba26eb99 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1341090, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.465942} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.828745] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-cbf7a2c2-7fb7-405b-bf5c-0776ba26eb99 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk to [datastore2] 9c989152-ec69-478e-a0b0-62983852576c/9c989152-ec69-478e-a0b0-62983852576c.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 903.828963] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-cbf7a2c2-7fb7-405b-bf5c-0776ba26eb99 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 9c989152-ec69-478e-a0b0-62983852576c] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 903.829259] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ed1573bb-815c-40f9-82e3-76ffcfdcbea2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.835207] env[62066]: DEBUG oslo_vmware.api [None req-cbf7a2c2-7fb7-405b-bf5c-0776ba26eb99 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Waiting for the task: (returnval){ [ 903.835207] env[62066]: value = "task-1341091" [ 903.835207] env[62066]: _type = "Task" [ 903.835207] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.843367] env[62066]: DEBUG oslo_vmware.api [None req-cbf7a2c2-7fb7-405b-bf5c-0776ba26eb99 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1341091, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.870384] env[62066]: DEBUG nova.network.neutron [None req-adc52c21-6d1b-4740-84d8-9cdb3fbc5d28 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: 90c33434-c127-450b-9ff0-75181b4ac385] Successfully updated port: 1b4bbccc-7650-4977-9133-da7bda63ceeb {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 904.106216] env[62066]: DEBUG nova.compute.utils [None req-93fb3d16-c0a4-4bf0-b776-e2995ef54e48 tempest-ServerTagsTestJSON-2015212290 tempest-ServerTagsTestJSON-2015212290-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 904.107614] env[62066]: DEBUG nova.compute.manager [None req-93fb3d16-c0a4-4bf0-b776-e2995ef54e48 tempest-ServerTagsTestJSON-2015212290 tempest-ServerTagsTestJSON-2015212290-project-member] [instance: f4a84152-f222-46a2-9a35-8a0a0078fc12] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 904.107798] env[62066]: DEBUG nova.network.neutron [None req-93fb3d16-c0a4-4bf0-b776-e2995ef54e48 tempest-ServerTagsTestJSON-2015212290 tempest-ServerTagsTestJSON-2015212290-project-member] [instance: f4a84152-f222-46a2-9a35-8a0a0078fc12] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 904.148239] env[62066]: DEBUG nova.policy [None req-93fb3d16-c0a4-4bf0-b776-e2995ef54e48 tempest-ServerTagsTestJSON-2015212290 tempest-ServerTagsTestJSON-2015212290-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '791ae284a03e429d8ab8faed2b006f84', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'abb1fa60628341849e205d03e1f7e8ed', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 904.344190] env[62066]: DEBUG oslo_vmware.api [None req-cbf7a2c2-7fb7-405b-bf5c-0776ba26eb99 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1341091, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.057695} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.344494] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-cbf7a2c2-7fb7-405b-bf5c-0776ba26eb99 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 9c989152-ec69-478e-a0b0-62983852576c] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 904.345250] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-934957af-2be1-4a31-9657-09d39208c610 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.369616] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-cbf7a2c2-7fb7-405b-bf5c-0776ba26eb99 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 9c989152-ec69-478e-a0b0-62983852576c] Reconfiguring VM instance instance-00000056 to attach disk [datastore2] 9c989152-ec69-478e-a0b0-62983852576c/9c989152-ec69-478e-a0b0-62983852576c.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 904.370303] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3f9796a1-f1e5-4ad9-a967-fd712731c61e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.386385] env[62066]: DEBUG oslo_concurrency.lockutils [None req-adc52c21-6d1b-4740-84d8-9cdb3fbc5d28 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Acquiring lock "refresh_cache-90c33434-c127-450b-9ff0-75181b4ac385" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 904.386530] env[62066]: DEBUG oslo_concurrency.lockutils [None req-adc52c21-6d1b-4740-84d8-9cdb3fbc5d28 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Acquired lock "refresh_cache-90c33434-c127-450b-9ff0-75181b4ac385" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 904.386675] env[62066]: DEBUG nova.network.neutron [None req-adc52c21-6d1b-4740-84d8-9cdb3fbc5d28 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: 90c33434-c127-450b-9ff0-75181b4ac385] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 904.393459] env[62066]: DEBUG oslo_vmware.api [None req-cbf7a2c2-7fb7-405b-bf5c-0776ba26eb99 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Waiting for the task: (returnval){ [ 904.393459] env[62066]: value = "task-1341092" [ 904.393459] env[62066]: _type = "Task" [ 904.393459] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.404534] env[62066]: DEBUG oslo_vmware.api [None req-cbf7a2c2-7fb7-405b-bf5c-0776ba26eb99 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1341092, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.464389] env[62066]: DEBUG nova.network.neutron [None req-93fb3d16-c0a4-4bf0-b776-e2995ef54e48 tempest-ServerTagsTestJSON-2015212290 tempest-ServerTagsTestJSON-2015212290-project-member] [instance: f4a84152-f222-46a2-9a35-8a0a0078fc12] Successfully created port: 35ae652b-5fe5-468b-b4d4-75e727382145 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 904.611364] env[62066]: DEBUG nova.compute.manager [None req-93fb3d16-c0a4-4bf0-b776-e2995ef54e48 tempest-ServerTagsTestJSON-2015212290 tempest-ServerTagsTestJSON-2015212290-project-member] [instance: f4a84152-f222-46a2-9a35-8a0a0078fc12] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 904.675161] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Acquiring lock "refresh_cache-02fb3e92-5dd6-4b1e-a6e4-d60d3fc85b07" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 904.675315] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Acquired lock "refresh_cache-02fb3e92-5dd6-4b1e-a6e4-d60d3fc85b07" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 904.675463] env[62066]: DEBUG nova.network.neutron [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] [instance: 02fb3e92-5dd6-4b1e-a6e4-d60d3fc85b07] Forcefully refreshing network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 904.821870] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f534c36-f08e-483f-81ff-b61b7af6a0e1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.829951] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d6e872a-f515-4c01-b22a-f1bf8a62093d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.861076] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28b02b43-5b32-46cc-9adc-c318d3de9c9c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.868912] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac637c85-bbab-4532-93ec-3190b2ca822b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.882963] env[62066]: DEBUG nova.compute.provider_tree [None req-36f2458b-83e5-4859-9178-b3e847334e66 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 904.903632] env[62066]: DEBUG oslo_vmware.api [None req-cbf7a2c2-7fb7-405b-bf5c-0776ba26eb99 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1341092, 'name': ReconfigVM_Task, 'duration_secs': 0.279192} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.903904] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-cbf7a2c2-7fb7-405b-bf5c-0776ba26eb99 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 9c989152-ec69-478e-a0b0-62983852576c] Reconfigured VM instance instance-00000056 to attach disk [datastore2] 9c989152-ec69-478e-a0b0-62983852576c/9c989152-ec69-478e-a0b0-62983852576c.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 904.904563] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e33417c8-4bab-460e-bd33-19ec8d39e20b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.910757] env[62066]: DEBUG oslo_vmware.api [None req-cbf7a2c2-7fb7-405b-bf5c-0776ba26eb99 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Waiting for the task: (returnval){ [ 904.910757] env[62066]: value = "task-1341093" [ 904.910757] env[62066]: _type = "Task" [ 904.910757] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.919164] env[62066]: DEBUG oslo_vmware.api [None req-cbf7a2c2-7fb7-405b-bf5c-0776ba26eb99 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1341093, 'name': Rename_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.926781] env[62066]: DEBUG nova.network.neutron [None req-adc52c21-6d1b-4740-84d8-9cdb3fbc5d28 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: 90c33434-c127-450b-9ff0-75181b4ac385] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 905.084167] env[62066]: DEBUG nova.network.neutron [None req-adc52c21-6d1b-4740-84d8-9cdb3fbc5d28 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: 90c33434-c127-450b-9ff0-75181b4ac385] Updating instance_info_cache with network_info: [{"id": "1b4bbccc-7650-4977-9133-da7bda63ceeb", "address": "fa:16:3e:22:10:c6", "network": {"id": "7617f887-bdd0-400a-8b5f-606d7785df78", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-558766873-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8475ad5a900548cba568360999c846ea", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2b6a4065-12af-4fb9-ac47-ec9143f7297e", "external-id": "nsx-vlan-transportzone-95", "segmentation_id": 95, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1b4bbccc-76", "ovs_interfaceid": "1b4bbccc-7650-4977-9133-da7bda63ceeb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 905.387463] env[62066]: DEBUG nova.scheduler.client.report [None req-36f2458b-83e5-4859-9178-b3e847334e66 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 905.420797] env[62066]: DEBUG oslo_vmware.api [None req-cbf7a2c2-7fb7-405b-bf5c-0776ba26eb99 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1341093, 'name': Rename_Task, 'duration_secs': 0.14665} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.421141] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-cbf7a2c2-7fb7-405b-bf5c-0776ba26eb99 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 9c989152-ec69-478e-a0b0-62983852576c] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 905.421401] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1b90df5f-e0ba-4a51-a365-020b6673a75f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.427222] env[62066]: DEBUG oslo_vmware.api [None req-cbf7a2c2-7fb7-405b-bf5c-0776ba26eb99 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Waiting for the task: (returnval){ [ 905.427222] env[62066]: value = "task-1341094" [ 905.427222] env[62066]: _type = "Task" [ 905.427222] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.434415] env[62066]: DEBUG oslo_vmware.api [None req-cbf7a2c2-7fb7-405b-bf5c-0776ba26eb99 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1341094, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.585054] env[62066]: DEBUG oslo_concurrency.lockutils [None req-adc52c21-6d1b-4740-84d8-9cdb3fbc5d28 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Releasing lock "refresh_cache-90c33434-c127-450b-9ff0-75181b4ac385" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 905.585966] env[62066]: DEBUG nova.compute.manager [None req-adc52c21-6d1b-4740-84d8-9cdb3fbc5d28 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: 90c33434-c127-450b-9ff0-75181b4ac385] Instance network_info: |[{"id": "1b4bbccc-7650-4977-9133-da7bda63ceeb", "address": "fa:16:3e:22:10:c6", "network": {"id": "7617f887-bdd0-400a-8b5f-606d7785df78", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-558766873-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8475ad5a900548cba568360999c846ea", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2b6a4065-12af-4fb9-ac47-ec9143f7297e", "external-id": "nsx-vlan-transportzone-95", "segmentation_id": 95, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1b4bbccc-76", "ovs_interfaceid": "1b4bbccc-7650-4977-9133-da7bda63ceeb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 905.585966] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-adc52c21-6d1b-4740-84d8-9cdb3fbc5d28 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: 90c33434-c127-450b-9ff0-75181b4ac385] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:22:10:c6', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '2b6a4065-12af-4fb9-ac47-ec9143f7297e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1b4bbccc-7650-4977-9133-da7bda63ceeb', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 905.593501] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-adc52c21-6d1b-4740-84d8-9cdb3fbc5d28 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Creating folder: Project (8475ad5a900548cba568360999c846ea). Parent ref: group-v285980. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 905.593823] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8d830269-21a8-46fe-8c30-89e202364537 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.605117] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-adc52c21-6d1b-4740-84d8-9cdb3fbc5d28 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Created folder: Project (8475ad5a900548cba568360999c846ea) in parent group-v285980. [ 905.605317] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-adc52c21-6d1b-4740-84d8-9cdb3fbc5d28 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Creating folder: Instances. Parent ref: group-v286092. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 905.605588] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b07d4704-863f-4bc3-a74b-907c994be6b6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.619042] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-adc52c21-6d1b-4740-84d8-9cdb3fbc5d28 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Created folder: Instances in parent group-v286092. [ 905.619205] env[62066]: DEBUG oslo.service.loopingcall [None req-adc52c21-6d1b-4740-84d8-9cdb3fbc5d28 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 905.619490] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 90c33434-c127-450b-9ff0-75181b4ac385] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 905.619661] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-dd02292d-9dae-42b9-a4ad-e05323e3b143 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.635418] env[62066]: DEBUG nova.compute.manager [None req-93fb3d16-c0a4-4bf0-b776-e2995ef54e48 tempest-ServerTagsTestJSON-2015212290 tempest-ServerTagsTestJSON-2015212290-project-member] [instance: f4a84152-f222-46a2-9a35-8a0a0078fc12] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 905.642152] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 905.642152] env[62066]: value = "task-1341097" [ 905.642152] env[62066]: _type = "Task" [ 905.642152] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.651495] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1341097, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.663759] env[62066]: DEBUG nova.virt.hardware [None req-93fb3d16-c0a4-4bf0-b776-e2995ef54e48 tempest-ServerTagsTestJSON-2015212290 tempest-ServerTagsTestJSON-2015212290-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-23T13:40:41Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-23T13:40:23Z,direct_url=,disk_format='vmdk',id=50ff584c-3b50-4395-af07-3e66769bc9f7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='eb52f7069a374c61ae946f052007c6d9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-23T13:40:24Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 905.664036] env[62066]: DEBUG nova.virt.hardware [None req-93fb3d16-c0a4-4bf0-b776-e2995ef54e48 tempest-ServerTagsTestJSON-2015212290 tempest-ServerTagsTestJSON-2015212290-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 905.664188] env[62066]: DEBUG nova.virt.hardware [None req-93fb3d16-c0a4-4bf0-b776-e2995ef54e48 tempest-ServerTagsTestJSON-2015212290 tempest-ServerTagsTestJSON-2015212290-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 905.664378] env[62066]: DEBUG nova.virt.hardware [None req-93fb3d16-c0a4-4bf0-b776-e2995ef54e48 tempest-ServerTagsTestJSON-2015212290 tempest-ServerTagsTestJSON-2015212290-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 905.664529] env[62066]: DEBUG nova.virt.hardware [None req-93fb3d16-c0a4-4bf0-b776-e2995ef54e48 tempest-ServerTagsTestJSON-2015212290 tempest-ServerTagsTestJSON-2015212290-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 905.664683] env[62066]: DEBUG nova.virt.hardware [None req-93fb3d16-c0a4-4bf0-b776-e2995ef54e48 tempest-ServerTagsTestJSON-2015212290 tempest-ServerTagsTestJSON-2015212290-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 905.664901] env[62066]: DEBUG nova.virt.hardware [None req-93fb3d16-c0a4-4bf0-b776-e2995ef54e48 tempest-ServerTagsTestJSON-2015212290 tempest-ServerTagsTestJSON-2015212290-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 905.665880] env[62066]: DEBUG nova.virt.hardware [None req-93fb3d16-c0a4-4bf0-b776-e2995ef54e48 tempest-ServerTagsTestJSON-2015212290 tempest-ServerTagsTestJSON-2015212290-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 905.666149] env[62066]: DEBUG nova.virt.hardware [None req-93fb3d16-c0a4-4bf0-b776-e2995ef54e48 tempest-ServerTagsTestJSON-2015212290 tempest-ServerTagsTestJSON-2015212290-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 905.666514] env[62066]: DEBUG nova.virt.hardware [None req-93fb3d16-c0a4-4bf0-b776-e2995ef54e48 tempest-ServerTagsTestJSON-2015212290 tempest-ServerTagsTestJSON-2015212290-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 905.666742] env[62066]: DEBUG nova.virt.hardware [None req-93fb3d16-c0a4-4bf0-b776-e2995ef54e48 tempest-ServerTagsTestJSON-2015212290 tempest-ServerTagsTestJSON-2015212290-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 905.668429] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99851b49-a44b-4503-8414-befb09bc9fbd {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.677701] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da7c3f64-b956-46a9-b7c9-ec8944dc2f1e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.803430] env[62066]: DEBUG nova.compute.manager [req-af934cf7-19c2-4070-94a5-70a3103dbcf2 req-fa079147-c851-488c-9d30-b08a959bc9a6 service nova] [instance: 90c33434-c127-450b-9ff0-75181b4ac385] Received event network-changed-1b4bbccc-7650-4977-9133-da7bda63ceeb {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 905.803663] env[62066]: DEBUG nova.compute.manager [req-af934cf7-19c2-4070-94a5-70a3103dbcf2 req-fa079147-c851-488c-9d30-b08a959bc9a6 service nova] [instance: 90c33434-c127-450b-9ff0-75181b4ac385] Refreshing instance network info cache due to event network-changed-1b4bbccc-7650-4977-9133-da7bda63ceeb. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 905.804029] env[62066]: DEBUG oslo_concurrency.lockutils [req-af934cf7-19c2-4070-94a5-70a3103dbcf2 req-fa079147-c851-488c-9d30-b08a959bc9a6 service nova] Acquiring lock "refresh_cache-90c33434-c127-450b-9ff0-75181b4ac385" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 905.804078] env[62066]: DEBUG oslo_concurrency.lockutils [req-af934cf7-19c2-4070-94a5-70a3103dbcf2 req-fa079147-c851-488c-9d30-b08a959bc9a6 service nova] Acquired lock "refresh_cache-90c33434-c127-450b-9ff0-75181b4ac385" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 905.804261] env[62066]: DEBUG nova.network.neutron [req-af934cf7-19c2-4070-94a5-70a3103dbcf2 req-fa079147-c851-488c-9d30-b08a959bc9a6 service nova] [instance: 90c33434-c127-450b-9ff0-75181b4ac385] Refreshing network info cache for port 1b4bbccc-7650-4977-9133-da7bda63ceeb {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 905.886138] env[62066]: DEBUG nova.network.neutron [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] [instance: 02fb3e92-5dd6-4b1e-a6e4-d60d3fc85b07] Updating instance_info_cache with network_info: [{"id": "16bb647a-353c-4bf8-832d-8f6e49a6ce4b", "address": "fa:16:3e:47:44:80", "network": {"id": "9f3396ce-ee48-4eaf-8cc7-3e1711f37537", "bridge": "br-int", "label": "tempest-ServersTestJSON-2129775093-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6497ab02f327476d8ff81c2ecc0371e5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7ab8d568-adb0-4f3b-b6cc-68413e6546ae", "external-id": "nsx-vlan-transportzone-86", "segmentation_id": 86, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap16bb647a-35", "ovs_interfaceid": "16bb647a-353c-4bf8-832d-8f6e49a6ce4b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 905.892847] env[62066]: DEBUG oslo_concurrency.lockutils [None req-36f2458b-83e5-4859-9178-b3e847334e66 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.288s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 905.893025] env[62066]: DEBUG nova.compute.manager [None req-36f2458b-83e5-4859-9178-b3e847334e66 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 4e39b9b0-4ab0-4f48-8a3f-6bc1895597eb] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 905.896194] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a96d87c9-14b1-48f6-a5e4-af778c649411 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 11.587s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 905.896420] env[62066]: DEBUG nova.objects.instance [None req-a96d87c9-14b1-48f6-a5e4-af778c649411 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Lazy-loading 'resources' on Instance uuid de2e2be7-efdb-45a8-842a-640ab9deb1d9 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 905.938419] env[62066]: DEBUG oslo_vmware.api [None req-cbf7a2c2-7fb7-405b-bf5c-0776ba26eb99 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1341094, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.151712] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1341097, 'name': CreateVM_Task} progress is 25%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.388857] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Releasing lock "refresh_cache-02fb3e92-5dd6-4b1e-a6e4-d60d3fc85b07" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 906.389262] env[62066]: DEBUG nova.compute.manager [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] [instance: 02fb3e92-5dd6-4b1e-a6e4-d60d3fc85b07] Updated the network info_cache for instance {{(pid=62066) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 906.389262] env[62066]: DEBUG oslo_service.periodic_task [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 906.389766] env[62066]: DEBUG oslo_service.periodic_task [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 906.389766] env[62066]: DEBUG oslo_service.periodic_task [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 906.390037] env[62066]: DEBUG oslo_service.periodic_task [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 906.390251] env[62066]: DEBUG oslo_service.periodic_task [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 906.390402] env[62066]: DEBUG oslo_service.periodic_task [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 906.390533] env[62066]: DEBUG nova.compute.manager [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62066) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 906.390681] env[62066]: DEBUG oslo_service.periodic_task [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Running periodic task ComputeManager.update_available_resource {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 906.398995] env[62066]: DEBUG nova.compute.utils [None req-36f2458b-83e5-4859-9178-b3e847334e66 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 906.403575] env[62066]: DEBUG nova.compute.manager [None req-36f2458b-83e5-4859-9178-b3e847334e66 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 4e39b9b0-4ab0-4f48-8a3f-6bc1895597eb] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 906.403806] env[62066]: DEBUG nova.network.neutron [None req-36f2458b-83e5-4859-9178-b3e847334e66 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 4e39b9b0-4ab0-4f48-8a3f-6bc1895597eb] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 906.442880] env[62066]: DEBUG oslo_vmware.api [None req-cbf7a2c2-7fb7-405b-bf5c-0776ba26eb99 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1341094, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.470396] env[62066]: DEBUG nova.policy [None req-36f2458b-83e5-4859-9178-b3e847334e66 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8910a229218b4ec5ad72c893badfc598', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6497ab02f327476d8ff81c2ecc0371e5', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 906.599010] env[62066]: DEBUG nova.network.neutron [req-af934cf7-19c2-4070-94a5-70a3103dbcf2 req-fa079147-c851-488c-9d30-b08a959bc9a6 service nova] [instance: 90c33434-c127-450b-9ff0-75181b4ac385] Updated VIF entry in instance network info cache for port 1b4bbccc-7650-4977-9133-da7bda63ceeb. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 906.599393] env[62066]: DEBUG nova.network.neutron [req-af934cf7-19c2-4070-94a5-70a3103dbcf2 req-fa079147-c851-488c-9d30-b08a959bc9a6 service nova] [instance: 90c33434-c127-450b-9ff0-75181b4ac385] Updating instance_info_cache with network_info: [{"id": "1b4bbccc-7650-4977-9133-da7bda63ceeb", "address": "fa:16:3e:22:10:c6", "network": {"id": "7617f887-bdd0-400a-8b5f-606d7785df78", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-558766873-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8475ad5a900548cba568360999c846ea", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2b6a4065-12af-4fb9-ac47-ec9143f7297e", "external-id": "nsx-vlan-transportzone-95", "segmentation_id": 95, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1b4bbccc-76", "ovs_interfaceid": "1b4bbccc-7650-4977-9133-da7bda63ceeb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 906.623270] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20075dc2-f7ec-4ee9-a306-1ca5c12f2bf3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.637817] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53e6f390-cac9-42d7-aa15-20e19f1c676e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.651067] env[62066]: DEBUG nova.network.neutron [None req-93fb3d16-c0a4-4bf0-b776-e2995ef54e48 tempest-ServerTagsTestJSON-2015212290 tempest-ServerTagsTestJSON-2015212290-project-member] [instance: f4a84152-f222-46a2-9a35-8a0a0078fc12] Successfully updated port: 35ae652b-5fe5-468b-b4d4-75e727382145 {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 906.684509] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50b88e08-a41a-45cf-81c8-963f0c1f43b6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.690250] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1341097, 'name': CreateVM_Task, 'duration_secs': 0.62436} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.692308] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 90c33434-c127-450b-9ff0-75181b4ac385] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 906.692956] env[62066]: DEBUG oslo_concurrency.lockutils [None req-adc52c21-6d1b-4740-84d8-9cdb3fbc5d28 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 906.693299] env[62066]: DEBUG oslo_concurrency.lockutils [None req-adc52c21-6d1b-4740-84d8-9cdb3fbc5d28 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Acquired lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 906.693487] env[62066]: DEBUG oslo_concurrency.lockutils [None req-adc52c21-6d1b-4740-84d8-9cdb3fbc5d28 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 906.695732] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b68bb2eb-0d29-4ed6-9096-6677e3617cea {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.699283] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25b697ef-8898-4ba3-9c79-c595a4a8411c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.707747] env[62066]: DEBUG oslo_vmware.api [None req-adc52c21-6d1b-4740-84d8-9cdb3fbc5d28 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Waiting for the task: (returnval){ [ 906.707747] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]5268e452-498e-9917-4f1c-6878e7e20b5e" [ 906.707747] env[62066]: _type = "Task" [ 906.707747] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.717239] env[62066]: DEBUG nova.compute.provider_tree [None req-a96d87c9-14b1-48f6-a5e4-af778c649411 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 906.726303] env[62066]: DEBUG oslo_vmware.api [None req-adc52c21-6d1b-4740-84d8-9cdb3fbc5d28 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]5268e452-498e-9917-4f1c-6878e7e20b5e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.808983] env[62066]: DEBUG nova.network.neutron [None req-36f2458b-83e5-4859-9178-b3e847334e66 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 4e39b9b0-4ab0-4f48-8a3f-6bc1895597eb] Successfully created port: 843a9051-1a6f-42b2-b482-2baf7ffb6ca4 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 906.844248] env[62066]: DEBUG oslo_concurrency.lockutils [None req-dc7187c9-b41f-4691-8bae-33272b590d9f tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Acquiring lock "c8e0d47c-4421-4e00-9183-206fceeabc40" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 906.845088] env[62066]: DEBUG oslo_concurrency.lockutils [None req-dc7187c9-b41f-4691-8bae-33272b590d9f tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Lock "c8e0d47c-4421-4e00-9183-206fceeabc40" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 906.894573] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 906.907042] env[62066]: DEBUG nova.compute.manager [None req-36f2458b-83e5-4859-9178-b3e847334e66 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 4e39b9b0-4ab0-4f48-8a3f-6bc1895597eb] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 906.941509] env[62066]: DEBUG oslo_vmware.api [None req-cbf7a2c2-7fb7-405b-bf5c-0776ba26eb99 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1341094, 'name': PowerOnVM_Task, 'duration_secs': 1.022604} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.941781] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-cbf7a2c2-7fb7-405b-bf5c-0776ba26eb99 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 9c989152-ec69-478e-a0b0-62983852576c] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 906.941986] env[62066]: INFO nova.compute.manager [None req-cbf7a2c2-7fb7-405b-bf5c-0776ba26eb99 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 9c989152-ec69-478e-a0b0-62983852576c] Took 7.68 seconds to spawn the instance on the hypervisor. [ 906.942279] env[62066]: DEBUG nova.compute.manager [None req-cbf7a2c2-7fb7-405b-bf5c-0776ba26eb99 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 9c989152-ec69-478e-a0b0-62983852576c] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 906.942966] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66b78ce7-cfb1-4274-b18b-98d514dd7fe8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.102665] env[62066]: DEBUG oslo_concurrency.lockutils [req-af934cf7-19c2-4070-94a5-70a3103dbcf2 req-fa079147-c851-488c-9d30-b08a959bc9a6 service nova] Releasing lock "refresh_cache-90c33434-c127-450b-9ff0-75181b4ac385" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 907.156067] env[62066]: DEBUG oslo_concurrency.lockutils [None req-93fb3d16-c0a4-4bf0-b776-e2995ef54e48 tempest-ServerTagsTestJSON-2015212290 tempest-ServerTagsTestJSON-2015212290-project-member] Acquiring lock "refresh_cache-f4a84152-f222-46a2-9a35-8a0a0078fc12" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 907.156067] env[62066]: DEBUG oslo_concurrency.lockutils [None req-93fb3d16-c0a4-4bf0-b776-e2995ef54e48 tempest-ServerTagsTestJSON-2015212290 tempest-ServerTagsTestJSON-2015212290-project-member] Acquired lock "refresh_cache-f4a84152-f222-46a2-9a35-8a0a0078fc12" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 907.156067] env[62066]: DEBUG nova.network.neutron [None req-93fb3d16-c0a4-4bf0-b776-e2995ef54e48 tempest-ServerTagsTestJSON-2015212290 tempest-ServerTagsTestJSON-2015212290-project-member] [instance: f4a84152-f222-46a2-9a35-8a0a0078fc12] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 907.227455] env[62066]: DEBUG nova.scheduler.client.report [None req-a96d87c9-14b1-48f6-a5e4-af778c649411 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 907.230509] env[62066]: DEBUG oslo_vmware.api [None req-adc52c21-6d1b-4740-84d8-9cdb3fbc5d28 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]5268e452-498e-9917-4f1c-6878e7e20b5e, 'name': SearchDatastore_Task, 'duration_secs': 0.020459} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.231209] env[62066]: DEBUG oslo_concurrency.lockutils [None req-adc52c21-6d1b-4740-84d8-9cdb3fbc5d28 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Releasing lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 907.231437] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-adc52c21-6d1b-4740-84d8-9cdb3fbc5d28 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: 90c33434-c127-450b-9ff0-75181b4ac385] Processing image 50ff584c-3b50-4395-af07-3e66769bc9f7 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 907.231666] env[62066]: DEBUG oslo_concurrency.lockutils [None req-adc52c21-6d1b-4740-84d8-9cdb3fbc5d28 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 907.231817] env[62066]: DEBUG oslo_concurrency.lockutils [None req-adc52c21-6d1b-4740-84d8-9cdb3fbc5d28 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Acquired lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 907.231995] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-adc52c21-6d1b-4740-84d8-9cdb3fbc5d28 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 907.232270] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d1eb5aaa-500b-4e4d-9c2b-1b2ce384d42b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.242515] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-adc52c21-6d1b-4740-84d8-9cdb3fbc5d28 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 907.242707] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-adc52c21-6d1b-4740-84d8-9cdb3fbc5d28 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 907.243669] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-36aeec5f-ef7f-47e1-9d6b-26e8d15c0b86 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.249612] env[62066]: DEBUG oslo_vmware.api [None req-adc52c21-6d1b-4740-84d8-9cdb3fbc5d28 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Waiting for the task: (returnval){ [ 907.249612] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]52f4d282-2d6d-c3d7-ddc6-bf3a8bb5220c" [ 907.249612] env[62066]: _type = "Task" [ 907.249612] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.258088] env[62066]: DEBUG oslo_vmware.api [None req-adc52c21-6d1b-4740-84d8-9cdb3fbc5d28 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52f4d282-2d6d-c3d7-ddc6-bf3a8bb5220c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.350677] env[62066]: DEBUG nova.compute.utils [None req-dc7187c9-b41f-4691-8bae-33272b590d9f tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 907.458573] env[62066]: INFO nova.compute.manager [None req-cbf7a2c2-7fb7-405b-bf5c-0776ba26eb99 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 9c989152-ec69-478e-a0b0-62983852576c] Took 20.77 seconds to build instance. [ 907.688032] env[62066]: DEBUG nova.network.neutron [None req-93fb3d16-c0a4-4bf0-b776-e2995ef54e48 tempest-ServerTagsTestJSON-2015212290 tempest-ServerTagsTestJSON-2015212290-project-member] [instance: f4a84152-f222-46a2-9a35-8a0a0078fc12] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 907.732116] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a96d87c9-14b1-48f6-a5e4-af778c649411 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.836s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 907.734576] env[62066]: DEBUG oslo_concurrency.lockutils [None req-81c18bf6-1c8d-45a0-a66c-dcda0761f479 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.063s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 907.734938] env[62066]: DEBUG nova.objects.instance [None req-81c18bf6-1c8d-45a0-a66c-dcda0761f479 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Lazy-loading 'pci_requests' on Instance uuid 8ff6fbb9-c90f-498d-9a85-d220a8c2f794 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 907.756828] env[62066]: INFO nova.scheduler.client.report [None req-a96d87c9-14b1-48f6-a5e4-af778c649411 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Deleted allocations for instance de2e2be7-efdb-45a8-842a-640ab9deb1d9 [ 907.761459] env[62066]: DEBUG oslo_vmware.api [None req-adc52c21-6d1b-4740-84d8-9cdb3fbc5d28 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52f4d282-2d6d-c3d7-ddc6-bf3a8bb5220c, 'name': SearchDatastore_Task, 'duration_secs': 0.016183} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.764550] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fc4149b8-f8e2-4ecd-8b65-cb2da30da426 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.770424] env[62066]: DEBUG oslo_vmware.api [None req-adc52c21-6d1b-4740-84d8-9cdb3fbc5d28 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Waiting for the task: (returnval){ [ 907.770424] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]52c38bdf-bffc-1a26-cab3-bd96dd6cd65b" [ 907.770424] env[62066]: _type = "Task" [ 907.770424] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.778374] env[62066]: DEBUG oslo_vmware.api [None req-adc52c21-6d1b-4740-84d8-9cdb3fbc5d28 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52c38bdf-bffc-1a26-cab3-bd96dd6cd65b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.854182] env[62066]: DEBUG oslo_concurrency.lockutils [None req-dc7187c9-b41f-4691-8bae-33272b590d9f tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Lock "c8e0d47c-4421-4e00-9183-206fceeabc40" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.009s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 907.885315] env[62066]: DEBUG nova.network.neutron [None req-93fb3d16-c0a4-4bf0-b776-e2995ef54e48 tempest-ServerTagsTestJSON-2015212290 tempest-ServerTagsTestJSON-2015212290-project-member] [instance: f4a84152-f222-46a2-9a35-8a0a0078fc12] Updating instance_info_cache with network_info: [{"id": "35ae652b-5fe5-468b-b4d4-75e727382145", "address": "fa:16:3e:65:8d:48", "network": {"id": "5f5893c8-7890-4f62-8ffa-87ebd9b85b90", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-1079654002-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "abb1fa60628341849e205d03e1f7e8ed", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c20f5114-0866-45b3-9a7c-62f113ff83fa", "external-id": "nsx-vlan-transportzone-47", "segmentation_id": 47, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap35ae652b-5f", "ovs_interfaceid": "35ae652b-5fe5-468b-b4d4-75e727382145", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 907.916921] env[62066]: DEBUG nova.compute.manager [None req-36f2458b-83e5-4859-9178-b3e847334e66 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 4e39b9b0-4ab0-4f48-8a3f-6bc1895597eb] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 907.942703] env[62066]: DEBUG nova.virt.hardware [None req-36f2458b-83e5-4859-9178-b3e847334e66 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-23T13:40:41Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-23T13:40:23Z,direct_url=,disk_format='vmdk',id=50ff584c-3b50-4395-af07-3e66769bc9f7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='eb52f7069a374c61ae946f052007c6d9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-23T13:40:24Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 907.942957] env[62066]: DEBUG nova.virt.hardware [None req-36f2458b-83e5-4859-9178-b3e847334e66 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 907.943136] env[62066]: DEBUG nova.virt.hardware [None req-36f2458b-83e5-4859-9178-b3e847334e66 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 907.943327] env[62066]: DEBUG nova.virt.hardware [None req-36f2458b-83e5-4859-9178-b3e847334e66 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 907.943476] env[62066]: DEBUG nova.virt.hardware [None req-36f2458b-83e5-4859-9178-b3e847334e66 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 907.943625] env[62066]: DEBUG nova.virt.hardware [None req-36f2458b-83e5-4859-9178-b3e847334e66 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 907.943833] env[62066]: DEBUG nova.virt.hardware [None req-36f2458b-83e5-4859-9178-b3e847334e66 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 907.943994] env[62066]: DEBUG nova.virt.hardware [None req-36f2458b-83e5-4859-9178-b3e847334e66 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 907.944178] env[62066]: DEBUG nova.virt.hardware [None req-36f2458b-83e5-4859-9178-b3e847334e66 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 907.944343] env[62066]: DEBUG nova.virt.hardware [None req-36f2458b-83e5-4859-9178-b3e847334e66 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 907.944515] env[62066]: DEBUG nova.virt.hardware [None req-36f2458b-83e5-4859-9178-b3e847334e66 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 907.945387] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8168bcf5-a8c3-4d7d-93cc-57523a8f3b3e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.953290] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e61f2815-d417-4d8c-b4ed-d3a98d9a4c0f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.966782] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cbf7a2c2-7fb7-405b-bf5c-0776ba26eb99 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Lock "9c989152-ec69-478e-a0b0-62983852576c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.289s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 908.119863] env[62066]: DEBUG nova.compute.manager [req-3374c398-983f-48a7-be5c-5f4c6bbfad23 req-9268695e-af5a-4f62-8d1b-f0f4f1869cb5 service nova] [instance: f4a84152-f222-46a2-9a35-8a0a0078fc12] Received event network-vif-plugged-35ae652b-5fe5-468b-b4d4-75e727382145 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 908.120226] env[62066]: DEBUG oslo_concurrency.lockutils [req-3374c398-983f-48a7-be5c-5f4c6bbfad23 req-9268695e-af5a-4f62-8d1b-f0f4f1869cb5 service nova] Acquiring lock "f4a84152-f222-46a2-9a35-8a0a0078fc12-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 908.120462] env[62066]: DEBUG oslo_concurrency.lockutils [req-3374c398-983f-48a7-be5c-5f4c6bbfad23 req-9268695e-af5a-4f62-8d1b-f0f4f1869cb5 service nova] Lock "f4a84152-f222-46a2-9a35-8a0a0078fc12-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 908.120643] env[62066]: DEBUG oslo_concurrency.lockutils [req-3374c398-983f-48a7-be5c-5f4c6bbfad23 req-9268695e-af5a-4f62-8d1b-f0f4f1869cb5 service nova] Lock "f4a84152-f222-46a2-9a35-8a0a0078fc12-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 908.120822] env[62066]: DEBUG nova.compute.manager [req-3374c398-983f-48a7-be5c-5f4c6bbfad23 req-9268695e-af5a-4f62-8d1b-f0f4f1869cb5 service nova] [instance: f4a84152-f222-46a2-9a35-8a0a0078fc12] No waiting events found dispatching network-vif-plugged-35ae652b-5fe5-468b-b4d4-75e727382145 {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 908.121033] env[62066]: WARNING nova.compute.manager [req-3374c398-983f-48a7-be5c-5f4c6bbfad23 req-9268695e-af5a-4f62-8d1b-f0f4f1869cb5 service nova] [instance: f4a84152-f222-46a2-9a35-8a0a0078fc12] Received unexpected event network-vif-plugged-35ae652b-5fe5-468b-b4d4-75e727382145 for instance with vm_state building and task_state spawning. [ 908.121224] env[62066]: DEBUG nova.compute.manager [req-3374c398-983f-48a7-be5c-5f4c6bbfad23 req-9268695e-af5a-4f62-8d1b-f0f4f1869cb5 service nova] [instance: f4a84152-f222-46a2-9a35-8a0a0078fc12] Received event network-changed-35ae652b-5fe5-468b-b4d4-75e727382145 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 908.121390] env[62066]: DEBUG nova.compute.manager [req-3374c398-983f-48a7-be5c-5f4c6bbfad23 req-9268695e-af5a-4f62-8d1b-f0f4f1869cb5 service nova] [instance: f4a84152-f222-46a2-9a35-8a0a0078fc12] Refreshing instance network info cache due to event network-changed-35ae652b-5fe5-468b-b4d4-75e727382145. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 908.121564] env[62066]: DEBUG oslo_concurrency.lockutils [req-3374c398-983f-48a7-be5c-5f4c6bbfad23 req-9268695e-af5a-4f62-8d1b-f0f4f1869cb5 service nova] Acquiring lock "refresh_cache-f4a84152-f222-46a2-9a35-8a0a0078fc12" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 908.238760] env[62066]: DEBUG nova.objects.instance [None req-81c18bf6-1c8d-45a0-a66c-dcda0761f479 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Lazy-loading 'numa_topology' on Instance uuid 8ff6fbb9-c90f-498d-9a85-d220a8c2f794 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 908.268261] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a96d87c9-14b1-48f6-a5e4-af778c649411 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Lock "de2e2be7-efdb-45a8-842a-640ab9deb1d9" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 17.111s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 908.287659] env[62066]: DEBUG oslo_vmware.api [None req-adc52c21-6d1b-4740-84d8-9cdb3fbc5d28 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52c38bdf-bffc-1a26-cab3-bd96dd6cd65b, 'name': SearchDatastore_Task, 'duration_secs': 0.009253} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.287961] env[62066]: DEBUG oslo_concurrency.lockutils [None req-adc52c21-6d1b-4740-84d8-9cdb3fbc5d28 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Releasing lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 908.288274] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-adc52c21-6d1b-4740-84d8-9cdb3fbc5d28 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk to [datastore2] 90c33434-c127-450b-9ff0-75181b4ac385/90c33434-c127-450b-9ff0-75181b4ac385.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 908.288583] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6437236f-4f1a-411f-9ba4-ecf9e1c18c8c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.296475] env[62066]: DEBUG oslo_vmware.api [None req-adc52c21-6d1b-4740-84d8-9cdb3fbc5d28 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Waiting for the task: (returnval){ [ 908.296475] env[62066]: value = "task-1341098" [ 908.296475] env[62066]: _type = "Task" [ 908.296475] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.306630] env[62066]: DEBUG oslo_vmware.api [None req-adc52c21-6d1b-4740-84d8-9cdb3fbc5d28 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341098, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.352225] env[62066]: DEBUG nova.network.neutron [None req-36f2458b-83e5-4859-9178-b3e847334e66 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 4e39b9b0-4ab0-4f48-8a3f-6bc1895597eb] Successfully updated port: 843a9051-1a6f-42b2-b482-2baf7ffb6ca4 {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 908.388318] env[62066]: DEBUG oslo_concurrency.lockutils [None req-93fb3d16-c0a4-4bf0-b776-e2995ef54e48 tempest-ServerTagsTestJSON-2015212290 tempest-ServerTagsTestJSON-2015212290-project-member] Releasing lock "refresh_cache-f4a84152-f222-46a2-9a35-8a0a0078fc12" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 908.388912] env[62066]: DEBUG nova.compute.manager [None req-93fb3d16-c0a4-4bf0-b776-e2995ef54e48 tempest-ServerTagsTestJSON-2015212290 tempest-ServerTagsTestJSON-2015212290-project-member] [instance: f4a84152-f222-46a2-9a35-8a0a0078fc12] Instance network_info: |[{"id": "35ae652b-5fe5-468b-b4d4-75e727382145", "address": "fa:16:3e:65:8d:48", "network": {"id": "5f5893c8-7890-4f62-8ffa-87ebd9b85b90", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-1079654002-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "abb1fa60628341849e205d03e1f7e8ed", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c20f5114-0866-45b3-9a7c-62f113ff83fa", "external-id": "nsx-vlan-transportzone-47", "segmentation_id": 47, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap35ae652b-5f", "ovs_interfaceid": "35ae652b-5fe5-468b-b4d4-75e727382145", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 908.389077] env[62066]: DEBUG oslo_concurrency.lockutils [req-3374c398-983f-48a7-be5c-5f4c6bbfad23 req-9268695e-af5a-4f62-8d1b-f0f4f1869cb5 service nova] Acquired lock "refresh_cache-f4a84152-f222-46a2-9a35-8a0a0078fc12" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 908.389137] env[62066]: DEBUG nova.network.neutron [req-3374c398-983f-48a7-be5c-5f4c6bbfad23 req-9268695e-af5a-4f62-8d1b-f0f4f1869cb5 service nova] [instance: f4a84152-f222-46a2-9a35-8a0a0078fc12] Refreshing network info cache for port 35ae652b-5fe5-468b-b4d4-75e727382145 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 908.390393] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-93fb3d16-c0a4-4bf0-b776-e2995ef54e48 tempest-ServerTagsTestJSON-2015212290 tempest-ServerTagsTestJSON-2015212290-project-member] [instance: f4a84152-f222-46a2-9a35-8a0a0078fc12] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:65:8d:48', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c20f5114-0866-45b3-9a7c-62f113ff83fa', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '35ae652b-5fe5-468b-b4d4-75e727382145', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 908.398310] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-93fb3d16-c0a4-4bf0-b776-e2995ef54e48 tempest-ServerTagsTestJSON-2015212290 tempest-ServerTagsTestJSON-2015212290-project-member] Creating folder: Project (abb1fa60628341849e205d03e1f7e8ed). Parent ref: group-v285980. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 908.399309] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-271c9d08-61b6-4ef8-b9b3-b4b49dfc3d63 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.409935] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-93fb3d16-c0a4-4bf0-b776-e2995ef54e48 tempest-ServerTagsTestJSON-2015212290 tempest-ServerTagsTestJSON-2015212290-project-member] Created folder: Project (abb1fa60628341849e205d03e1f7e8ed) in parent group-v285980. [ 908.410276] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-93fb3d16-c0a4-4bf0-b776-e2995ef54e48 tempest-ServerTagsTestJSON-2015212290 tempest-ServerTagsTestJSON-2015212290-project-member] Creating folder: Instances. Parent ref: group-v286095. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 908.410417] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-afb984e6-b453-49bf-bffe-58d209e5be47 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.420345] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-93fb3d16-c0a4-4bf0-b776-e2995ef54e48 tempest-ServerTagsTestJSON-2015212290 tempest-ServerTagsTestJSON-2015212290-project-member] Created folder: Instances in parent group-v286095. [ 908.420598] env[62066]: DEBUG oslo.service.loopingcall [None req-93fb3d16-c0a4-4bf0-b776-e2995ef54e48 tempest-ServerTagsTestJSON-2015212290 tempest-ServerTagsTestJSON-2015212290-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 908.420792] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f4a84152-f222-46a2-9a35-8a0a0078fc12] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 908.421019] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6d9523f6-d250-436a-8340-969765448503 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.442839] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 908.442839] env[62066]: value = "task-1341101" [ 908.442839] env[62066]: _type = "Task" [ 908.442839] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.447825] env[62066]: DEBUG nova.compute.manager [req-e50213bb-51cd-4fef-bc59-d315f2fd8386 req-b1792d89-bd16-4738-84cc-56f6825b7379 service nova] [instance: 4e39b9b0-4ab0-4f48-8a3f-6bc1895597eb] Received event network-vif-plugged-843a9051-1a6f-42b2-b482-2baf7ffb6ca4 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 908.448061] env[62066]: DEBUG oslo_concurrency.lockutils [req-e50213bb-51cd-4fef-bc59-d315f2fd8386 req-b1792d89-bd16-4738-84cc-56f6825b7379 service nova] Acquiring lock "4e39b9b0-4ab0-4f48-8a3f-6bc1895597eb-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 908.449221] env[62066]: DEBUG oslo_concurrency.lockutils [req-e50213bb-51cd-4fef-bc59-d315f2fd8386 req-b1792d89-bd16-4738-84cc-56f6825b7379 service nova] Lock "4e39b9b0-4ab0-4f48-8a3f-6bc1895597eb-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 908.449400] env[62066]: DEBUG oslo_concurrency.lockutils [req-e50213bb-51cd-4fef-bc59-d315f2fd8386 req-b1792d89-bd16-4738-84cc-56f6825b7379 service nova] Lock "4e39b9b0-4ab0-4f48-8a3f-6bc1895597eb-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 908.449636] env[62066]: DEBUG nova.compute.manager [req-e50213bb-51cd-4fef-bc59-d315f2fd8386 req-b1792d89-bd16-4738-84cc-56f6825b7379 service nova] [instance: 4e39b9b0-4ab0-4f48-8a3f-6bc1895597eb] No waiting events found dispatching network-vif-plugged-843a9051-1a6f-42b2-b482-2baf7ffb6ca4 {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 908.451497] env[62066]: WARNING nova.compute.manager [req-e50213bb-51cd-4fef-bc59-d315f2fd8386 req-b1792d89-bd16-4738-84cc-56f6825b7379 service nova] [instance: 4e39b9b0-4ab0-4f48-8a3f-6bc1895597eb] Received unexpected event network-vif-plugged-843a9051-1a6f-42b2-b482-2baf7ffb6ca4 for instance with vm_state building and task_state spawning. [ 908.457174] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1341101, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.744106] env[62066]: INFO nova.compute.claims [None req-81c18bf6-1c8d-45a0-a66c-dcda0761f479 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] [instance: 8ff6fbb9-c90f-498d-9a85-d220a8c2f794] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 908.808511] env[62066]: DEBUG oslo_vmware.api [None req-adc52c21-6d1b-4740-84d8-9cdb3fbc5d28 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341098, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.486968} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.808511] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-adc52c21-6d1b-4740-84d8-9cdb3fbc5d28 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk to [datastore2] 90c33434-c127-450b-9ff0-75181b4ac385/90c33434-c127-450b-9ff0-75181b4ac385.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 908.808511] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-adc52c21-6d1b-4740-84d8-9cdb3fbc5d28 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: 90c33434-c127-450b-9ff0-75181b4ac385] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 908.808511] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-34906d0e-a45d-4fca-a86b-f71e06fe05b4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.814709] env[62066]: DEBUG oslo_vmware.api [None req-adc52c21-6d1b-4740-84d8-9cdb3fbc5d28 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Waiting for the task: (returnval){ [ 908.814709] env[62066]: value = "task-1341102" [ 908.814709] env[62066]: _type = "Task" [ 908.814709] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.822607] env[62066]: DEBUG oslo_vmware.api [None req-adc52c21-6d1b-4740-84d8-9cdb3fbc5d28 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341102, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.858720] env[62066]: DEBUG oslo_concurrency.lockutils [None req-36f2458b-83e5-4859-9178-b3e847334e66 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Acquiring lock "refresh_cache-4e39b9b0-4ab0-4f48-8a3f-6bc1895597eb" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 908.858720] env[62066]: DEBUG oslo_concurrency.lockutils [None req-36f2458b-83e5-4859-9178-b3e847334e66 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Acquired lock "refresh_cache-4e39b9b0-4ab0-4f48-8a3f-6bc1895597eb" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 908.859210] env[62066]: DEBUG nova.network.neutron [None req-36f2458b-83e5-4859-9178-b3e847334e66 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 4e39b9b0-4ab0-4f48-8a3f-6bc1895597eb] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 908.940277] env[62066]: DEBUG oslo_concurrency.lockutils [None req-dc7187c9-b41f-4691-8bae-33272b590d9f tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Acquiring lock "c8e0d47c-4421-4e00-9183-206fceeabc40" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 908.940526] env[62066]: DEBUG oslo_concurrency.lockutils [None req-dc7187c9-b41f-4691-8bae-33272b590d9f tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Lock "c8e0d47c-4421-4e00-9183-206fceeabc40" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 908.940842] env[62066]: INFO nova.compute.manager [None req-dc7187c9-b41f-4691-8bae-33272b590d9f tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: c8e0d47c-4421-4e00-9183-206fceeabc40] Attaching volume 338bd175-a96d-417e-a5b0-380b7d939a85 to /dev/sdb [ 908.955132] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1341101, 'name': CreateVM_Task} progress is 99%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.985480] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1da8270-2756-44b4-b7c1-f4ade8a51bfe {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.997697] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3320fb9e-b453-4483-9a27-cb5f4544e278 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.014158] env[62066]: DEBUG nova.virt.block_device [None req-dc7187c9-b41f-4691-8bae-33272b590d9f tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: c8e0d47c-4421-4e00-9183-206fceeabc40] Updating existing volume attachment record: 847ed3fa-1c18-4a68-850c-ae54d01aa7ba {{(pid=62066) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 909.185544] env[62066]: DEBUG nova.network.neutron [req-3374c398-983f-48a7-be5c-5f4c6bbfad23 req-9268695e-af5a-4f62-8d1b-f0f4f1869cb5 service nova] [instance: f4a84152-f222-46a2-9a35-8a0a0078fc12] Updated VIF entry in instance network info cache for port 35ae652b-5fe5-468b-b4d4-75e727382145. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 909.186143] env[62066]: DEBUG nova.network.neutron [req-3374c398-983f-48a7-be5c-5f4c6bbfad23 req-9268695e-af5a-4f62-8d1b-f0f4f1869cb5 service nova] [instance: f4a84152-f222-46a2-9a35-8a0a0078fc12] Updating instance_info_cache with network_info: [{"id": "35ae652b-5fe5-468b-b4d4-75e727382145", "address": "fa:16:3e:65:8d:48", "network": {"id": "5f5893c8-7890-4f62-8ffa-87ebd9b85b90", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-1079654002-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "abb1fa60628341849e205d03e1f7e8ed", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c20f5114-0866-45b3-9a7c-62f113ff83fa", "external-id": "nsx-vlan-transportzone-47", "segmentation_id": 47, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap35ae652b-5f", "ovs_interfaceid": "35ae652b-5fe5-468b-b4d4-75e727382145", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 909.325013] env[62066]: DEBUG oslo_vmware.api [None req-adc52c21-6d1b-4740-84d8-9cdb3fbc5d28 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341102, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064722} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.325328] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-adc52c21-6d1b-4740-84d8-9cdb3fbc5d28 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: 90c33434-c127-450b-9ff0-75181b4ac385] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 909.326100] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6205436-42c6-4442-a2e4-46643a9517a4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.349906] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-adc52c21-6d1b-4740-84d8-9cdb3fbc5d28 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: 90c33434-c127-450b-9ff0-75181b4ac385] Reconfiguring VM instance instance-00000057 to attach disk [datastore2] 90c33434-c127-450b-9ff0-75181b4ac385/90c33434-c127-450b-9ff0-75181b4ac385.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 909.352610] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2e1ced00-e392-4677-9b22-d001745d05d4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.368264] env[62066]: DEBUG nova.compute.manager [None req-740993dc-fc35-4aa8-9a06-8f4c0e5f95e1 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 9c989152-ec69-478e-a0b0-62983852576c] Stashing vm_state: active {{(pid=62066) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 909.379054] env[62066]: DEBUG oslo_vmware.api [None req-adc52c21-6d1b-4740-84d8-9cdb3fbc5d28 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Waiting for the task: (returnval){ [ 909.379054] env[62066]: value = "task-1341106" [ 909.379054] env[62066]: _type = "Task" [ 909.379054] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.387414] env[62066]: DEBUG oslo_vmware.api [None req-adc52c21-6d1b-4740-84d8-9cdb3fbc5d28 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341106, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.406317] env[62066]: DEBUG nova.network.neutron [None req-36f2458b-83e5-4859-9178-b3e847334e66 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 4e39b9b0-4ab0-4f48-8a3f-6bc1895597eb] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 909.455664] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1341101, 'name': CreateVM_Task} progress is 99%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.545698] env[62066]: DEBUG nova.network.neutron [None req-36f2458b-83e5-4859-9178-b3e847334e66 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 4e39b9b0-4ab0-4f48-8a3f-6bc1895597eb] Updating instance_info_cache with network_info: [{"id": "843a9051-1a6f-42b2-b482-2baf7ffb6ca4", "address": "fa:16:3e:b7:ff:a5", "network": {"id": "9f3396ce-ee48-4eaf-8cc7-3e1711f37537", "bridge": "br-int", "label": "tempest-ServersTestJSON-2129775093-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6497ab02f327476d8ff81c2ecc0371e5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7ab8d568-adb0-4f3b-b6cc-68413e6546ae", "external-id": "nsx-vlan-transportzone-86", "segmentation_id": 86, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap843a9051-1a", "ovs_interfaceid": "843a9051-1a6f-42b2-b482-2baf7ffb6ca4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 909.689326] env[62066]: DEBUG oslo_concurrency.lockutils [req-3374c398-983f-48a7-be5c-5f4c6bbfad23 req-9268695e-af5a-4f62-8d1b-f0f4f1869cb5 service nova] Releasing lock "refresh_cache-f4a84152-f222-46a2-9a35-8a0a0078fc12" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 909.887765] env[62066]: DEBUG oslo_concurrency.lockutils [None req-740993dc-fc35-4aa8-9a06-8f4c0e5f95e1 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 909.897664] env[62066]: DEBUG oslo_vmware.api [None req-adc52c21-6d1b-4740-84d8-9cdb3fbc5d28 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341106, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.959359] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1341101, 'name': CreateVM_Task} progress is 99%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.963606] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a25a378-0225-449d-aca0-b9ec764dfee2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.971685] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0061067c-f295-484a-aec9-9f9a41beaa33 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.004908] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67bf4f10-82c8-4993-8a5c-91b3bf45e739 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.012435] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15be3052-e570-44a0-8ba0-30b2446eb582 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.025657] env[62066]: DEBUG nova.compute.provider_tree [None req-81c18bf6-1c8d-45a0-a66c-dcda0761f479 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 910.049013] env[62066]: DEBUG oslo_concurrency.lockutils [None req-36f2458b-83e5-4859-9178-b3e847334e66 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Releasing lock "refresh_cache-4e39b9b0-4ab0-4f48-8a3f-6bc1895597eb" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 910.049449] env[62066]: DEBUG nova.compute.manager [None req-36f2458b-83e5-4859-9178-b3e847334e66 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 4e39b9b0-4ab0-4f48-8a3f-6bc1895597eb] Instance network_info: |[{"id": "843a9051-1a6f-42b2-b482-2baf7ffb6ca4", "address": "fa:16:3e:b7:ff:a5", "network": {"id": "9f3396ce-ee48-4eaf-8cc7-3e1711f37537", "bridge": "br-int", "label": "tempest-ServersTestJSON-2129775093-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6497ab02f327476d8ff81c2ecc0371e5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7ab8d568-adb0-4f3b-b6cc-68413e6546ae", "external-id": "nsx-vlan-transportzone-86", "segmentation_id": 86, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap843a9051-1a", "ovs_interfaceid": "843a9051-1a6f-42b2-b482-2baf7ffb6ca4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 910.049814] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-36f2458b-83e5-4859-9178-b3e847334e66 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 4e39b9b0-4ab0-4f48-8a3f-6bc1895597eb] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b7:ff:a5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7ab8d568-adb0-4f3b-b6cc-68413e6546ae', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '843a9051-1a6f-42b2-b482-2baf7ffb6ca4', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 910.058287] env[62066]: DEBUG oslo.service.loopingcall [None req-36f2458b-83e5-4859-9178-b3e847334e66 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 910.059014] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4e39b9b0-4ab0-4f48-8a3f-6bc1895597eb] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 910.059197] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d843e7c0-c15a-452d-9c7f-3516b28bc926 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.088257] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 910.088257] env[62066]: value = "task-1341107" [ 910.088257] env[62066]: _type = "Task" [ 910.088257] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.099513] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1341107, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.391126] env[62066]: DEBUG oslo_vmware.api [None req-adc52c21-6d1b-4740-84d8-9cdb3fbc5d28 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341106, 'name': ReconfigVM_Task, 'duration_secs': 0.71405} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.391440] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-adc52c21-6d1b-4740-84d8-9cdb3fbc5d28 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: 90c33434-c127-450b-9ff0-75181b4ac385] Reconfigured VM instance instance-00000057 to attach disk [datastore2] 90c33434-c127-450b-9ff0-75181b4ac385/90c33434-c127-450b-9ff0-75181b4ac385.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 910.392087] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e92dc988-36ab-41a6-abdc-95d385a82ee6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.397703] env[62066]: DEBUG oslo_vmware.api [None req-adc52c21-6d1b-4740-84d8-9cdb3fbc5d28 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Waiting for the task: (returnval){ [ 910.397703] env[62066]: value = "task-1341108" [ 910.397703] env[62066]: _type = "Task" [ 910.397703] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.404984] env[62066]: DEBUG oslo_vmware.api [None req-adc52c21-6d1b-4740-84d8-9cdb3fbc5d28 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341108, 'name': Rename_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.454288] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1341101, 'name': CreateVM_Task, 'duration_secs': 1.593746} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.454413] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f4a84152-f222-46a2-9a35-8a0a0078fc12] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 910.455118] env[62066]: DEBUG oslo_concurrency.lockutils [None req-93fb3d16-c0a4-4bf0-b776-e2995ef54e48 tempest-ServerTagsTestJSON-2015212290 tempest-ServerTagsTestJSON-2015212290-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 910.455282] env[62066]: DEBUG oslo_concurrency.lockutils [None req-93fb3d16-c0a4-4bf0-b776-e2995ef54e48 tempest-ServerTagsTestJSON-2015212290 tempest-ServerTagsTestJSON-2015212290-project-member] Acquired lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 910.455595] env[62066]: DEBUG oslo_concurrency.lockutils [None req-93fb3d16-c0a4-4bf0-b776-e2995ef54e48 tempest-ServerTagsTestJSON-2015212290 tempest-ServerTagsTestJSON-2015212290-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 910.455859] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6d419534-810e-4cba-89d6-5b219ae362f2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.460710] env[62066]: DEBUG oslo_vmware.api [None req-93fb3d16-c0a4-4bf0-b776-e2995ef54e48 tempest-ServerTagsTestJSON-2015212290 tempest-ServerTagsTestJSON-2015212290-project-member] Waiting for the task: (returnval){ [ 910.460710] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]524f8b4d-97cd-ace4-d9a9-9d284ba3de0e" [ 910.460710] env[62066]: _type = "Task" [ 910.460710] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.468621] env[62066]: DEBUG oslo_vmware.api [None req-93fb3d16-c0a4-4bf0-b776-e2995ef54e48 tempest-ServerTagsTestJSON-2015212290 tempest-ServerTagsTestJSON-2015212290-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]524f8b4d-97cd-ace4-d9a9-9d284ba3de0e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.500322] env[62066]: DEBUG nova.compute.manager [req-8e196c3c-2bbd-435b-9dc8-8a276d06402c req-d40ef172-8000-44e3-af91-f2e4157aeef7 service nova] [instance: 4e39b9b0-4ab0-4f48-8a3f-6bc1895597eb] Received event network-changed-843a9051-1a6f-42b2-b482-2baf7ffb6ca4 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 910.500517] env[62066]: DEBUG nova.compute.manager [req-8e196c3c-2bbd-435b-9dc8-8a276d06402c req-d40ef172-8000-44e3-af91-f2e4157aeef7 service nova] [instance: 4e39b9b0-4ab0-4f48-8a3f-6bc1895597eb] Refreshing instance network info cache due to event network-changed-843a9051-1a6f-42b2-b482-2baf7ffb6ca4. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 910.501171] env[62066]: DEBUG oslo_concurrency.lockutils [req-8e196c3c-2bbd-435b-9dc8-8a276d06402c req-d40ef172-8000-44e3-af91-f2e4157aeef7 service nova] Acquiring lock "refresh_cache-4e39b9b0-4ab0-4f48-8a3f-6bc1895597eb" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 910.501171] env[62066]: DEBUG oslo_concurrency.lockutils [req-8e196c3c-2bbd-435b-9dc8-8a276d06402c req-d40ef172-8000-44e3-af91-f2e4157aeef7 service nova] Acquired lock "refresh_cache-4e39b9b0-4ab0-4f48-8a3f-6bc1895597eb" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 910.501171] env[62066]: DEBUG nova.network.neutron [req-8e196c3c-2bbd-435b-9dc8-8a276d06402c req-d40ef172-8000-44e3-af91-f2e4157aeef7 service nova] [instance: 4e39b9b0-4ab0-4f48-8a3f-6bc1895597eb] Refreshing network info cache for port 843a9051-1a6f-42b2-b482-2baf7ffb6ca4 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 910.528478] env[62066]: DEBUG nova.scheduler.client.report [None req-81c18bf6-1c8d-45a0-a66c-dcda0761f479 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 910.598158] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1341107, 'name': CreateVM_Task, 'duration_secs': 0.480847} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.598334] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4e39b9b0-4ab0-4f48-8a3f-6bc1895597eb] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 910.598973] env[62066]: DEBUG oslo_concurrency.lockutils [None req-36f2458b-83e5-4859-9178-b3e847334e66 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 910.908490] env[62066]: DEBUG oslo_vmware.api [None req-adc52c21-6d1b-4740-84d8-9cdb3fbc5d28 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341108, 'name': Rename_Task, 'duration_secs': 0.198996} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.908490] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-adc52c21-6d1b-4740-84d8-9cdb3fbc5d28 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: 90c33434-c127-450b-9ff0-75181b4ac385] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 910.908860] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-98521858-cd01-403f-85e5-d053373a8146 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.914892] env[62066]: DEBUG oslo_vmware.api [None req-adc52c21-6d1b-4740-84d8-9cdb3fbc5d28 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Waiting for the task: (returnval){ [ 910.914892] env[62066]: value = "task-1341109" [ 910.914892] env[62066]: _type = "Task" [ 910.914892] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.922136] env[62066]: DEBUG oslo_vmware.api [None req-adc52c21-6d1b-4740-84d8-9cdb3fbc5d28 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341109, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.972039] env[62066]: DEBUG oslo_vmware.api [None req-93fb3d16-c0a4-4bf0-b776-e2995ef54e48 tempest-ServerTagsTestJSON-2015212290 tempest-ServerTagsTestJSON-2015212290-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]524f8b4d-97cd-ace4-d9a9-9d284ba3de0e, 'name': SearchDatastore_Task, 'duration_secs': 0.011187} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.972039] env[62066]: DEBUG oslo_concurrency.lockutils [None req-93fb3d16-c0a4-4bf0-b776-e2995ef54e48 tempest-ServerTagsTestJSON-2015212290 tempest-ServerTagsTestJSON-2015212290-project-member] Releasing lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 910.972039] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-93fb3d16-c0a4-4bf0-b776-e2995ef54e48 tempest-ServerTagsTestJSON-2015212290 tempest-ServerTagsTestJSON-2015212290-project-member] [instance: f4a84152-f222-46a2-9a35-8a0a0078fc12] Processing image 50ff584c-3b50-4395-af07-3e66769bc9f7 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 910.972039] env[62066]: DEBUG oslo_concurrency.lockutils [None req-93fb3d16-c0a4-4bf0-b776-e2995ef54e48 tempest-ServerTagsTestJSON-2015212290 tempest-ServerTagsTestJSON-2015212290-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 910.972039] env[62066]: DEBUG oslo_concurrency.lockutils [None req-93fb3d16-c0a4-4bf0-b776-e2995ef54e48 tempest-ServerTagsTestJSON-2015212290 tempest-ServerTagsTestJSON-2015212290-project-member] Acquired lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 910.972039] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-93fb3d16-c0a4-4bf0-b776-e2995ef54e48 tempest-ServerTagsTestJSON-2015212290 tempest-ServerTagsTestJSON-2015212290-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 910.972408] env[62066]: DEBUG oslo_concurrency.lockutils [None req-36f2458b-83e5-4859-9178-b3e847334e66 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Acquired lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 910.972487] env[62066]: DEBUG oslo_concurrency.lockutils [None req-36f2458b-83e5-4859-9178-b3e847334e66 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 910.972717] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3ef712a4-4fee-4c5d-89cc-5549ce0378b8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.976016] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9a611590-8333-4ece-abd2-4f1252cfd39b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.980760] env[62066]: DEBUG oslo_vmware.api [None req-36f2458b-83e5-4859-9178-b3e847334e66 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Waiting for the task: (returnval){ [ 910.980760] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]52d31f20-2b0e-538d-e980-c6009190e0c0" [ 910.980760] env[62066]: _type = "Task" [ 910.980760] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.984771] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-93fb3d16-c0a4-4bf0-b776-e2995ef54e48 tempest-ServerTagsTestJSON-2015212290 tempest-ServerTagsTestJSON-2015212290-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 910.984955] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-93fb3d16-c0a4-4bf0-b776-e2995ef54e48 tempest-ServerTagsTestJSON-2015212290 tempest-ServerTagsTestJSON-2015212290-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 910.985995] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6adb5d12-acbe-4ce9-a251-138405d13169 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.992104] env[62066]: DEBUG oslo_vmware.api [None req-36f2458b-83e5-4859-9178-b3e847334e66 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52d31f20-2b0e-538d-e980-c6009190e0c0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.995751] env[62066]: DEBUG oslo_vmware.api [None req-93fb3d16-c0a4-4bf0-b776-e2995ef54e48 tempest-ServerTagsTestJSON-2015212290 tempest-ServerTagsTestJSON-2015212290-project-member] Waiting for the task: (returnval){ [ 910.995751] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]528236e6-acb0-9376-f7c5-205d48a312c8" [ 910.995751] env[62066]: _type = "Task" [ 910.995751] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.004659] env[62066]: DEBUG oslo_vmware.api [None req-93fb3d16-c0a4-4bf0-b776-e2995ef54e48 tempest-ServerTagsTestJSON-2015212290 tempest-ServerTagsTestJSON-2015212290-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]528236e6-acb0-9376-f7c5-205d48a312c8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.035772] env[62066]: DEBUG oslo_concurrency.lockutils [None req-81c18bf6-1c8d-45a0-a66c-dcda0761f479 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.301s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 911.040025] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cb83f312-6ced-4dc9-b42a-28a22511dbae tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 14.445s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 911.040025] env[62066]: DEBUG nova.objects.instance [None req-cb83f312-6ced-4dc9-b42a-28a22511dbae tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Lazy-loading 'resources' on Instance uuid 2634fcc0-96bd-4513-8f92-515f83fe23d7 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 911.073546] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e9681fdd-3744-4c1e-81e7-a2a0f833c385 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Acquiring lock "a10a6dc0-7df7-40a1-888a-d1414dbeb1f7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 911.073793] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e9681fdd-3744-4c1e-81e7-a2a0f833c385 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Lock "a10a6dc0-7df7-40a1-888a-d1414dbeb1f7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 911.096637] env[62066]: INFO nova.network.neutron [None req-81c18bf6-1c8d-45a0-a66c-dcda0761f479 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] [instance: 8ff6fbb9-c90f-498d-9a85-d220a8c2f794] Updating port 8c324833-7bf9-4007-85aa-5b16871f63c7 with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 911.203871] env[62066]: DEBUG nova.network.neutron [req-8e196c3c-2bbd-435b-9dc8-8a276d06402c req-d40ef172-8000-44e3-af91-f2e4157aeef7 service nova] [instance: 4e39b9b0-4ab0-4f48-8a3f-6bc1895597eb] Updated VIF entry in instance network info cache for port 843a9051-1a6f-42b2-b482-2baf7ffb6ca4. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 911.204273] env[62066]: DEBUG nova.network.neutron [req-8e196c3c-2bbd-435b-9dc8-8a276d06402c req-d40ef172-8000-44e3-af91-f2e4157aeef7 service nova] [instance: 4e39b9b0-4ab0-4f48-8a3f-6bc1895597eb] Updating instance_info_cache with network_info: [{"id": "843a9051-1a6f-42b2-b482-2baf7ffb6ca4", "address": "fa:16:3e:b7:ff:a5", "network": {"id": "9f3396ce-ee48-4eaf-8cc7-3e1711f37537", "bridge": "br-int", "label": "tempest-ServersTestJSON-2129775093-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6497ab02f327476d8ff81c2ecc0371e5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7ab8d568-adb0-4f3b-b6cc-68413e6546ae", "external-id": "nsx-vlan-transportzone-86", "segmentation_id": 86, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap843a9051-1a", "ovs_interfaceid": "843a9051-1a6f-42b2-b482-2baf7ffb6ca4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 911.425352] env[62066]: DEBUG oslo_vmware.api [None req-adc52c21-6d1b-4740-84d8-9cdb3fbc5d28 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341109, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.491186] env[62066]: DEBUG oslo_vmware.api [None req-36f2458b-83e5-4859-9178-b3e847334e66 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52d31f20-2b0e-538d-e980-c6009190e0c0, 'name': SearchDatastore_Task, 'duration_secs': 0.011223} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.491530] env[62066]: DEBUG oslo_concurrency.lockutils [None req-36f2458b-83e5-4859-9178-b3e847334e66 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Releasing lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 911.491786] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-36f2458b-83e5-4859-9178-b3e847334e66 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 4e39b9b0-4ab0-4f48-8a3f-6bc1895597eb] Processing image 50ff584c-3b50-4395-af07-3e66769bc9f7 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 911.492011] env[62066]: DEBUG oslo_concurrency.lockutils [None req-36f2458b-83e5-4859-9178-b3e847334e66 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 911.504224] env[62066]: DEBUG oslo_vmware.api [None req-93fb3d16-c0a4-4bf0-b776-e2995ef54e48 tempest-ServerTagsTestJSON-2015212290 tempest-ServerTagsTestJSON-2015212290-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]528236e6-acb0-9376-f7c5-205d48a312c8, 'name': SearchDatastore_Task, 'duration_secs': 0.008419} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.505124] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1f4e9620-bfb1-4d69-b9ee-d3d84c09c566 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.510320] env[62066]: DEBUG oslo_vmware.api [None req-93fb3d16-c0a4-4bf0-b776-e2995ef54e48 tempest-ServerTagsTestJSON-2015212290 tempest-ServerTagsTestJSON-2015212290-project-member] Waiting for the task: (returnval){ [ 911.510320] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]52d834c0-7f4e-395d-c915-4f11660fda1d" [ 911.510320] env[62066]: _type = "Task" [ 911.510320] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.517676] env[62066]: DEBUG oslo_vmware.api [None req-93fb3d16-c0a4-4bf0-b776-e2995ef54e48 tempest-ServerTagsTestJSON-2015212290 tempest-ServerTagsTestJSON-2015212290-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52d834c0-7f4e-395d-c915-4f11660fda1d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.575970] env[62066]: DEBUG nova.compute.manager [None req-e9681fdd-3744-4c1e-81e7-a2a0f833c385 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] [instance: a10a6dc0-7df7-40a1-888a-d1414dbeb1f7] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 911.707341] env[62066]: DEBUG oslo_concurrency.lockutils [req-8e196c3c-2bbd-435b-9dc8-8a276d06402c req-d40ef172-8000-44e3-af91-f2e4157aeef7 service nova] Releasing lock "refresh_cache-4e39b9b0-4ab0-4f48-8a3f-6bc1895597eb" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 911.740464] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a4361b5-edf8-4e89-8492-3330f9813ca2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.747652] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2dcae302-8288-4c7e-8b2a-deefff105726 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.776551] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d837cd1-ad6b-499a-99a4-102f008b0d62 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.783391] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4228a6db-ac0a-406b-8cdf-d9dda1052b46 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.796413] env[62066]: DEBUG nova.compute.provider_tree [None req-cb83f312-6ced-4dc9-b42a-28a22511dbae tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 911.926039] env[62066]: DEBUG oslo_vmware.api [None req-adc52c21-6d1b-4740-84d8-9cdb3fbc5d28 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341109, 'name': PowerOnVM_Task, 'duration_secs': 0.757201} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.926356] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-adc52c21-6d1b-4740-84d8-9cdb3fbc5d28 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: 90c33434-c127-450b-9ff0-75181b4ac385] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 911.926573] env[62066]: INFO nova.compute.manager [None req-adc52c21-6d1b-4740-84d8-9cdb3fbc5d28 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: 90c33434-c127-450b-9ff0-75181b4ac385] Took 8.59 seconds to spawn the instance on the hypervisor. [ 911.926695] env[62066]: DEBUG nova.compute.manager [None req-adc52c21-6d1b-4740-84d8-9cdb3fbc5d28 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: 90c33434-c127-450b-9ff0-75181b4ac385] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 911.927447] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5a3bf58-4965-4dc8-8f0c-0a2764905913 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.020185] env[62066]: DEBUG oslo_vmware.api [None req-93fb3d16-c0a4-4bf0-b776-e2995ef54e48 tempest-ServerTagsTestJSON-2015212290 tempest-ServerTagsTestJSON-2015212290-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52d834c0-7f4e-395d-c915-4f11660fda1d, 'name': SearchDatastore_Task, 'duration_secs': 0.008897} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.020485] env[62066]: DEBUG oslo_concurrency.lockutils [None req-93fb3d16-c0a4-4bf0-b776-e2995ef54e48 tempest-ServerTagsTestJSON-2015212290 tempest-ServerTagsTestJSON-2015212290-project-member] Releasing lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 912.020761] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-93fb3d16-c0a4-4bf0-b776-e2995ef54e48 tempest-ServerTagsTestJSON-2015212290 tempest-ServerTagsTestJSON-2015212290-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk to [datastore2] f4a84152-f222-46a2-9a35-8a0a0078fc12/f4a84152-f222-46a2-9a35-8a0a0078fc12.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 912.021123] env[62066]: DEBUG oslo_concurrency.lockutils [None req-36f2458b-83e5-4859-9178-b3e847334e66 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Acquired lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 912.021367] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-36f2458b-83e5-4859-9178-b3e847334e66 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 912.021873] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9304cd42-30dc-47ec-860b-cac634b9fe30 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.023824] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4236b25e-944e-47e0-95d2-c190a239fd1f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.030300] env[62066]: DEBUG oslo_vmware.api [None req-93fb3d16-c0a4-4bf0-b776-e2995ef54e48 tempest-ServerTagsTestJSON-2015212290 tempest-ServerTagsTestJSON-2015212290-project-member] Waiting for the task: (returnval){ [ 912.030300] env[62066]: value = "task-1341111" [ 912.030300] env[62066]: _type = "Task" [ 912.030300] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.034837] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-36f2458b-83e5-4859-9178-b3e847334e66 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 912.034837] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-36f2458b-83e5-4859-9178-b3e847334e66 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 912.035506] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ea4d04fc-3bad-457a-84ab-de96eb1a2032 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.040700] env[62066]: DEBUG oslo_vmware.api [None req-93fb3d16-c0a4-4bf0-b776-e2995ef54e48 tempest-ServerTagsTestJSON-2015212290 tempest-ServerTagsTestJSON-2015212290-project-member] Task: {'id': task-1341111, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.043733] env[62066]: DEBUG oslo_vmware.api [None req-36f2458b-83e5-4859-9178-b3e847334e66 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Waiting for the task: (returnval){ [ 912.043733] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]52a1188d-363d-51d1-82fb-6ffbf3eaf06e" [ 912.043733] env[62066]: _type = "Task" [ 912.043733] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.051244] env[62066]: DEBUG oslo_vmware.api [None req-36f2458b-83e5-4859-9178-b3e847334e66 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52a1188d-363d-51d1-82fb-6ffbf3eaf06e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.096168] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e9681fdd-3744-4c1e-81e7-a2a0f833c385 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 912.299573] env[62066]: DEBUG nova.scheduler.client.report [None req-cb83f312-6ced-4dc9-b42a-28a22511dbae tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 912.443210] env[62066]: INFO nova.compute.manager [None req-adc52c21-6d1b-4740-84d8-9cdb3fbc5d28 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: 90c33434-c127-450b-9ff0-75181b4ac385] Took 21.76 seconds to build instance. [ 912.540456] env[62066]: DEBUG oslo_vmware.api [None req-93fb3d16-c0a4-4bf0-b776-e2995ef54e48 tempest-ServerTagsTestJSON-2015212290 tempest-ServerTagsTestJSON-2015212290-project-member] Task: {'id': task-1341111, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.502478} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.540737] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-93fb3d16-c0a4-4bf0-b776-e2995ef54e48 tempest-ServerTagsTestJSON-2015212290 tempest-ServerTagsTestJSON-2015212290-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk to [datastore2] f4a84152-f222-46a2-9a35-8a0a0078fc12/f4a84152-f222-46a2-9a35-8a0a0078fc12.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 912.540954] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-93fb3d16-c0a4-4bf0-b776-e2995ef54e48 tempest-ServerTagsTestJSON-2015212290 tempest-ServerTagsTestJSON-2015212290-project-member] [instance: f4a84152-f222-46a2-9a35-8a0a0078fc12] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 912.541276] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-bb385681-7907-4b8e-a543-2095e1b50e32 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.550560] env[62066]: DEBUG oslo_vmware.api [None req-93fb3d16-c0a4-4bf0-b776-e2995ef54e48 tempest-ServerTagsTestJSON-2015212290 tempest-ServerTagsTestJSON-2015212290-project-member] Waiting for the task: (returnval){ [ 912.550560] env[62066]: value = "task-1341112" [ 912.550560] env[62066]: _type = "Task" [ 912.550560] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.556097] env[62066]: DEBUG oslo_vmware.api [None req-36f2458b-83e5-4859-9178-b3e847334e66 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52a1188d-363d-51d1-82fb-6ffbf3eaf06e, 'name': SearchDatastore_Task, 'duration_secs': 0.008262} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.557201] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-57e8e390-92fb-44f0-abb8-191a5fcd7f5d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.565032] env[62066]: DEBUG oslo_vmware.api [None req-93fb3d16-c0a4-4bf0-b776-e2995ef54e48 tempest-ServerTagsTestJSON-2015212290 tempest-ServerTagsTestJSON-2015212290-project-member] Task: {'id': task-1341112, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.565640] env[62066]: DEBUG oslo_vmware.api [None req-36f2458b-83e5-4859-9178-b3e847334e66 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Waiting for the task: (returnval){ [ 912.565640] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]521c759b-ab20-5dc6-d6f0-e1a801183839" [ 912.565640] env[62066]: _type = "Task" [ 912.565640] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.573370] env[62066]: DEBUG oslo_vmware.api [None req-36f2458b-83e5-4859-9178-b3e847334e66 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]521c759b-ab20-5dc6-d6f0-e1a801183839, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.673879] env[62066]: DEBUG nova.compute.manager [req-546367ed-59d1-412e-972c-0d92a55109c0 req-37029eee-fe14-4691-bfd1-3fda4c23541e service nova] [instance: 8ff6fbb9-c90f-498d-9a85-d220a8c2f794] Received event network-vif-plugged-8c324833-7bf9-4007-85aa-5b16871f63c7 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 912.674111] env[62066]: DEBUG oslo_concurrency.lockutils [req-546367ed-59d1-412e-972c-0d92a55109c0 req-37029eee-fe14-4691-bfd1-3fda4c23541e service nova] Acquiring lock "8ff6fbb9-c90f-498d-9a85-d220a8c2f794-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 912.674327] env[62066]: DEBUG oslo_concurrency.lockutils [req-546367ed-59d1-412e-972c-0d92a55109c0 req-37029eee-fe14-4691-bfd1-3fda4c23541e service nova] Lock "8ff6fbb9-c90f-498d-9a85-d220a8c2f794-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 912.674514] env[62066]: DEBUG oslo_concurrency.lockutils [req-546367ed-59d1-412e-972c-0d92a55109c0 req-37029eee-fe14-4691-bfd1-3fda4c23541e service nova] Lock "8ff6fbb9-c90f-498d-9a85-d220a8c2f794-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 912.674723] env[62066]: DEBUG nova.compute.manager [req-546367ed-59d1-412e-972c-0d92a55109c0 req-37029eee-fe14-4691-bfd1-3fda4c23541e service nova] [instance: 8ff6fbb9-c90f-498d-9a85-d220a8c2f794] No waiting events found dispatching network-vif-plugged-8c324833-7bf9-4007-85aa-5b16871f63c7 {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 912.674869] env[62066]: WARNING nova.compute.manager [req-546367ed-59d1-412e-972c-0d92a55109c0 req-37029eee-fe14-4691-bfd1-3fda4c23541e service nova] [instance: 8ff6fbb9-c90f-498d-9a85-d220a8c2f794] Received unexpected event network-vif-plugged-8c324833-7bf9-4007-85aa-5b16871f63c7 for instance with vm_state shelved_offloaded and task_state spawning. [ 912.805857] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cb83f312-6ced-4dc9-b42a-28a22511dbae tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.768s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 912.808111] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ad6adfa4-e5e7-49bc-8c71-f4f4f133daea tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.579s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 912.809558] env[62066]: INFO nova.compute.claims [None req-ad6adfa4-e5e7-49bc-8c71-f4f4f133daea tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] [instance: 15afaea0-65de-4f32-851c-365003e5498f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 912.834182] env[62066]: INFO nova.scheduler.client.report [None req-cb83f312-6ced-4dc9-b42a-28a22511dbae tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Deleted allocations for instance 2634fcc0-96bd-4513-8f92-515f83fe23d7 [ 912.876846] env[62066]: DEBUG oslo_concurrency.lockutils [None req-81c18bf6-1c8d-45a0-a66c-dcda0761f479 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Acquiring lock "refresh_cache-8ff6fbb9-c90f-498d-9a85-d220a8c2f794" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 912.877134] env[62066]: DEBUG oslo_concurrency.lockutils [None req-81c18bf6-1c8d-45a0-a66c-dcda0761f479 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Acquired lock "refresh_cache-8ff6fbb9-c90f-498d-9a85-d220a8c2f794" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 912.877462] env[62066]: DEBUG nova.network.neutron [None req-81c18bf6-1c8d-45a0-a66c-dcda0761f479 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] [instance: 8ff6fbb9-c90f-498d-9a85-d220a8c2f794] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 912.945160] env[62066]: DEBUG oslo_concurrency.lockutils [None req-adc52c21-6d1b-4740-84d8-9cdb3fbc5d28 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Lock "90c33434-c127-450b-9ff0-75181b4ac385" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 23.274s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 913.060390] env[62066]: DEBUG oslo_vmware.api [None req-93fb3d16-c0a4-4bf0-b776-e2995ef54e48 tempest-ServerTagsTestJSON-2015212290 tempest-ServerTagsTestJSON-2015212290-project-member] Task: {'id': task-1341112, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.085067} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.060666] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-93fb3d16-c0a4-4bf0-b776-e2995ef54e48 tempest-ServerTagsTestJSON-2015212290 tempest-ServerTagsTestJSON-2015212290-project-member] [instance: f4a84152-f222-46a2-9a35-8a0a0078fc12] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 913.061472] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce0bbaae-9d7b-4c50-ab12-865c15ba4216 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.087339] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-93fb3d16-c0a4-4bf0-b776-e2995ef54e48 tempest-ServerTagsTestJSON-2015212290 tempest-ServerTagsTestJSON-2015212290-project-member] [instance: f4a84152-f222-46a2-9a35-8a0a0078fc12] Reconfiguring VM instance instance-00000058 to attach disk [datastore2] f4a84152-f222-46a2-9a35-8a0a0078fc12/f4a84152-f222-46a2-9a35-8a0a0078fc12.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 913.087610] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ceda86eb-9d85-411a-b182-33f95d331542 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.106610] env[62066]: DEBUG oslo_vmware.api [None req-36f2458b-83e5-4859-9178-b3e847334e66 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]521c759b-ab20-5dc6-d6f0-e1a801183839, 'name': SearchDatastore_Task, 'duration_secs': 0.02707} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.107679] env[62066]: DEBUG oslo_concurrency.lockutils [None req-36f2458b-83e5-4859-9178-b3e847334e66 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Releasing lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 913.107947] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-36f2458b-83e5-4859-9178-b3e847334e66 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk to [datastore2] 4e39b9b0-4ab0-4f48-8a3f-6bc1895597eb/4e39b9b0-4ab0-4f48-8a3f-6bc1895597eb.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 913.108274] env[62066]: DEBUG oslo_vmware.api [None req-93fb3d16-c0a4-4bf0-b776-e2995ef54e48 tempest-ServerTagsTestJSON-2015212290 tempest-ServerTagsTestJSON-2015212290-project-member] Waiting for the task: (returnval){ [ 913.108274] env[62066]: value = "task-1341113" [ 913.108274] env[62066]: _type = "Task" [ 913.108274] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.108465] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d8d0d61f-acb1-4a09-bb41-7923e846c4b2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.117723] env[62066]: DEBUG oslo_vmware.api [None req-93fb3d16-c0a4-4bf0-b776-e2995ef54e48 tempest-ServerTagsTestJSON-2015212290 tempest-ServerTagsTestJSON-2015212290-project-member] Task: {'id': task-1341113, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.118765] env[62066]: DEBUG oslo_vmware.api [None req-36f2458b-83e5-4859-9178-b3e847334e66 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Waiting for the task: (returnval){ [ 913.118765] env[62066]: value = "task-1341114" [ 913.118765] env[62066]: _type = "Task" [ 913.118765] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.127529] env[62066]: DEBUG oslo_vmware.api [None req-36f2458b-83e5-4859-9178-b3e847334e66 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': task-1341114, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.341960] env[62066]: DEBUG oslo_concurrency.lockutils [None req-cb83f312-6ced-4dc9-b42a-28a22511dbae tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Lock "2634fcc0-96bd-4513-8f92-515f83fe23d7" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 19.632s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 913.435543] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ab952be5-ba87-4d38-b219-8de60eb09bdf tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Acquiring lock "90c33434-c127-450b-9ff0-75181b4ac385" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 913.435808] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ab952be5-ba87-4d38-b219-8de60eb09bdf tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Lock "90c33434-c127-450b-9ff0-75181b4ac385" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 913.436044] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ab952be5-ba87-4d38-b219-8de60eb09bdf tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Acquiring lock "90c33434-c127-450b-9ff0-75181b4ac385-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 913.436247] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ab952be5-ba87-4d38-b219-8de60eb09bdf tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Lock "90c33434-c127-450b-9ff0-75181b4ac385-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 913.436418] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ab952be5-ba87-4d38-b219-8de60eb09bdf tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Lock "90c33434-c127-450b-9ff0-75181b4ac385-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 913.438587] env[62066]: INFO nova.compute.manager [None req-ab952be5-ba87-4d38-b219-8de60eb09bdf tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: 90c33434-c127-450b-9ff0-75181b4ac385] Terminating instance [ 913.440396] env[62066]: DEBUG nova.compute.manager [None req-ab952be5-ba87-4d38-b219-8de60eb09bdf tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: 90c33434-c127-450b-9ff0-75181b4ac385] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 913.440592] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-ab952be5-ba87-4d38-b219-8de60eb09bdf tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: 90c33434-c127-450b-9ff0-75181b4ac385] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 913.441500] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d7ccf7d-bd82-4a7f-8bb2-031cdcd00aa2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.449133] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab952be5-ba87-4d38-b219-8de60eb09bdf tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: 90c33434-c127-450b-9ff0-75181b4ac385] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 913.449370] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ecb654b9-17bc-4ffe-b6b8-c7a08e419890 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.455926] env[62066]: DEBUG oslo_vmware.api [None req-ab952be5-ba87-4d38-b219-8de60eb09bdf tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Waiting for the task: (returnval){ [ 913.455926] env[62066]: value = "task-1341115" [ 913.455926] env[62066]: _type = "Task" [ 913.455926] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.464288] env[62066]: DEBUG oslo_vmware.api [None req-ab952be5-ba87-4d38-b219-8de60eb09bdf tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341115, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.571472] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-dc7187c9-b41f-4691-8bae-33272b590d9f tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: c8e0d47c-4421-4e00-9183-206fceeabc40] Volume attach. Driver type: vmdk {{(pid=62066) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 913.571863] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-dc7187c9-b41f-4691-8bae-33272b590d9f tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: c8e0d47c-4421-4e00-9183-206fceeabc40] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-286099', 'volume_id': '338bd175-a96d-417e-a5b0-380b7d939a85', 'name': 'volume-338bd175-a96d-417e-a5b0-380b7d939a85', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'c8e0d47c-4421-4e00-9183-206fceeabc40', 'attached_at': '', 'detached_at': '', 'volume_id': '338bd175-a96d-417e-a5b0-380b7d939a85', 'serial': '338bd175-a96d-417e-a5b0-380b7d939a85'} {{(pid=62066) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 913.572957] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b535143-54b3-4aee-8d96-5bb0fa5f5d6c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.596476] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ba86fa8-2ef1-4d3f-ac29-9e5b162109ab {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.622647] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-dc7187c9-b41f-4691-8bae-33272b590d9f tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: c8e0d47c-4421-4e00-9183-206fceeabc40] Reconfiguring VM instance instance-0000004f to attach disk [datastore2] volume-338bd175-a96d-417e-a5b0-380b7d939a85/volume-338bd175-a96d-417e-a5b0-380b7d939a85.vmdk or device None with type thin {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 913.626377] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-176feb5b-5015-447d-9d78-bdd58db52e00 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.649559] env[62066]: DEBUG oslo_vmware.api [None req-93fb3d16-c0a4-4bf0-b776-e2995ef54e48 tempest-ServerTagsTestJSON-2015212290 tempest-ServerTagsTestJSON-2015212290-project-member] Task: {'id': task-1341113, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.654365] env[62066]: DEBUG oslo_vmware.api [None req-36f2458b-83e5-4859-9178-b3e847334e66 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': task-1341114, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.658021] env[62066]: DEBUG oslo_vmware.api [None req-dc7187c9-b41f-4691-8bae-33272b590d9f tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Waiting for the task: (returnval){ [ 913.658021] env[62066]: value = "task-1341116" [ 913.658021] env[62066]: _type = "Task" [ 913.658021] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.666521] env[62066]: DEBUG oslo_vmware.api [None req-dc7187c9-b41f-4691-8bae-33272b590d9f tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Task: {'id': task-1341116, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.780581] env[62066]: DEBUG nova.network.neutron [None req-81c18bf6-1c8d-45a0-a66c-dcda0761f479 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] [instance: 8ff6fbb9-c90f-498d-9a85-d220a8c2f794] Updating instance_info_cache with network_info: [{"id": "8c324833-7bf9-4007-85aa-5b16871f63c7", "address": "fa:16:3e:00:85:34", "network": {"id": "0e17b5e2-526e-4d4c-a463-5e34408044cf", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1588316661-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.184", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1f05d8a66d0c4a479e49a947e4b4bbff", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ffcecdaa-a7b8-49fc-9371-dbdb7744688e", "external-id": "nsx-vlan-transportzone-994", "segmentation_id": 994, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8c324833-7b", "ovs_interfaceid": "8c324833-7bf9-4007-85aa-5b16871f63c7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 913.897750] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b9bedc72-5782-477a-80b5-653c800704af tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Acquiring lock "5e96def2-0cbd-4bd9-93f4-6a365a0142b0" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 913.897750] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b9bedc72-5782-477a-80b5-653c800704af tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Lock "5e96def2-0cbd-4bd9-93f4-6a365a0142b0" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 913.897750] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b9bedc72-5782-477a-80b5-653c800704af tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Acquiring lock "5e96def2-0cbd-4bd9-93f4-6a365a0142b0-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 913.897750] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b9bedc72-5782-477a-80b5-653c800704af tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Lock "5e96def2-0cbd-4bd9-93f4-6a365a0142b0-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 913.897750] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b9bedc72-5782-477a-80b5-653c800704af tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Lock "5e96def2-0cbd-4bd9-93f4-6a365a0142b0-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 913.899833] env[62066]: INFO nova.compute.manager [None req-b9bedc72-5782-477a-80b5-653c800704af tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] [instance: 5e96def2-0cbd-4bd9-93f4-6a365a0142b0] Terminating instance [ 913.902721] env[62066]: DEBUG nova.compute.manager [None req-b9bedc72-5782-477a-80b5-653c800704af tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] [instance: 5e96def2-0cbd-4bd9-93f4-6a365a0142b0] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 913.902984] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-b9bedc72-5782-477a-80b5-653c800704af tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] [instance: 5e96def2-0cbd-4bd9-93f4-6a365a0142b0] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 913.903965] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ea5ff6a-1a3c-4872-8974-3c300529cb9f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.915024] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-b9bedc72-5782-477a-80b5-653c800704af tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] [instance: 5e96def2-0cbd-4bd9-93f4-6a365a0142b0] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 913.915024] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d6b42eca-7ea9-40e2-a09c-e362ff9a4079 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.924048] env[62066]: DEBUG oslo_vmware.api [None req-b9bedc72-5782-477a-80b5-653c800704af tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Waiting for the task: (returnval){ [ 913.924048] env[62066]: value = "task-1341117" [ 913.924048] env[62066]: _type = "Task" [ 913.924048] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.931321] env[62066]: DEBUG oslo_vmware.api [None req-b9bedc72-5782-477a-80b5-653c800704af tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Task: {'id': task-1341117, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.966812] env[62066]: DEBUG oslo_vmware.api [None req-ab952be5-ba87-4d38-b219-8de60eb09bdf tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341115, 'name': PowerOffVM_Task, 'duration_secs': 0.247229} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.967183] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab952be5-ba87-4d38-b219-8de60eb09bdf tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: 90c33434-c127-450b-9ff0-75181b4ac385] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 913.967381] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-ab952be5-ba87-4d38-b219-8de60eb09bdf tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: 90c33434-c127-450b-9ff0-75181b4ac385] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 913.969968] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8bdcd8fc-95da-4661-87bd-51dca80ff8cd {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.039275] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5dca9fda-a294-45bc-b896-9e2cdcd8f677 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.043677] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-ab952be5-ba87-4d38-b219-8de60eb09bdf tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: 90c33434-c127-450b-9ff0-75181b4ac385] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 914.043935] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-ab952be5-ba87-4d38-b219-8de60eb09bdf tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: 90c33434-c127-450b-9ff0-75181b4ac385] Deleting contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 914.044202] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-ab952be5-ba87-4d38-b219-8de60eb09bdf tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Deleting the datastore file [datastore2] 90c33434-c127-450b-9ff0-75181b4ac385 {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 914.044874] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4bb7c1a6-e45e-427b-b869-99f3e0b540c8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.049710] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebe3395f-6b1f-4992-bb66-e9acb368726f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.058030] env[62066]: DEBUG oslo_vmware.api [None req-ab952be5-ba87-4d38-b219-8de60eb09bdf tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Waiting for the task: (returnval){ [ 914.058030] env[62066]: value = "task-1341119" [ 914.058030] env[62066]: _type = "Task" [ 914.058030] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.090127] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbea54d1-6705-4722-93cf-a4fa62352ee6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.096324] env[62066]: DEBUG oslo_vmware.api [None req-ab952be5-ba87-4d38-b219-8de60eb09bdf tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341119, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.102492] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e47e12a-c5e9-4399-96c3-73b1773a47b4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.123442] env[62066]: DEBUG nova.compute.provider_tree [None req-ad6adfa4-e5e7-49bc-8c71-f4f4f133daea tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 914.140670] env[62066]: DEBUG oslo_vmware.api [None req-93fb3d16-c0a4-4bf0-b776-e2995ef54e48 tempest-ServerTagsTestJSON-2015212290 tempest-ServerTagsTestJSON-2015212290-project-member] Task: {'id': task-1341113, 'name': ReconfigVM_Task, 'duration_secs': 0.875952} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.145656] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-93fb3d16-c0a4-4bf0-b776-e2995ef54e48 tempest-ServerTagsTestJSON-2015212290 tempest-ServerTagsTestJSON-2015212290-project-member] [instance: f4a84152-f222-46a2-9a35-8a0a0078fc12] Reconfigured VM instance instance-00000058 to attach disk [datastore2] f4a84152-f222-46a2-9a35-8a0a0078fc12/f4a84152-f222-46a2-9a35-8a0a0078fc12.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 914.146775] env[62066]: DEBUG oslo_vmware.api [None req-36f2458b-83e5-4859-9178-b3e847334e66 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': task-1341114, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.796014} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.147012] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d826e6ff-663d-47f3-8e14-4f54ab2436d4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.148826] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-36f2458b-83e5-4859-9178-b3e847334e66 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk to [datastore2] 4e39b9b0-4ab0-4f48-8a3f-6bc1895597eb/4e39b9b0-4ab0-4f48-8a3f-6bc1895597eb.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 914.149109] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-36f2458b-83e5-4859-9178-b3e847334e66 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 4e39b9b0-4ab0-4f48-8a3f-6bc1895597eb] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 914.149404] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4d81ff7b-44f7-41db-9c6f-bd887ae0daa5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.158258] env[62066]: DEBUG oslo_vmware.api [None req-36f2458b-83e5-4859-9178-b3e847334e66 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Waiting for the task: (returnval){ [ 914.158258] env[62066]: value = "task-1341121" [ 914.158258] env[62066]: _type = "Task" [ 914.158258] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.158616] env[62066]: DEBUG oslo_vmware.api [None req-93fb3d16-c0a4-4bf0-b776-e2995ef54e48 tempest-ServerTagsTestJSON-2015212290 tempest-ServerTagsTestJSON-2015212290-project-member] Waiting for the task: (returnval){ [ 914.158616] env[62066]: value = "task-1341120" [ 914.158616] env[62066]: _type = "Task" [ 914.158616] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.174983] env[62066]: DEBUG oslo_vmware.api [None req-dc7187c9-b41f-4691-8bae-33272b590d9f tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Task: {'id': task-1341116, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.182423] env[62066]: DEBUG oslo_vmware.api [None req-36f2458b-83e5-4859-9178-b3e847334e66 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': task-1341121, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.182423] env[62066]: DEBUG oslo_vmware.api [None req-93fb3d16-c0a4-4bf0-b776-e2995ef54e48 tempest-ServerTagsTestJSON-2015212290 tempest-ServerTagsTestJSON-2015212290-project-member] Task: {'id': task-1341120, 'name': Rename_Task} progress is 10%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.283495] env[62066]: DEBUG oslo_concurrency.lockutils [None req-81c18bf6-1c8d-45a0-a66c-dcda0761f479 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Releasing lock "refresh_cache-8ff6fbb9-c90f-498d-9a85-d220a8c2f794" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 914.312792] env[62066]: DEBUG nova.virt.hardware [None req-81c18bf6-1c8d-45a0-a66c-dcda0761f479 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-23T13:40:41Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='12fd819bb36bea1b277763b23d533a4b',container_format='bare',created_at=2024-10-23T13:48:30Z,direct_url=,disk_format='vmdk',id=1ebd2ae1-0ae3-4486-822c-c75cafe27fe3,min_disk=1,min_ram=0,name='tempest-AttachVolumeShelveTestJSON-server-109117280-shelved',owner='1f05d8a66d0c4a479e49a947e4b4bbff',properties=ImageMetaProps,protected=,size=31661056,status='active',tags=,updated_at=2024-10-23T13:48:48Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 914.312998] env[62066]: DEBUG nova.virt.hardware [None req-81c18bf6-1c8d-45a0-a66c-dcda0761f479 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 914.313295] env[62066]: DEBUG nova.virt.hardware [None req-81c18bf6-1c8d-45a0-a66c-dcda0761f479 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 914.313371] env[62066]: DEBUG nova.virt.hardware [None req-81c18bf6-1c8d-45a0-a66c-dcda0761f479 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 914.313520] env[62066]: DEBUG nova.virt.hardware [None req-81c18bf6-1c8d-45a0-a66c-dcda0761f479 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 914.313675] env[62066]: DEBUG nova.virt.hardware [None req-81c18bf6-1c8d-45a0-a66c-dcda0761f479 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 914.313944] env[62066]: DEBUG nova.virt.hardware [None req-81c18bf6-1c8d-45a0-a66c-dcda0761f479 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 914.314119] env[62066]: DEBUG nova.virt.hardware [None req-81c18bf6-1c8d-45a0-a66c-dcda0761f479 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 914.314258] env[62066]: DEBUG nova.virt.hardware [None req-81c18bf6-1c8d-45a0-a66c-dcda0761f479 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 914.314431] env[62066]: DEBUG nova.virt.hardware [None req-81c18bf6-1c8d-45a0-a66c-dcda0761f479 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 914.314610] env[62066]: DEBUG nova.virt.hardware [None req-81c18bf6-1c8d-45a0-a66c-dcda0761f479 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 914.315537] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d85e9a6b-2f95-441e-a448-be045e0792f9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.324035] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11ca6254-6f90-4232-bac5-1367948f18e5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.340716] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-81c18bf6-1c8d-45a0-a66c-dcda0761f479 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] [instance: 8ff6fbb9-c90f-498d-9a85-d220a8c2f794] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:00:85:34', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ffcecdaa-a7b8-49fc-9371-dbdb7744688e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8c324833-7bf9-4007-85aa-5b16871f63c7', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 914.348076] env[62066]: DEBUG oslo.service.loopingcall [None req-81c18bf6-1c8d-45a0-a66c-dcda0761f479 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 914.348357] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8ff6fbb9-c90f-498d-9a85-d220a8c2f794] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 914.348577] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4befea35-3761-41f2-9bd4-120ee25a7613 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.366949] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 914.366949] env[62066]: value = "task-1341122" [ 914.366949] env[62066]: _type = "Task" [ 914.366949] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.374419] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1341122, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.432737] env[62066]: DEBUG oslo_vmware.api [None req-b9bedc72-5782-477a-80b5-653c800704af tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Task: {'id': task-1341117, 'name': PowerOffVM_Task, 'duration_secs': 0.209696} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.433036] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-b9bedc72-5782-477a-80b5-653c800704af tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] [instance: 5e96def2-0cbd-4bd9-93f4-6a365a0142b0] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 914.433267] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-b9bedc72-5782-477a-80b5-653c800704af tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] [instance: 5e96def2-0cbd-4bd9-93f4-6a365a0142b0] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 914.433543] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-472953de-eee7-410c-b008-b47bffc325c5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.492053] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-b9bedc72-5782-477a-80b5-653c800704af tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] [instance: 5e96def2-0cbd-4bd9-93f4-6a365a0142b0] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 914.492461] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-b9bedc72-5782-477a-80b5-653c800704af tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] [instance: 5e96def2-0cbd-4bd9-93f4-6a365a0142b0] Deleting contents of the VM from datastore datastore1 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 914.492679] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-b9bedc72-5782-477a-80b5-653c800704af tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Deleting the datastore file [datastore1] 5e96def2-0cbd-4bd9-93f4-6a365a0142b0 {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 914.492962] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3fd51f5b-ec28-4561-b135-2c63efa1f3e7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.499224] env[62066]: DEBUG oslo_vmware.api [None req-b9bedc72-5782-477a-80b5-653c800704af tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Waiting for the task: (returnval){ [ 914.499224] env[62066]: value = "task-1341124" [ 914.499224] env[62066]: _type = "Task" [ 914.499224] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.509075] env[62066]: DEBUG oslo_vmware.api [None req-b9bedc72-5782-477a-80b5-653c800704af tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Task: {'id': task-1341124, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.567834] env[62066]: DEBUG oslo_vmware.api [None req-ab952be5-ba87-4d38-b219-8de60eb09bdf tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341119, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.157901} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.568130] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-ab952be5-ba87-4d38-b219-8de60eb09bdf tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 914.568336] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-ab952be5-ba87-4d38-b219-8de60eb09bdf tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: 90c33434-c127-450b-9ff0-75181b4ac385] Deleted contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 914.568514] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-ab952be5-ba87-4d38-b219-8de60eb09bdf tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: 90c33434-c127-450b-9ff0-75181b4ac385] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 914.568687] env[62066]: INFO nova.compute.manager [None req-ab952be5-ba87-4d38-b219-8de60eb09bdf tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: 90c33434-c127-450b-9ff0-75181b4ac385] Took 1.13 seconds to destroy the instance on the hypervisor. [ 914.568928] env[62066]: DEBUG oslo.service.loopingcall [None req-ab952be5-ba87-4d38-b219-8de60eb09bdf tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 914.569165] env[62066]: DEBUG nova.compute.manager [-] [instance: 90c33434-c127-450b-9ff0-75181b4ac385] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 914.569261] env[62066]: DEBUG nova.network.neutron [-] [instance: 90c33434-c127-450b-9ff0-75181b4ac385] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 914.630867] env[62066]: DEBUG nova.scheduler.client.report [None req-ad6adfa4-e5e7-49bc-8c71-f4f4f133daea tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 914.672060] env[62066]: DEBUG oslo_vmware.api [None req-dc7187c9-b41f-4691-8bae-33272b590d9f tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Task: {'id': task-1341116, 'name': ReconfigVM_Task, 'duration_secs': 0.684327} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.673569] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-dc7187c9-b41f-4691-8bae-33272b590d9f tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: c8e0d47c-4421-4e00-9183-206fceeabc40] Reconfigured VM instance instance-0000004f to attach disk [datastore2] volume-338bd175-a96d-417e-a5b0-380b7d939a85/volume-338bd175-a96d-417e-a5b0-380b7d939a85.vmdk or device None with type thin {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 914.683902] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1b609a6e-9347-4082-a1b1-211bef85709f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.693819] env[62066]: DEBUG oslo_vmware.api [None req-36f2458b-83e5-4859-9178-b3e847334e66 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': task-1341121, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.085391} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.694164] env[62066]: DEBUG oslo_vmware.api [None req-93fb3d16-c0a4-4bf0-b776-e2995ef54e48 tempest-ServerTagsTestJSON-2015212290 tempest-ServerTagsTestJSON-2015212290-project-member] Task: {'id': task-1341120, 'name': Rename_Task, 'duration_secs': 0.148173} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.696726] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-36f2458b-83e5-4859-9178-b3e847334e66 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 4e39b9b0-4ab0-4f48-8a3f-6bc1895597eb] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 914.696726] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-93fb3d16-c0a4-4bf0-b776-e2995ef54e48 tempest-ServerTagsTestJSON-2015212290 tempest-ServerTagsTestJSON-2015212290-project-member] [instance: f4a84152-f222-46a2-9a35-8a0a0078fc12] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 914.696726] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b03f916f-0d25-43f6-b3a7-ea29e2840e83 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.698560] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4c2be1aa-1d39-4ed6-94f3-a9701b4ce042 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.702017] env[62066]: DEBUG oslo_vmware.api [None req-dc7187c9-b41f-4691-8bae-33272b590d9f tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Waiting for the task: (returnval){ [ 914.702017] env[62066]: value = "task-1341125" [ 914.702017] env[62066]: _type = "Task" [ 914.702017] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.708332] env[62066]: DEBUG nova.compute.manager [req-c1938e26-2efd-46f2-99b5-7de7b277e80d req-b1bd0bbd-be66-4bea-aab0-b523c7f8995e service nova] [instance: 8ff6fbb9-c90f-498d-9a85-d220a8c2f794] Received event network-changed-8c324833-7bf9-4007-85aa-5b16871f63c7 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 914.708520] env[62066]: DEBUG nova.compute.manager [req-c1938e26-2efd-46f2-99b5-7de7b277e80d req-b1bd0bbd-be66-4bea-aab0-b523c7f8995e service nova] [instance: 8ff6fbb9-c90f-498d-9a85-d220a8c2f794] Refreshing instance network info cache due to event network-changed-8c324833-7bf9-4007-85aa-5b16871f63c7. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 914.708776] env[62066]: DEBUG oslo_concurrency.lockutils [req-c1938e26-2efd-46f2-99b5-7de7b277e80d req-b1bd0bbd-be66-4bea-aab0-b523c7f8995e service nova] Acquiring lock "refresh_cache-8ff6fbb9-c90f-498d-9a85-d220a8c2f794" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 914.708934] env[62066]: DEBUG oslo_concurrency.lockutils [req-c1938e26-2efd-46f2-99b5-7de7b277e80d req-b1bd0bbd-be66-4bea-aab0-b523c7f8995e service nova] Acquired lock "refresh_cache-8ff6fbb9-c90f-498d-9a85-d220a8c2f794" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 914.709152] env[62066]: DEBUG nova.network.neutron [req-c1938e26-2efd-46f2-99b5-7de7b277e80d req-b1bd0bbd-be66-4bea-aab0-b523c7f8995e service nova] [instance: 8ff6fbb9-c90f-498d-9a85-d220a8c2f794] Refreshing network info cache for port 8c324833-7bf9-4007-85aa-5b16871f63c7 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 914.735133] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-36f2458b-83e5-4859-9178-b3e847334e66 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 4e39b9b0-4ab0-4f48-8a3f-6bc1895597eb] Reconfiguring VM instance instance-00000059 to attach disk [datastore2] 4e39b9b0-4ab0-4f48-8a3f-6bc1895597eb/4e39b9b0-4ab0-4f48-8a3f-6bc1895597eb.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 914.735526] env[62066]: DEBUG oslo_vmware.api [None req-93fb3d16-c0a4-4bf0-b776-e2995ef54e48 tempest-ServerTagsTestJSON-2015212290 tempest-ServerTagsTestJSON-2015212290-project-member] Waiting for the task: (returnval){ [ 914.735526] env[62066]: value = "task-1341126" [ 914.735526] env[62066]: _type = "Task" [ 914.735526] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.735736] env[62066]: DEBUG oslo_vmware.api [None req-dc7187c9-b41f-4691-8bae-33272b590d9f tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Task: {'id': task-1341125, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.737066] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7f5dbefc-bee2-4a1f-a0bf-247fcbee355a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.761032] env[62066]: DEBUG oslo_vmware.api [None req-93fb3d16-c0a4-4bf0-b776-e2995ef54e48 tempest-ServerTagsTestJSON-2015212290 tempest-ServerTagsTestJSON-2015212290-project-member] Task: {'id': task-1341126, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.762436] env[62066]: DEBUG oslo_vmware.api [None req-36f2458b-83e5-4859-9178-b3e847334e66 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Waiting for the task: (returnval){ [ 914.762436] env[62066]: value = "task-1341127" [ 914.762436] env[62066]: _type = "Task" [ 914.762436] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.770950] env[62066]: DEBUG oslo_vmware.api [None req-36f2458b-83e5-4859-9178-b3e847334e66 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': task-1341127, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.879503] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1341122, 'name': CreateVM_Task} progress is 99%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.013561] env[62066]: DEBUG oslo_vmware.api [None req-b9bedc72-5782-477a-80b5-653c800704af tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Task: {'id': task-1341124, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.184684} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.013967] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-b9bedc72-5782-477a-80b5-653c800704af tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 915.014312] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-b9bedc72-5782-477a-80b5-653c800704af tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] [instance: 5e96def2-0cbd-4bd9-93f4-6a365a0142b0] Deleted contents of the VM from datastore datastore1 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 915.014718] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-b9bedc72-5782-477a-80b5-653c800704af tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] [instance: 5e96def2-0cbd-4bd9-93f4-6a365a0142b0] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 915.014935] env[62066]: INFO nova.compute.manager [None req-b9bedc72-5782-477a-80b5-653c800704af tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] [instance: 5e96def2-0cbd-4bd9-93f4-6a365a0142b0] Took 1.11 seconds to destroy the instance on the hypervisor. [ 915.015334] env[62066]: DEBUG oslo.service.loopingcall [None req-b9bedc72-5782-477a-80b5-653c800704af tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 915.015642] env[62066]: DEBUG nova.compute.manager [-] [instance: 5e96def2-0cbd-4bd9-93f4-6a365a0142b0] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 915.015791] env[62066]: DEBUG nova.network.neutron [-] [instance: 5e96def2-0cbd-4bd9-93f4-6a365a0142b0] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 915.138968] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ad6adfa4-e5e7-49bc-8c71-f4f4f133daea tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.331s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 915.139545] env[62066]: DEBUG nova.compute.manager [None req-ad6adfa4-e5e7-49bc-8c71-f4f4f133daea tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] [instance: 15afaea0-65de-4f32-851c-365003e5498f] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 915.142750] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3ba3ebd2-6915-4f7b-8349-9384d9a8c223 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 15.527s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 915.143015] env[62066]: DEBUG nova.objects.instance [None req-3ba3ebd2-6915-4f7b-8349-9384d9a8c223 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Lazy-loading 'resources' on Instance uuid 8807ce4d-532f-469c-a302-464c61c7efeb {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 915.212066] env[62066]: DEBUG oslo_vmware.api [None req-dc7187c9-b41f-4691-8bae-33272b590d9f tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Task: {'id': task-1341125, 'name': ReconfigVM_Task, 'duration_secs': 0.183287} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.212807] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-dc7187c9-b41f-4691-8bae-33272b590d9f tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: c8e0d47c-4421-4e00-9183-206fceeabc40] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-286099', 'volume_id': '338bd175-a96d-417e-a5b0-380b7d939a85', 'name': 'volume-338bd175-a96d-417e-a5b0-380b7d939a85', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'c8e0d47c-4421-4e00-9183-206fceeabc40', 'attached_at': '', 'detached_at': '', 'volume_id': '338bd175-a96d-417e-a5b0-380b7d939a85', 'serial': '338bd175-a96d-417e-a5b0-380b7d939a85'} {{(pid=62066) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 915.260557] env[62066]: DEBUG oslo_vmware.api [None req-93fb3d16-c0a4-4bf0-b776-e2995ef54e48 tempest-ServerTagsTestJSON-2015212290 tempest-ServerTagsTestJSON-2015212290-project-member] Task: {'id': task-1341126, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.272568] env[62066]: DEBUG oslo_vmware.api [None req-36f2458b-83e5-4859-9178-b3e847334e66 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': task-1341127, 'name': ReconfigVM_Task, 'duration_secs': 0.295627} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.272690] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-36f2458b-83e5-4859-9178-b3e847334e66 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 4e39b9b0-4ab0-4f48-8a3f-6bc1895597eb] Reconfigured VM instance instance-00000059 to attach disk [datastore2] 4e39b9b0-4ab0-4f48-8a3f-6bc1895597eb/4e39b9b0-4ab0-4f48-8a3f-6bc1895597eb.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 915.273772] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0aaff813-7d97-41dc-89d4-22fc12bc2393 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.279808] env[62066]: DEBUG oslo_vmware.api [None req-36f2458b-83e5-4859-9178-b3e847334e66 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Waiting for the task: (returnval){ [ 915.279808] env[62066]: value = "task-1341128" [ 915.279808] env[62066]: _type = "Task" [ 915.279808] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.289515] env[62066]: DEBUG oslo_vmware.api [None req-36f2458b-83e5-4859-9178-b3e847334e66 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': task-1341128, 'name': Rename_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.377390] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1341122, 'name': CreateVM_Task, 'duration_secs': 0.513705} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.377561] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8ff6fbb9-c90f-498d-9a85-d220a8c2f794] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 915.378276] env[62066]: DEBUG oslo_concurrency.lockutils [None req-81c18bf6-1c8d-45a0-a66c-dcda0761f479 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1ebd2ae1-0ae3-4486-822c-c75cafe27fe3" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 915.378447] env[62066]: DEBUG oslo_concurrency.lockutils [None req-81c18bf6-1c8d-45a0-a66c-dcda0761f479 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1ebd2ae1-0ae3-4486-822c-c75cafe27fe3" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 915.378846] env[62066]: DEBUG oslo_concurrency.lockutils [None req-81c18bf6-1c8d-45a0-a66c-dcda0761f479 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/1ebd2ae1-0ae3-4486-822c-c75cafe27fe3" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 915.379146] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2ccdd754-df2d-4a11-b125-e9752013028b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.384222] env[62066]: DEBUG oslo_vmware.api [None req-81c18bf6-1c8d-45a0-a66c-dcda0761f479 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Waiting for the task: (returnval){ [ 915.384222] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]52124ae6-5eb1-b476-9d39-1b84cf554b8b" [ 915.384222] env[62066]: _type = "Task" [ 915.384222] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.387272] env[62066]: DEBUG nova.network.neutron [-] [instance: 90c33434-c127-450b-9ff0-75181b4ac385] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 915.392990] env[62066]: DEBUG oslo_vmware.api [None req-81c18bf6-1c8d-45a0-a66c-dcda0761f479 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52124ae6-5eb1-b476-9d39-1b84cf554b8b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.543891] env[62066]: DEBUG nova.network.neutron [req-c1938e26-2efd-46f2-99b5-7de7b277e80d req-b1bd0bbd-be66-4bea-aab0-b523c7f8995e service nova] [instance: 8ff6fbb9-c90f-498d-9a85-d220a8c2f794] Updated VIF entry in instance network info cache for port 8c324833-7bf9-4007-85aa-5b16871f63c7. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 915.543891] env[62066]: DEBUG nova.network.neutron [req-c1938e26-2efd-46f2-99b5-7de7b277e80d req-b1bd0bbd-be66-4bea-aab0-b523c7f8995e service nova] [instance: 8ff6fbb9-c90f-498d-9a85-d220a8c2f794] Updating instance_info_cache with network_info: [{"id": "8c324833-7bf9-4007-85aa-5b16871f63c7", "address": "fa:16:3e:00:85:34", "network": {"id": "0e17b5e2-526e-4d4c-a463-5e34408044cf", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1588316661-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.184", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1f05d8a66d0c4a479e49a947e4b4bbff", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ffcecdaa-a7b8-49fc-9371-dbdb7744688e", "external-id": "nsx-vlan-transportzone-994", "segmentation_id": 994, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8c324833-7b", "ovs_interfaceid": "8c324833-7bf9-4007-85aa-5b16871f63c7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 915.646716] env[62066]: DEBUG nova.compute.utils [None req-ad6adfa4-e5e7-49bc-8c71-f4f4f133daea tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 915.651433] env[62066]: DEBUG nova.compute.manager [None req-ad6adfa4-e5e7-49bc-8c71-f4f4f133daea tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] [instance: 15afaea0-65de-4f32-851c-365003e5498f] Not allocating networking since 'none' was specified. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 915.733484] env[62066]: DEBUG nova.network.neutron [-] [instance: 5e96def2-0cbd-4bd9-93f4-6a365a0142b0] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 915.762214] env[62066]: DEBUG oslo_vmware.api [None req-93fb3d16-c0a4-4bf0-b776-e2995ef54e48 tempest-ServerTagsTestJSON-2015212290 tempest-ServerTagsTestJSON-2015212290-project-member] Task: {'id': task-1341126, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.789648] env[62066]: DEBUG oslo_vmware.api [None req-36f2458b-83e5-4859-9178-b3e847334e66 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': task-1341128, 'name': Rename_Task, 'duration_secs': 0.131453} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.792303] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-36f2458b-83e5-4859-9178-b3e847334e66 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 4e39b9b0-4ab0-4f48-8a3f-6bc1895597eb] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 915.792756] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6bf5016f-fd57-4731-9a23-e0287f0daf0c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.800572] env[62066]: DEBUG oslo_vmware.api [None req-36f2458b-83e5-4859-9178-b3e847334e66 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Waiting for the task: (returnval){ [ 915.800572] env[62066]: value = "task-1341129" [ 915.800572] env[62066]: _type = "Task" [ 915.800572] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.810360] env[62066]: DEBUG oslo_vmware.api [None req-36f2458b-83e5-4859-9178-b3e847334e66 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': task-1341129, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.859426] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8eff841-1cf6-420b-8157-8017a1af91ea {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.867983] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40129ba6-2d39-4277-9063-b34788e06d7a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.904194] env[62066]: INFO nova.compute.manager [-] [instance: 90c33434-c127-450b-9ff0-75181b4ac385] Took 1.33 seconds to deallocate network for instance. [ 915.909956] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a67b08db-8575-4198-bcb3-9c62d308ead3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.922397] env[62066]: DEBUG oslo_concurrency.lockutils [None req-81c18bf6-1c8d-45a0-a66c-dcda0761f479 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1ebd2ae1-0ae3-4486-822c-c75cafe27fe3" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 915.922657] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-81c18bf6-1c8d-45a0-a66c-dcda0761f479 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] [instance: 8ff6fbb9-c90f-498d-9a85-d220a8c2f794] Processing image 1ebd2ae1-0ae3-4486-822c-c75cafe27fe3 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 915.922894] env[62066]: DEBUG oslo_concurrency.lockutils [None req-81c18bf6-1c8d-45a0-a66c-dcda0761f479 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/1ebd2ae1-0ae3-4486-822c-c75cafe27fe3/1ebd2ae1-0ae3-4486-822c-c75cafe27fe3.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 915.923063] env[62066]: DEBUG oslo_concurrency.lockutils [None req-81c18bf6-1c8d-45a0-a66c-dcda0761f479 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Acquired lock "[datastore1] devstack-image-cache_base/1ebd2ae1-0ae3-4486-822c-c75cafe27fe3/1ebd2ae1-0ae3-4486-822c-c75cafe27fe3.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 915.923243] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-81c18bf6-1c8d-45a0-a66c-dcda0761f479 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 915.923537] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-bff2fd9b-469d-4dc5-97de-dad69027fbef {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.926569] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a04a26e7-9596-44fa-a22e-fb1ec68b6174 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.940986] env[62066]: DEBUG nova.compute.provider_tree [None req-3ba3ebd2-6915-4f7b-8349-9384d9a8c223 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 915.943236] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-81c18bf6-1c8d-45a0-a66c-dcda0761f479 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 915.944116] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-81c18bf6-1c8d-45a0-a66c-dcda0761f479 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 915.944332] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b8f4453e-d553-4b59-a8bc-f0a12a3d94cb {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.949503] env[62066]: DEBUG oslo_vmware.api [None req-81c18bf6-1c8d-45a0-a66c-dcda0761f479 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Waiting for the task: (returnval){ [ 915.949503] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]527a05de-94bf-8576-0133-a84bc9f471fe" [ 915.949503] env[62066]: _type = "Task" [ 915.949503] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.958992] env[62066]: DEBUG oslo_vmware.api [None req-81c18bf6-1c8d-45a0-a66c-dcda0761f479 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]527a05de-94bf-8576-0133-a84bc9f471fe, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.045919] env[62066]: DEBUG oslo_concurrency.lockutils [req-c1938e26-2efd-46f2-99b5-7de7b277e80d req-b1bd0bbd-be66-4bea-aab0-b523c7f8995e service nova] Releasing lock "refresh_cache-8ff6fbb9-c90f-498d-9a85-d220a8c2f794" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 916.153296] env[62066]: DEBUG nova.compute.manager [None req-ad6adfa4-e5e7-49bc-8c71-f4f4f133daea tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] [instance: 15afaea0-65de-4f32-851c-365003e5498f] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 916.237834] env[62066]: INFO nova.compute.manager [-] [instance: 5e96def2-0cbd-4bd9-93f4-6a365a0142b0] Took 1.22 seconds to deallocate network for instance. [ 916.250373] env[62066]: DEBUG nova.objects.instance [None req-dc7187c9-b41f-4691-8bae-33272b590d9f tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Lazy-loading 'flavor' on Instance uuid c8e0d47c-4421-4e00-9183-206fceeabc40 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 916.261550] env[62066]: DEBUG oslo_vmware.api [None req-93fb3d16-c0a4-4bf0-b776-e2995ef54e48 tempest-ServerTagsTestJSON-2015212290 tempest-ServerTagsTestJSON-2015212290-project-member] Task: {'id': task-1341126, 'name': PowerOnVM_Task, 'duration_secs': 1.222755} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.262378] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-93fb3d16-c0a4-4bf0-b776-e2995ef54e48 tempest-ServerTagsTestJSON-2015212290 tempest-ServerTagsTestJSON-2015212290-project-member] [instance: f4a84152-f222-46a2-9a35-8a0a0078fc12] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 916.262593] env[62066]: INFO nova.compute.manager [None req-93fb3d16-c0a4-4bf0-b776-e2995ef54e48 tempest-ServerTagsTestJSON-2015212290 tempest-ServerTagsTestJSON-2015212290-project-member] [instance: f4a84152-f222-46a2-9a35-8a0a0078fc12] Took 10.63 seconds to spawn the instance on the hypervisor. [ 916.262775] env[62066]: DEBUG nova.compute.manager [None req-93fb3d16-c0a4-4bf0-b776-e2995ef54e48 tempest-ServerTagsTestJSON-2015212290 tempest-ServerTagsTestJSON-2015212290-project-member] [instance: f4a84152-f222-46a2-9a35-8a0a0078fc12] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 916.263565] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cee1f01c-b5cf-4971-9a96-dd94bcb699e2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.310464] env[62066]: DEBUG oslo_vmware.api [None req-36f2458b-83e5-4859-9178-b3e847334e66 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': task-1341129, 'name': PowerOnVM_Task} progress is 87%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.417233] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ab952be5-ba87-4d38-b219-8de60eb09bdf tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 916.445863] env[62066]: DEBUG nova.scheduler.client.report [None req-3ba3ebd2-6915-4f7b-8349-9384d9a8c223 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 916.459468] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-81c18bf6-1c8d-45a0-a66c-dcda0761f479 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] [instance: 8ff6fbb9-c90f-498d-9a85-d220a8c2f794] Preparing fetch location {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 916.459763] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-81c18bf6-1c8d-45a0-a66c-dcda0761f479 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] [instance: 8ff6fbb9-c90f-498d-9a85-d220a8c2f794] Fetch image to [datastore1] OSTACK_IMG_7d0cd4f2-d1f4-4ceb-9c76-75a77646fd93/OSTACK_IMG_7d0cd4f2-d1f4-4ceb-9c76-75a77646fd93.vmdk {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 916.459992] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-81c18bf6-1c8d-45a0-a66c-dcda0761f479 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] [instance: 8ff6fbb9-c90f-498d-9a85-d220a8c2f794] Downloading stream optimized image 1ebd2ae1-0ae3-4486-822c-c75cafe27fe3 to [datastore1] OSTACK_IMG_7d0cd4f2-d1f4-4ceb-9c76-75a77646fd93/OSTACK_IMG_7d0cd4f2-d1f4-4ceb-9c76-75a77646fd93.vmdk on the data store datastore1 as vApp {{(pid=62066) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 916.460202] env[62066]: DEBUG nova.virt.vmwareapi.images [None req-81c18bf6-1c8d-45a0-a66c-dcda0761f479 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] [instance: 8ff6fbb9-c90f-498d-9a85-d220a8c2f794] Downloading image file data 1ebd2ae1-0ae3-4486-822c-c75cafe27fe3 to the ESX as VM named 'OSTACK_IMG_7d0cd4f2-d1f4-4ceb-9c76-75a77646fd93' {{(pid=62066) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 916.538472] env[62066]: DEBUG oslo_vmware.rw_handles [None req-81c18bf6-1c8d-45a0-a66c-dcda0761f479 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 916.538472] env[62066]: value = "resgroup-9" [ 916.538472] env[62066]: _type = "ResourcePool" [ 916.538472] env[62066]: }. {{(pid=62066) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 916.538790] env[62066]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-beb728e3-700e-407f-bcb7-4e0825d020ea {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.559810] env[62066]: DEBUG oslo_vmware.rw_handles [None req-81c18bf6-1c8d-45a0-a66c-dcda0761f479 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Lease: (returnval){ [ 916.559810] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]525e97f8-5b6e-585c-f9ed-54c63164d125" [ 916.559810] env[62066]: _type = "HttpNfcLease" [ 916.559810] env[62066]: } obtained for vApp import into resource pool (val){ [ 916.559810] env[62066]: value = "resgroup-9" [ 916.559810] env[62066]: _type = "ResourcePool" [ 916.559810] env[62066]: }. {{(pid=62066) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 916.560137] env[62066]: DEBUG oslo_vmware.api [None req-81c18bf6-1c8d-45a0-a66c-dcda0761f479 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Waiting for the lease: (returnval){ [ 916.560137] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]525e97f8-5b6e-585c-f9ed-54c63164d125" [ 916.560137] env[62066]: _type = "HttpNfcLease" [ 916.560137] env[62066]: } to be ready. {{(pid=62066) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 916.568346] env[62066]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 916.568346] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]525e97f8-5b6e-585c-f9ed-54c63164d125" [ 916.568346] env[62066]: _type = "HttpNfcLease" [ 916.568346] env[62066]: } is initializing. {{(pid=62066) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 916.744923] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b9bedc72-5782-477a-80b5-653c800704af tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 916.755476] env[62066]: DEBUG oslo_concurrency.lockutils [None req-dc7187c9-b41f-4691-8bae-33272b590d9f tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Lock "c8e0d47c-4421-4e00-9183-206fceeabc40" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.815s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 916.772220] env[62066]: DEBUG nova.compute.manager [req-ca5bc5ba-c873-44a6-824e-4ffd0894ba4b req-2b2d6efd-7ab0-43ad-839e-b2884c9ae580 service nova] [instance: 90c33434-c127-450b-9ff0-75181b4ac385] Received event network-vif-deleted-1b4bbccc-7650-4977-9133-da7bda63ceeb {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 916.772476] env[62066]: DEBUG nova.compute.manager [req-ca5bc5ba-c873-44a6-824e-4ffd0894ba4b req-2b2d6efd-7ab0-43ad-839e-b2884c9ae580 service nova] [instance: 5e96def2-0cbd-4bd9-93f4-6a365a0142b0] Received event network-vif-deleted-0932c7ac-630e-4cef-8991-45b438f39483 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 916.779255] env[62066]: INFO nova.compute.manager [None req-93fb3d16-c0a4-4bf0-b776-e2995ef54e48 tempest-ServerTagsTestJSON-2015212290 tempest-ServerTagsTestJSON-2015212290-project-member] [instance: f4a84152-f222-46a2-9a35-8a0a0078fc12] Took 26.10 seconds to build instance. [ 916.812294] env[62066]: DEBUG oslo_vmware.api [None req-36f2458b-83e5-4859-9178-b3e847334e66 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': task-1341129, 'name': PowerOnVM_Task, 'duration_secs': 0.712252} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.812294] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-36f2458b-83e5-4859-9178-b3e847334e66 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 4e39b9b0-4ab0-4f48-8a3f-6bc1895597eb] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 916.812294] env[62066]: INFO nova.compute.manager [None req-36f2458b-83e5-4859-9178-b3e847334e66 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 4e39b9b0-4ab0-4f48-8a3f-6bc1895597eb] Took 8.89 seconds to spawn the instance on the hypervisor. [ 916.812294] env[62066]: DEBUG nova.compute.manager [None req-36f2458b-83e5-4859-9178-b3e847334e66 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 4e39b9b0-4ab0-4f48-8a3f-6bc1895597eb] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 916.812704] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7aaf7c33-e13e-4baa-8294-7f23cea1ef92 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.954845] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3ba3ebd2-6915-4f7b-8349-9384d9a8c223 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.812s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 916.956934] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 10.062s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 916.957161] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 916.957327] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62066) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 916.957639] env[62066]: DEBUG oslo_concurrency.lockutils [None req-740993dc-fc35-4aa8-9a06-8f4c0e5f95e1 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 7.070s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 916.959601] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf2c26da-2324-459c-b23a-306c89e78a10 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.969255] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ba498ae-d23a-4def-87b8-f229a29bdc0e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.985224] env[62066]: INFO nova.scheduler.client.report [None req-3ba3ebd2-6915-4f7b-8349-9384d9a8c223 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Deleted allocations for instance 8807ce4d-532f-469c-a302-464c61c7efeb [ 916.986618] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aefba031-57ab-4299-8ca4-af31b7a66492 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.996195] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73311245-ccbc-4ba9-9de1-ad64980fe736 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.029146] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180149MB free_disk=153GB free_vcpus=48 pci_devices=None {{(pid=62066) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 917.029319] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 917.067519] env[62066]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 917.067519] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]525e97f8-5b6e-585c-f9ed-54c63164d125" [ 917.067519] env[62066]: _type = "HttpNfcLease" [ 917.067519] env[62066]: } is initializing. {{(pid=62066) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 917.163028] env[62066]: DEBUG nova.compute.manager [None req-ad6adfa4-e5e7-49bc-8c71-f4f4f133daea tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] [instance: 15afaea0-65de-4f32-851c-365003e5498f] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 917.185775] env[62066]: DEBUG nova.virt.hardware [None req-ad6adfa4-e5e7-49bc-8c71-f4f4f133daea tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-23T13:40:41Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-23T13:40:23Z,direct_url=,disk_format='vmdk',id=50ff584c-3b50-4395-af07-3e66769bc9f7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='eb52f7069a374c61ae946f052007c6d9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-23T13:40:24Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 917.186055] env[62066]: DEBUG nova.virt.hardware [None req-ad6adfa4-e5e7-49bc-8c71-f4f4f133daea tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 917.186228] env[62066]: DEBUG nova.virt.hardware [None req-ad6adfa4-e5e7-49bc-8c71-f4f4f133daea tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 917.186434] env[62066]: DEBUG nova.virt.hardware [None req-ad6adfa4-e5e7-49bc-8c71-f4f4f133daea tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 917.186603] env[62066]: DEBUG nova.virt.hardware [None req-ad6adfa4-e5e7-49bc-8c71-f4f4f133daea tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 917.186767] env[62066]: DEBUG nova.virt.hardware [None req-ad6adfa4-e5e7-49bc-8c71-f4f4f133daea tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 917.186987] env[62066]: DEBUG nova.virt.hardware [None req-ad6adfa4-e5e7-49bc-8c71-f4f4f133daea tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 917.187759] env[62066]: DEBUG nova.virt.hardware [None req-ad6adfa4-e5e7-49bc-8c71-f4f4f133daea tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 917.188226] env[62066]: DEBUG nova.virt.hardware [None req-ad6adfa4-e5e7-49bc-8c71-f4f4f133daea tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 917.188487] env[62066]: DEBUG nova.virt.hardware [None req-ad6adfa4-e5e7-49bc-8c71-f4f4f133daea tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 917.188683] env[62066]: DEBUG nova.virt.hardware [None req-ad6adfa4-e5e7-49bc-8c71-f4f4f133daea tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 917.189558] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cf920de-e87d-4f49-93ab-5ce41ccb2960 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.197764] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e073552d-f859-4ed9-b1a4-c97849877bd4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.211329] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-ad6adfa4-e5e7-49bc-8c71-f4f4f133daea tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] [instance: 15afaea0-65de-4f32-851c-365003e5498f] Instance VIF info [] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 917.217443] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-ad6adfa4-e5e7-49bc-8c71-f4f4f133daea tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Creating folder: Project (f3d429a303724c5a8cce22c7a564f856). Parent ref: group-v285980. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 917.217719] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4b2bd5cd-8ef7-4d27-b787-15854eba3c1f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.226542] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-ad6adfa4-e5e7-49bc-8c71-f4f4f133daea tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Created folder: Project (f3d429a303724c5a8cce22c7a564f856) in parent group-v285980. [ 917.226732] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-ad6adfa4-e5e7-49bc-8c71-f4f4f133daea tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Creating folder: Instances. Parent ref: group-v286103. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 917.226952] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3a8a38fb-bffc-4596-9c0b-756b453d0526 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.234746] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-ad6adfa4-e5e7-49bc-8c71-f4f4f133daea tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Created folder: Instances in parent group-v286103. [ 917.234972] env[62066]: DEBUG oslo.service.loopingcall [None req-ad6adfa4-e5e7-49bc-8c71-f4f4f133daea tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 917.235183] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 15afaea0-65de-4f32-851c-365003e5498f] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 917.235387] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-56e034e9-2c1f-488a-b728-6ee3e802511c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.250586] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 917.250586] env[62066]: value = "task-1341133" [ 917.250586] env[62066]: _type = "Task" [ 917.250586] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.257672] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1341133, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.281279] env[62066]: DEBUG oslo_concurrency.lockutils [None req-93fb3d16-c0a4-4bf0-b776-e2995ef54e48 tempest-ServerTagsTestJSON-2015212290 tempest-ServerTagsTestJSON-2015212290-project-member] Lock "f4a84152-f222-46a2-9a35-8a0a0078fc12" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 27.615s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 917.330463] env[62066]: INFO nova.compute.manager [None req-36f2458b-83e5-4859-9178-b3e847334e66 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 4e39b9b0-4ab0-4f48-8a3f-6bc1895597eb] Took 23.83 seconds to build instance. [ 917.395036] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1b0d97e8-08d8-404b-b2dc-335d2f49fb47 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Acquiring lock "c8e0d47c-4421-4e00-9183-206fceeabc40" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 917.395407] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1b0d97e8-08d8-404b-b2dc-335d2f49fb47 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Lock "c8e0d47c-4421-4e00-9183-206fceeabc40" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 917.395642] env[62066]: DEBUG nova.compute.manager [None req-1b0d97e8-08d8-404b-b2dc-335d2f49fb47 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: c8e0d47c-4421-4e00-9183-206fceeabc40] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 917.396626] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c0342aa-968a-4b8f-8a5c-57ce02a57bd3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.404499] env[62066]: DEBUG nova.compute.manager [None req-1b0d97e8-08d8-404b-b2dc-335d2f49fb47 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: c8e0d47c-4421-4e00-9183-206fceeabc40] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62066) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 917.405123] env[62066]: DEBUG nova.objects.instance [None req-1b0d97e8-08d8-404b-b2dc-335d2f49fb47 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Lazy-loading 'flavor' on Instance uuid c8e0d47c-4421-4e00-9183-206fceeabc40 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 917.469347] env[62066]: INFO nova.compute.claims [None req-740993dc-fc35-4aa8-9a06-8f4c0e5f95e1 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 9c989152-ec69-478e-a0b0-62983852576c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 917.495477] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3ba3ebd2-6915-4f7b-8349-9384d9a8c223 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Lock "8807ce4d-532f-469c-a302-464c61c7efeb" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 20.728s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 917.568355] env[62066]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 917.568355] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]525e97f8-5b6e-585c-f9ed-54c63164d125" [ 917.568355] env[62066]: _type = "HttpNfcLease" [ 917.568355] env[62066]: } is ready. {{(pid=62066) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 917.568662] env[62066]: DEBUG oslo_vmware.rw_handles [None req-81c18bf6-1c8d-45a0-a66c-dcda0761f479 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 917.568662] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]525e97f8-5b6e-585c-f9ed-54c63164d125" [ 917.568662] env[62066]: _type = "HttpNfcLease" [ 917.568662] env[62066]: }. {{(pid=62066) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 917.569399] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bdd7d1d2-b10c-4bd6-a520-dab707dac72a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.576542] env[62066]: DEBUG oslo_vmware.rw_handles [None req-81c18bf6-1c8d-45a0-a66c-dcda0761f479 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/523551ec-af05-8a45-7c98-55783b4fdbe9/disk-0.vmdk from lease info. {{(pid=62066) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 917.576724] env[62066]: DEBUG oslo_vmware.rw_handles [None req-81c18bf6-1c8d-45a0-a66c-dcda0761f479 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Creating HTTP connection to write to file with size = 31661056 and URL = https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/523551ec-af05-8a45-7c98-55783b4fdbe9/disk-0.vmdk. {{(pid=62066) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 917.639747] env[62066]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-dc4730b2-0a11-45a2-b1ec-c985d69e3c84 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.762648] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1341133, 'name': CreateVM_Task, 'duration_secs': 0.242053} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.762922] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 15afaea0-65de-4f32-851c-365003e5498f] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 917.763256] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ad6adfa4-e5e7-49bc-8c71-f4f4f133daea tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 917.763423] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ad6adfa4-e5e7-49bc-8c71-f4f4f133daea tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Acquired lock "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 917.763748] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ad6adfa4-e5e7-49bc-8c71-f4f4f133daea tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 917.764104] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a8011ef6-8a10-436e-ad84-1d9f714ef3d9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.768569] env[62066]: DEBUG oslo_vmware.api [None req-ad6adfa4-e5e7-49bc-8c71-f4f4f133daea tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Waiting for the task: (returnval){ [ 917.768569] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]52e05575-934d-6ad2-dd3e-b5d40c29d101" [ 917.768569] env[62066]: _type = "Task" [ 917.768569] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.775941] env[62066]: DEBUG oslo_vmware.api [None req-ad6adfa4-e5e7-49bc-8c71-f4f4f133daea tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52e05575-934d-6ad2-dd3e-b5d40c29d101, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.834757] env[62066]: DEBUG oslo_concurrency.lockutils [None req-36f2458b-83e5-4859-9178-b3e847334e66 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Lock "4e39b9b0-4ab0-4f48-8a3f-6bc1895597eb" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 25.343s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 917.911501] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-1b0d97e8-08d8-404b-b2dc-335d2f49fb47 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: c8e0d47c-4421-4e00-9183-206fceeabc40] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 917.911820] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f88406de-f183-49cb-8498-c46d52ae12e2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.920543] env[62066]: DEBUG oslo_vmware.api [None req-1b0d97e8-08d8-404b-b2dc-335d2f49fb47 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Waiting for the task: (returnval){ [ 917.920543] env[62066]: value = "task-1341134" [ 917.920543] env[62066]: _type = "Task" [ 917.920543] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.931287] env[62066]: DEBUG oslo_vmware.api [None req-1b0d97e8-08d8-404b-b2dc-335d2f49fb47 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Task: {'id': task-1341134, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.977123] env[62066]: INFO nova.compute.resource_tracker [None req-740993dc-fc35-4aa8-9a06-8f4c0e5f95e1 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 9c989152-ec69-478e-a0b0-62983852576c] Updating resource usage from migration aba80e34-5809-409a-9dcc-c487ec0ebe3f [ 918.198510] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b853a2a-d307-45d0-9251-224a60bf5b24 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.210800] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-122018c3-5afe-4f9b-8878-4430424e51f9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.248094] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93bae2ae-c8b2-46d5-b61d-8bedd659e9d1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.258607] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f820456-772c-4da4-a70a-9e48daa0856f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.276822] env[62066]: DEBUG nova.compute.provider_tree [None req-740993dc-fc35-4aa8-9a06-8f4c0e5f95e1 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 918.292767] env[62066]: DEBUG oslo_vmware.api [None req-ad6adfa4-e5e7-49bc-8c71-f4f4f133daea tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52e05575-934d-6ad2-dd3e-b5d40c29d101, 'name': SearchDatastore_Task, 'duration_secs': 0.011466} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.295316] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ad6adfa4-e5e7-49bc-8c71-f4f4f133daea tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Releasing lock "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 918.295640] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-ad6adfa4-e5e7-49bc-8c71-f4f4f133daea tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] [instance: 15afaea0-65de-4f32-851c-365003e5498f] Processing image 50ff584c-3b50-4395-af07-3e66769bc9f7 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 918.296323] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ad6adfa4-e5e7-49bc-8c71-f4f4f133daea tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 918.296323] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ad6adfa4-e5e7-49bc-8c71-f4f4f133daea tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Acquired lock "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 918.296449] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-ad6adfa4-e5e7-49bc-8c71-f4f4f133daea tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 918.297348] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-62c222fb-dcd6-4124-a640-b7cefef8b497 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.306181] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-ad6adfa4-e5e7-49bc-8c71-f4f4f133daea tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 918.306408] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-ad6adfa4-e5e7-49bc-8c71-f4f4f133daea tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 918.307207] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1823ff7a-4fce-484f-801c-9e859a9a80df {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.316740] env[62066]: DEBUG oslo_vmware.api [None req-ad6adfa4-e5e7-49bc-8c71-f4f4f133daea tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Waiting for the task: (returnval){ [ 918.316740] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]521b5fae-74e3-e80c-c85c-55e93462930d" [ 918.316740] env[62066]: _type = "Task" [ 918.316740] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.322477] env[62066]: DEBUG oslo_vmware.api [None req-ad6adfa4-e5e7-49bc-8c71-f4f4f133daea tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]521b5fae-74e3-e80c-c85c-55e93462930d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.432042] env[62066]: DEBUG oslo_vmware.api [None req-1b0d97e8-08d8-404b-b2dc-335d2f49fb47 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Task: {'id': task-1341134, 'name': PowerOffVM_Task, 'duration_secs': 0.362682} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.432362] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-1b0d97e8-08d8-404b-b2dc-335d2f49fb47 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: c8e0d47c-4421-4e00-9183-206fceeabc40] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 918.432543] env[62066]: DEBUG nova.compute.manager [None req-1b0d97e8-08d8-404b-b2dc-335d2f49fb47 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: c8e0d47c-4421-4e00-9183-206fceeabc40] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 918.433351] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d92e014-7aec-42a5-9039-65021dacefd4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.707441] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0e878f04-940e-4d47-ad0a-15c5be73ba89 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Acquiring lock "b08f27b1-429e-4d05-9a95-f45fd0cb2904" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 918.707441] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0e878f04-940e-4d47-ad0a-15c5be73ba89 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Lock "b08f27b1-429e-4d05-9a95-f45fd0cb2904" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 918.707441] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0e878f04-940e-4d47-ad0a-15c5be73ba89 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Acquiring lock "b08f27b1-429e-4d05-9a95-f45fd0cb2904-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 918.707441] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0e878f04-940e-4d47-ad0a-15c5be73ba89 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Lock "b08f27b1-429e-4d05-9a95-f45fd0cb2904-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 918.707441] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0e878f04-940e-4d47-ad0a-15c5be73ba89 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Lock "b08f27b1-429e-4d05-9a95-f45fd0cb2904-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 918.712039] env[62066]: INFO nova.compute.manager [None req-0e878f04-940e-4d47-ad0a-15c5be73ba89 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: b08f27b1-429e-4d05-9a95-f45fd0cb2904] Terminating instance [ 918.718020] env[62066]: DEBUG nova.compute.manager [None req-0e878f04-940e-4d47-ad0a-15c5be73ba89 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: b08f27b1-429e-4d05-9a95-f45fd0cb2904] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 918.718020] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-0e878f04-940e-4d47-ad0a-15c5be73ba89 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: b08f27b1-429e-4d05-9a95-f45fd0cb2904] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 918.718020] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3ee7217-e5d0-40aa-aa97-1fd19f891dff {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.728463] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e878f04-940e-4d47-ad0a-15c5be73ba89 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: b08f27b1-429e-4d05-9a95-f45fd0cb2904] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 918.729064] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-97832f3a-3ca3-46fb-9ddc-fb5536401152 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.741761] env[62066]: DEBUG oslo_vmware.api [None req-0e878f04-940e-4d47-ad0a-15c5be73ba89 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Waiting for the task: (returnval){ [ 918.741761] env[62066]: value = "task-1341135" [ 918.741761] env[62066]: _type = "Task" [ 918.741761] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.755719] env[62066]: DEBUG oslo_vmware.api [None req-0e878f04-940e-4d47-ad0a-15c5be73ba89 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': task-1341135, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.787445] env[62066]: DEBUG nova.scheduler.client.report [None req-740993dc-fc35-4aa8-9a06-8f4c0e5f95e1 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 918.829473] env[62066]: DEBUG oslo_vmware.api [None req-ad6adfa4-e5e7-49bc-8c71-f4f4f133daea tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]521b5fae-74e3-e80c-c85c-55e93462930d, 'name': SearchDatastore_Task, 'duration_secs': 0.011977} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.830835] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5e9cbac5-b517-4c42-861f-d7a3628160b0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.834494] env[62066]: DEBUG oslo_vmware.rw_handles [None req-81c18bf6-1c8d-45a0-a66c-dcda0761f479 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Completed reading data from the image iterator. {{(pid=62066) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 918.834789] env[62066]: DEBUG oslo_vmware.rw_handles [None req-81c18bf6-1c8d-45a0-a66c-dcda0761f479 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/523551ec-af05-8a45-7c98-55783b4fdbe9/disk-0.vmdk. {{(pid=62066) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 918.835703] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad1ddfa7-5919-439b-bce1-7fd2884ef783 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.842916] env[62066]: DEBUG oslo_vmware.api [None req-ad6adfa4-e5e7-49bc-8c71-f4f4f133daea tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Waiting for the task: (returnval){ [ 918.842916] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]5254a1b6-e1b6-5d31-7d04-423d35728284" [ 918.842916] env[62066]: _type = "Task" [ 918.842916] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.843937] env[62066]: DEBUG oslo_vmware.rw_handles [None req-81c18bf6-1c8d-45a0-a66c-dcda0761f479 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/523551ec-af05-8a45-7c98-55783b4fdbe9/disk-0.vmdk is in state: ready. {{(pid=62066) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 918.844160] env[62066]: DEBUG oslo_vmware.rw_handles [None req-81c18bf6-1c8d-45a0-a66c-dcda0761f479 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Releasing lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/523551ec-af05-8a45-7c98-55783b4fdbe9/disk-0.vmdk. {{(pid=62066) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 918.847866] env[62066]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-4703d2a0-9235-4ce0-82d3-a69104c821ec {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.857081] env[62066]: DEBUG oslo_vmware.api [None req-ad6adfa4-e5e7-49bc-8c71-f4f4f133daea tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]5254a1b6-e1b6-5d31-7d04-423d35728284, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.934144] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d267d329-25e8-414c-9540-a1c71897bff8 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Acquiring lock "20987cb6-e12c-48c5-8fae-d990b1bb6b8a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 918.934546] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d267d329-25e8-414c-9540-a1c71897bff8 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Lock "20987cb6-e12c-48c5-8fae-d990b1bb6b8a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 918.946997] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1b0d97e8-08d8-404b-b2dc-335d2f49fb47 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Lock "c8e0d47c-4421-4e00-9183-206fceeabc40" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.551s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 919.186068] env[62066]: DEBUG oslo_vmware.rw_handles [None req-81c18bf6-1c8d-45a0-a66c-dcda0761f479 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Closed VMDK write handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/523551ec-af05-8a45-7c98-55783b4fdbe9/disk-0.vmdk. {{(pid=62066) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 919.186450] env[62066]: INFO nova.virt.vmwareapi.images [None req-81c18bf6-1c8d-45a0-a66c-dcda0761f479 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] [instance: 8ff6fbb9-c90f-498d-9a85-d220a8c2f794] Downloaded image file data 1ebd2ae1-0ae3-4486-822c-c75cafe27fe3 [ 919.187679] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3792eeac-7144-4e6a-999f-b7af724ac457 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.204853] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-384a42d5-edac-4f9a-8a21-8f0400d05fa2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.253897] env[62066]: DEBUG oslo_vmware.api [None req-0e878f04-940e-4d47-ad0a-15c5be73ba89 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': task-1341135, 'name': PowerOffVM_Task, 'duration_secs': 0.324957} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.255443] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e878f04-940e-4d47-ad0a-15c5be73ba89 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: b08f27b1-429e-4d05-9a95-f45fd0cb2904] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 919.255636] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-0e878f04-940e-4d47-ad0a-15c5be73ba89 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: b08f27b1-429e-4d05-9a95-f45fd0cb2904] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 919.258347] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-48982595-a23c-46b0-badf-52a03190f49b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.260267] env[62066]: DEBUG oslo_concurrency.lockutils [None req-54702e84-4f3e-4046-b83c-40cd5c6f41f7 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Acquiring lock "651e5142-7ab0-4040-97b4-4599bd4e419a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 919.260490] env[62066]: DEBUG oslo_concurrency.lockutils [None req-54702e84-4f3e-4046-b83c-40cd5c6f41f7 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Lock "651e5142-7ab0-4040-97b4-4599bd4e419a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 919.263445] env[62066]: INFO nova.virt.vmwareapi.images [None req-81c18bf6-1c8d-45a0-a66c-dcda0761f479 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] [instance: 8ff6fbb9-c90f-498d-9a85-d220a8c2f794] The imported VM was unregistered [ 919.266423] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-81c18bf6-1c8d-45a0-a66c-dcda0761f479 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] [instance: 8ff6fbb9-c90f-498d-9a85-d220a8c2f794] Caching image {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 919.266658] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-81c18bf6-1c8d-45a0-a66c-dcda0761f479 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Creating directory with path [datastore1] devstack-image-cache_base/1ebd2ae1-0ae3-4486-822c-c75cafe27fe3 {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 919.267269] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9d6450f4-a743-41d7-9537-0af3455c292c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.283770] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-81c18bf6-1c8d-45a0-a66c-dcda0761f479 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Created directory with path [datastore1] devstack-image-cache_base/1ebd2ae1-0ae3-4486-822c-c75cafe27fe3 {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 919.283963] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-81c18bf6-1c8d-45a0-a66c-dcda0761f479 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Moving virtual disk from [datastore1] OSTACK_IMG_7d0cd4f2-d1f4-4ceb-9c76-75a77646fd93/OSTACK_IMG_7d0cd4f2-d1f4-4ceb-9c76-75a77646fd93.vmdk to [datastore1] devstack-image-cache_base/1ebd2ae1-0ae3-4486-822c-c75cafe27fe3/1ebd2ae1-0ae3-4486-822c-c75cafe27fe3.vmdk. {{(pid=62066) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 919.284231] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-4b1454ef-a415-4d7e-aa76-4ae783ffb251 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.290683] env[62066]: DEBUG oslo_vmware.api [None req-81c18bf6-1c8d-45a0-a66c-dcda0761f479 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Waiting for the task: (returnval){ [ 919.290683] env[62066]: value = "task-1341138" [ 919.290683] env[62066]: _type = "Task" [ 919.290683] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.296133] env[62066]: DEBUG oslo_concurrency.lockutils [None req-740993dc-fc35-4aa8-9a06-8f4c0e5f95e1 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.338s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 919.296335] env[62066]: INFO nova.compute.manager [None req-740993dc-fc35-4aa8-9a06-8f4c0e5f95e1 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 9c989152-ec69-478e-a0b0-62983852576c] Migrating [ 919.296557] env[62066]: DEBUG oslo_concurrency.lockutils [None req-740993dc-fc35-4aa8-9a06-8f4c0e5f95e1 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Acquiring lock "compute-rpcapi-router" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 919.296707] env[62066]: DEBUG oslo_concurrency.lockutils [None req-740993dc-fc35-4aa8-9a06-8f4c0e5f95e1 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Acquired lock "compute-rpcapi-router" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 919.301915] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e9681fdd-3744-4c1e-81e7-a2a0f833c385 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.206s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 919.303491] env[62066]: INFO nova.compute.claims [None req-e9681fdd-3744-4c1e-81e7-a2a0f833c385 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] [instance: a10a6dc0-7df7-40a1-888a-d1414dbeb1f7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 919.306401] env[62066]: DEBUG oslo_vmware.api [None req-81c18bf6-1c8d-45a0-a66c-dcda0761f479 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Task: {'id': task-1341138, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.330805] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-0e878f04-940e-4d47-ad0a-15c5be73ba89 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: b08f27b1-429e-4d05-9a95-f45fd0cb2904] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 919.331047] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-0e878f04-940e-4d47-ad0a-15c5be73ba89 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: b08f27b1-429e-4d05-9a95-f45fd0cb2904] Deleting contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 919.331315] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-0e878f04-940e-4d47-ad0a-15c5be73ba89 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Deleting the datastore file [datastore2] b08f27b1-429e-4d05-9a95-f45fd0cb2904 {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 919.332032] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5d6a22c7-618e-4b9c-b2b6-82d0a1498af0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.339458] env[62066]: DEBUG oslo_vmware.api [None req-0e878f04-940e-4d47-ad0a-15c5be73ba89 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Waiting for the task: (returnval){ [ 919.339458] env[62066]: value = "task-1341139" [ 919.339458] env[62066]: _type = "Task" [ 919.339458] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.350252] env[62066]: DEBUG oslo_vmware.api [None req-0e878f04-940e-4d47-ad0a-15c5be73ba89 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': task-1341139, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.355654] env[62066]: DEBUG oslo_vmware.api [None req-ad6adfa4-e5e7-49bc-8c71-f4f4f133daea tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]5254a1b6-e1b6-5d31-7d04-423d35728284, 'name': SearchDatastore_Task, 'duration_secs': 0.012391} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.355907] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ad6adfa4-e5e7-49bc-8c71-f4f4f133daea tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Releasing lock "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 919.356185] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-ad6adfa4-e5e7-49bc-8c71-f4f4f133daea tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk to [datastore1] 15afaea0-65de-4f32-851c-365003e5498f/15afaea0-65de-4f32-851c-365003e5498f.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 919.356440] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-951e6bfe-9459-497b-a882-f57a52546a94 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.362831] env[62066]: DEBUG oslo_vmware.api [None req-ad6adfa4-e5e7-49bc-8c71-f4f4f133daea tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Waiting for the task: (returnval){ [ 919.362831] env[62066]: value = "task-1341140" [ 919.362831] env[62066]: _type = "Task" [ 919.362831] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.370245] env[62066]: DEBUG oslo_vmware.api [None req-ad6adfa4-e5e7-49bc-8c71-f4f4f133daea tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Task: {'id': task-1341140, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.434263] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3bd520fa-2cd0-41e8-afc3-e67149da4105 tempest-ServerTagsTestJSON-2015212290 tempest-ServerTagsTestJSON-2015212290-project-member] Acquiring lock "f4a84152-f222-46a2-9a35-8a0a0078fc12" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 919.434544] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3bd520fa-2cd0-41e8-afc3-e67149da4105 tempest-ServerTagsTestJSON-2015212290 tempest-ServerTagsTestJSON-2015212290-project-member] Lock "f4a84152-f222-46a2-9a35-8a0a0078fc12" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 919.434757] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3bd520fa-2cd0-41e8-afc3-e67149da4105 tempest-ServerTagsTestJSON-2015212290 tempest-ServerTagsTestJSON-2015212290-project-member] Acquiring lock "f4a84152-f222-46a2-9a35-8a0a0078fc12-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 919.434951] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3bd520fa-2cd0-41e8-afc3-e67149da4105 tempest-ServerTagsTestJSON-2015212290 tempest-ServerTagsTestJSON-2015212290-project-member] Lock "f4a84152-f222-46a2-9a35-8a0a0078fc12-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 919.435147] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3bd520fa-2cd0-41e8-afc3-e67149da4105 tempest-ServerTagsTestJSON-2015212290 tempest-ServerTagsTestJSON-2015212290-project-member] Lock "f4a84152-f222-46a2-9a35-8a0a0078fc12-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 919.437252] env[62066]: DEBUG nova.compute.manager [None req-d267d329-25e8-414c-9540-a1c71897bff8 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 20987cb6-e12c-48c5-8fae-d990b1bb6b8a] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 919.440027] env[62066]: INFO nova.compute.manager [None req-3bd520fa-2cd0-41e8-afc3-e67149da4105 tempest-ServerTagsTestJSON-2015212290 tempest-ServerTagsTestJSON-2015212290-project-member] [instance: f4a84152-f222-46a2-9a35-8a0a0078fc12] Terminating instance [ 919.444163] env[62066]: DEBUG nova.compute.manager [None req-3bd520fa-2cd0-41e8-afc3-e67149da4105 tempest-ServerTagsTestJSON-2015212290 tempest-ServerTagsTestJSON-2015212290-project-member] [instance: f4a84152-f222-46a2-9a35-8a0a0078fc12] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 919.444163] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-3bd520fa-2cd0-41e8-afc3-e67149da4105 tempest-ServerTagsTestJSON-2015212290 tempest-ServerTagsTestJSON-2015212290-project-member] [instance: f4a84152-f222-46a2-9a35-8a0a0078fc12] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 919.444163] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a648906c-c52f-4073-bce4-bbf3655c6b18 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.451935] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-3bd520fa-2cd0-41e8-afc3-e67149da4105 tempest-ServerTagsTestJSON-2015212290 tempest-ServerTagsTestJSON-2015212290-project-member] [instance: f4a84152-f222-46a2-9a35-8a0a0078fc12] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 919.452245] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-45014062-9264-4138-a1bd-9bf286cbae80 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.458196] env[62066]: DEBUG oslo_vmware.api [None req-3bd520fa-2cd0-41e8-afc3-e67149da4105 tempest-ServerTagsTestJSON-2015212290 tempest-ServerTagsTestJSON-2015212290-project-member] Waiting for the task: (returnval){ [ 919.458196] env[62066]: value = "task-1341141" [ 919.458196] env[62066]: _type = "Task" [ 919.458196] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.466021] env[62066]: DEBUG oslo_vmware.api [None req-3bd520fa-2cd0-41e8-afc3-e67149da4105 tempest-ServerTagsTestJSON-2015212290 tempest-ServerTagsTestJSON-2015212290-project-member] Task: {'id': task-1341141, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.763272] env[62066]: DEBUG nova.compute.manager [None req-54702e84-4f3e-4046-b83c-40cd5c6f41f7 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 651e5142-7ab0-4040-97b4-4599bd4e419a] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 919.799826] env[62066]: INFO nova.compute.rpcapi [None req-740993dc-fc35-4aa8-9a06-8f4c0e5f95e1 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Automatically selected compute RPC version 6.3 from minimum service version 67 [ 919.800010] env[62066]: DEBUG oslo_concurrency.lockutils [None req-740993dc-fc35-4aa8-9a06-8f4c0e5f95e1 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Releasing lock "compute-rpcapi-router" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 919.817039] env[62066]: DEBUG oslo_vmware.api [None req-81c18bf6-1c8d-45a0-a66c-dcda0761f479 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Task: {'id': task-1341138, 'name': MoveVirtualDisk_Task} progress is 15%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.852557] env[62066]: DEBUG oslo_vmware.api [None req-0e878f04-940e-4d47-ad0a-15c5be73ba89 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': task-1341139, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.265818} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.853178] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-0e878f04-940e-4d47-ad0a-15c5be73ba89 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 919.853475] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-0e878f04-940e-4d47-ad0a-15c5be73ba89 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: b08f27b1-429e-4d05-9a95-f45fd0cb2904] Deleted contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 919.853678] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-0e878f04-940e-4d47-ad0a-15c5be73ba89 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: b08f27b1-429e-4d05-9a95-f45fd0cb2904] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 919.853991] env[62066]: INFO nova.compute.manager [None req-0e878f04-940e-4d47-ad0a-15c5be73ba89 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: b08f27b1-429e-4d05-9a95-f45fd0cb2904] Took 1.14 seconds to destroy the instance on the hypervisor. [ 919.854306] env[62066]: DEBUG oslo.service.loopingcall [None req-0e878f04-940e-4d47-ad0a-15c5be73ba89 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 919.854616] env[62066]: DEBUG nova.compute.manager [-] [instance: b08f27b1-429e-4d05-9a95-f45fd0cb2904] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 919.854695] env[62066]: DEBUG nova.network.neutron [-] [instance: b08f27b1-429e-4d05-9a95-f45fd0cb2904] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 919.875139] env[62066]: DEBUG oslo_vmware.api [None req-ad6adfa4-e5e7-49bc-8c71-f4f4f133daea tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Task: {'id': task-1341140, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.963965] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d267d329-25e8-414c-9540-a1c71897bff8 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 919.972045] env[62066]: DEBUG nova.objects.instance [None req-96e7ee9b-6f06-4817-a05d-ab2ef59667eb tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Lazy-loading 'flavor' on Instance uuid c8e0d47c-4421-4e00-9183-206fceeabc40 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 919.973429] env[62066]: DEBUG oslo_vmware.api [None req-3bd520fa-2cd0-41e8-afc3-e67149da4105 tempest-ServerTagsTestJSON-2015212290 tempest-ServerTagsTestJSON-2015212290-project-member] Task: {'id': task-1341141, 'name': PowerOffVM_Task, 'duration_secs': 0.190606} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.973684] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-3bd520fa-2cd0-41e8-afc3-e67149da4105 tempest-ServerTagsTestJSON-2015212290 tempest-ServerTagsTestJSON-2015212290-project-member] [instance: f4a84152-f222-46a2-9a35-8a0a0078fc12] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 919.973858] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-3bd520fa-2cd0-41e8-afc3-e67149da4105 tempest-ServerTagsTestJSON-2015212290 tempest-ServerTagsTestJSON-2015212290-project-member] [instance: f4a84152-f222-46a2-9a35-8a0a0078fc12] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 919.974336] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-886fa653-cc92-4ddb-a040-0c5e52c76bb5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.057057] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-3bd520fa-2cd0-41e8-afc3-e67149da4105 tempest-ServerTagsTestJSON-2015212290 tempest-ServerTagsTestJSON-2015212290-project-member] [instance: f4a84152-f222-46a2-9a35-8a0a0078fc12] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 920.057308] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-3bd520fa-2cd0-41e8-afc3-e67149da4105 tempest-ServerTagsTestJSON-2015212290 tempest-ServerTagsTestJSON-2015212290-project-member] [instance: f4a84152-f222-46a2-9a35-8a0a0078fc12] Deleting contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 920.057549] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-3bd520fa-2cd0-41e8-afc3-e67149da4105 tempest-ServerTagsTestJSON-2015212290 tempest-ServerTagsTestJSON-2015212290-project-member] Deleting the datastore file [datastore2] f4a84152-f222-46a2-9a35-8a0a0078fc12 {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 920.057770] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-62e8aced-836a-4be0-870b-5a58d15c85ad {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.067975] env[62066]: DEBUG oslo_vmware.api [None req-3bd520fa-2cd0-41e8-afc3-e67149da4105 tempest-ServerTagsTestJSON-2015212290 tempest-ServerTagsTestJSON-2015212290-project-member] Waiting for the task: (returnval){ [ 920.067975] env[62066]: value = "task-1341143" [ 920.067975] env[62066]: _type = "Task" [ 920.067975] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.080151] env[62066]: DEBUG oslo_vmware.api [None req-3bd520fa-2cd0-41e8-afc3-e67149da4105 tempest-ServerTagsTestJSON-2015212290 tempest-ServerTagsTestJSON-2015212290-project-member] Task: {'id': task-1341143, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.264468] env[62066]: DEBUG nova.compute.manager [req-283837cc-e086-4502-a7ab-74946642990f req-93f79f0b-a3ba-46b7-b49b-150e9fcb60f1 service nova] [instance: b08f27b1-429e-4d05-9a95-f45fd0cb2904] Received event network-vif-deleted-62d600d9-8ad3-43f8-a2b0-050747916fdb {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 920.265099] env[62066]: INFO nova.compute.manager [req-283837cc-e086-4502-a7ab-74946642990f req-93f79f0b-a3ba-46b7-b49b-150e9fcb60f1 service nova] [instance: b08f27b1-429e-4d05-9a95-f45fd0cb2904] Neutron deleted interface 62d600d9-8ad3-43f8-a2b0-050747916fdb; detaching it from the instance and deleting it from the info cache [ 920.265256] env[62066]: DEBUG nova.network.neutron [req-283837cc-e086-4502-a7ab-74946642990f req-93f79f0b-a3ba-46b7-b49b-150e9fcb60f1 service nova] [instance: b08f27b1-429e-4d05-9a95-f45fd0cb2904] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 920.284253] env[62066]: DEBUG oslo_concurrency.lockutils [None req-54702e84-4f3e-4046-b83c-40cd5c6f41f7 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 920.306084] env[62066]: DEBUG oslo_vmware.api [None req-81c18bf6-1c8d-45a0-a66c-dcda0761f479 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Task: {'id': task-1341138, 'name': MoveVirtualDisk_Task} progress is 32%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.325885] env[62066]: DEBUG oslo_concurrency.lockutils [None req-740993dc-fc35-4aa8-9a06-8f4c0e5f95e1 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Acquiring lock "refresh_cache-9c989152-ec69-478e-a0b0-62983852576c" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 920.326109] env[62066]: DEBUG oslo_concurrency.lockutils [None req-740993dc-fc35-4aa8-9a06-8f4c0e5f95e1 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Acquired lock "refresh_cache-9c989152-ec69-478e-a0b0-62983852576c" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 920.326320] env[62066]: DEBUG nova.network.neutron [None req-740993dc-fc35-4aa8-9a06-8f4c0e5f95e1 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 9c989152-ec69-478e-a0b0-62983852576c] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 920.376859] env[62066]: DEBUG oslo_vmware.api [None req-ad6adfa4-e5e7-49bc-8c71-f4f4f133daea tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Task: {'id': task-1341140, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.478075] env[62066]: DEBUG oslo_concurrency.lockutils [None req-96e7ee9b-6f06-4817-a05d-ab2ef59667eb tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Acquiring lock "refresh_cache-c8e0d47c-4421-4e00-9183-206fceeabc40" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 920.478075] env[62066]: DEBUG oslo_concurrency.lockutils [None req-96e7ee9b-6f06-4817-a05d-ab2ef59667eb tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Acquired lock "refresh_cache-c8e0d47c-4421-4e00-9183-206fceeabc40" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 920.478075] env[62066]: DEBUG nova.network.neutron [None req-96e7ee9b-6f06-4817-a05d-ab2ef59667eb tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: c8e0d47c-4421-4e00-9183-206fceeabc40] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 920.478413] env[62066]: DEBUG nova.objects.instance [None req-96e7ee9b-6f06-4817-a05d-ab2ef59667eb tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Lazy-loading 'info_cache' on Instance uuid c8e0d47c-4421-4e00-9183-206fceeabc40 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 920.554260] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2204e51-ab87-4050-9ee9-a7dc57e31b3c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.566639] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93866a04-d4ef-466e-bbe2-ed6c03648c42 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.578325] env[62066]: DEBUG oslo_vmware.api [None req-3bd520fa-2cd0-41e8-afc3-e67149da4105 tempest-ServerTagsTestJSON-2015212290 tempest-ServerTagsTestJSON-2015212290-project-member] Task: {'id': task-1341143, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.32619} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.610932] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-3bd520fa-2cd0-41e8-afc3-e67149da4105 tempest-ServerTagsTestJSON-2015212290 tempest-ServerTagsTestJSON-2015212290-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 920.611460] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-3bd520fa-2cd0-41e8-afc3-e67149da4105 tempest-ServerTagsTestJSON-2015212290 tempest-ServerTagsTestJSON-2015212290-project-member] [instance: f4a84152-f222-46a2-9a35-8a0a0078fc12] Deleted contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 920.611814] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-3bd520fa-2cd0-41e8-afc3-e67149da4105 tempest-ServerTagsTestJSON-2015212290 tempest-ServerTagsTestJSON-2015212290-project-member] [instance: f4a84152-f222-46a2-9a35-8a0a0078fc12] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 920.612164] env[62066]: INFO nova.compute.manager [None req-3bd520fa-2cd0-41e8-afc3-e67149da4105 tempest-ServerTagsTestJSON-2015212290 tempest-ServerTagsTestJSON-2015212290-project-member] [instance: f4a84152-f222-46a2-9a35-8a0a0078fc12] Took 1.17 seconds to destroy the instance on the hypervisor. [ 920.612615] env[62066]: DEBUG oslo.service.loopingcall [None req-3bd520fa-2cd0-41e8-afc3-e67149da4105 tempest-ServerTagsTestJSON-2015212290 tempest-ServerTagsTestJSON-2015212290-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 920.613327] env[62066]: DEBUG nova.compute.manager [-] [instance: f4a84152-f222-46a2-9a35-8a0a0078fc12] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 920.614071] env[62066]: DEBUG nova.network.neutron [-] [instance: f4a84152-f222-46a2-9a35-8a0a0078fc12] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 920.616937] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f32e9a3d-c3aa-40fb-9e93-150a70f4506d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.630453] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a93a9c9-df15-4056-a24d-666ef3e63982 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.650913] env[62066]: DEBUG nova.compute.provider_tree [None req-e9681fdd-3744-4c1e-81e7-a2a0f833c385 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 920.704276] env[62066]: DEBUG nova.network.neutron [-] [instance: b08f27b1-429e-4d05-9a95-f45fd0cb2904] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 920.770813] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1697d7ab-94c2-46e8-816f-0524fbaa80ea {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.782651] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4e4fea7-1721-4a29-a63e-f2f2fc0dd670 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.806547] env[62066]: DEBUG oslo_vmware.api [None req-81c18bf6-1c8d-45a0-a66c-dcda0761f479 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Task: {'id': task-1341138, 'name': MoveVirtualDisk_Task} progress is 49%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.819592] env[62066]: DEBUG nova.compute.manager [req-283837cc-e086-4502-a7ab-74946642990f req-93f79f0b-a3ba-46b7-b49b-150e9fcb60f1 service nova] [instance: b08f27b1-429e-4d05-9a95-f45fd0cb2904] Detach interface failed, port_id=62d600d9-8ad3-43f8-a2b0-050747916fdb, reason: Instance b08f27b1-429e-4d05-9a95-f45fd0cb2904 could not be found. {{(pid=62066) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 920.878962] env[62066]: DEBUG oslo_vmware.api [None req-ad6adfa4-e5e7-49bc-8c71-f4f4f133daea tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Task: {'id': task-1341140, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.430778} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.879330] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-ad6adfa4-e5e7-49bc-8c71-f4f4f133daea tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk to [datastore1] 15afaea0-65de-4f32-851c-365003e5498f/15afaea0-65de-4f32-851c-365003e5498f.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 920.879604] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-ad6adfa4-e5e7-49bc-8c71-f4f4f133daea tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] [instance: 15afaea0-65de-4f32-851c-365003e5498f] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 920.879895] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-cd8b21ad-ecad-4a9e-9cf8-8968bd134bb4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.888072] env[62066]: DEBUG oslo_vmware.api [None req-ad6adfa4-e5e7-49bc-8c71-f4f4f133daea tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Waiting for the task: (returnval){ [ 920.888072] env[62066]: value = "task-1341144" [ 920.888072] env[62066]: _type = "Task" [ 920.888072] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.897521] env[62066]: DEBUG oslo_vmware.api [None req-ad6adfa4-e5e7-49bc-8c71-f4f4f133daea tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Task: {'id': task-1341144, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.982177] env[62066]: DEBUG nova.objects.base [None req-96e7ee9b-6f06-4817-a05d-ab2ef59667eb tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Object Instance lazy-loaded attributes: flavor,info_cache {{(pid=62066) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 921.085726] env[62066]: DEBUG nova.network.neutron [None req-740993dc-fc35-4aa8-9a06-8f4c0e5f95e1 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 9c989152-ec69-478e-a0b0-62983852576c] Updating instance_info_cache with network_info: [{"id": "1c0d0075-7f22-441e-a283-21e9b5c5aecc", "address": "fa:16:3e:c5:c3:07", "network": {"id": "1c6c51e4-96f8-4175-b7f5-96b8a42fcb13", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-2006089809-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d609babc987a42e2a8ddb4bfb9c3b103", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a965790c-2d2f-4c2a-9ee7-745f4d53039b", "external-id": "nsx-vlan-transportzone-708", "segmentation_id": 708, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1c0d0075-7f", "ovs_interfaceid": "1c0d0075-7f22-441e-a283-21e9b5c5aecc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 921.154434] env[62066]: DEBUG nova.scheduler.client.report [None req-e9681fdd-3744-4c1e-81e7-a2a0f833c385 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 921.211641] env[62066]: INFO nova.compute.manager [-] [instance: b08f27b1-429e-4d05-9a95-f45fd0cb2904] Took 1.35 seconds to deallocate network for instance. [ 921.307963] env[62066]: DEBUG oslo_vmware.api [None req-81c18bf6-1c8d-45a0-a66c-dcda0761f479 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Task: {'id': task-1341138, 'name': MoveVirtualDisk_Task} progress is 71%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.401735] env[62066]: DEBUG oslo_vmware.api [None req-ad6adfa4-e5e7-49bc-8c71-f4f4f133daea tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Task: {'id': task-1341144, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.112364} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.402036] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-ad6adfa4-e5e7-49bc-8c71-f4f4f133daea tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] [instance: 15afaea0-65de-4f32-851c-365003e5498f] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 921.402939] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91a46e7b-8db3-4d74-b4ff-ae6cce4e2719 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.425556] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-ad6adfa4-e5e7-49bc-8c71-f4f4f133daea tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] [instance: 15afaea0-65de-4f32-851c-365003e5498f] Reconfiguring VM instance instance-0000005a to attach disk [datastore1] 15afaea0-65de-4f32-851c-365003e5498f/15afaea0-65de-4f32-851c-365003e5498f.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 921.425915] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-035103ba-1054-4978-824b-b04ebced8f3c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.446630] env[62066]: DEBUG oslo_vmware.api [None req-ad6adfa4-e5e7-49bc-8c71-f4f4f133daea tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Waiting for the task: (returnval){ [ 921.446630] env[62066]: value = "task-1341145" [ 921.446630] env[62066]: _type = "Task" [ 921.446630] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.456994] env[62066]: DEBUG oslo_vmware.api [None req-ad6adfa4-e5e7-49bc-8c71-f4f4f133daea tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Task: {'id': task-1341145, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.588933] env[62066]: DEBUG oslo_concurrency.lockutils [None req-740993dc-fc35-4aa8-9a06-8f4c0e5f95e1 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Releasing lock "refresh_cache-9c989152-ec69-478e-a0b0-62983852576c" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 921.659912] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e9681fdd-3744-4c1e-81e7-a2a0f833c385 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.358s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 921.660628] env[62066]: DEBUG nova.compute.manager [None req-e9681fdd-3744-4c1e-81e7-a2a0f833c385 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] [instance: a10a6dc0-7df7-40a1-888a-d1414dbeb1f7] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 921.663538] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ab952be5-ba87-4d38-b219-8de60eb09bdf tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.247s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 921.663733] env[62066]: DEBUG nova.objects.instance [None req-ab952be5-ba87-4d38-b219-8de60eb09bdf tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Lazy-loading 'resources' on Instance uuid 90c33434-c127-450b-9ff0-75181b4ac385 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 921.698636] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0c3b1401-b067-4cd8-8015-7ffe20d4d016 tempest-ServersTestFqdnHostnames-1367628746 tempest-ServersTestFqdnHostnames-1367628746-project-member] Acquiring lock "afaca059-fc4e-4de2-8fa0-d2f226ea7051" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 921.698975] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0c3b1401-b067-4cd8-8015-7ffe20d4d016 tempest-ServersTestFqdnHostnames-1367628746 tempest-ServersTestFqdnHostnames-1367628746-project-member] Lock "afaca059-fc4e-4de2-8fa0-d2f226ea7051" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 921.699666] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0c3b1401-b067-4cd8-8015-7ffe20d4d016 tempest-ServersTestFqdnHostnames-1367628746 tempest-ServersTestFqdnHostnames-1367628746-project-member] Acquiring lock "afaca059-fc4e-4de2-8fa0-d2f226ea7051-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 921.699909] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0c3b1401-b067-4cd8-8015-7ffe20d4d016 tempest-ServersTestFqdnHostnames-1367628746 tempest-ServersTestFqdnHostnames-1367628746-project-member] Lock "afaca059-fc4e-4de2-8fa0-d2f226ea7051-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 921.700166] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0c3b1401-b067-4cd8-8015-7ffe20d4d016 tempest-ServersTestFqdnHostnames-1367628746 tempest-ServersTestFqdnHostnames-1367628746-project-member] Lock "afaca059-fc4e-4de2-8fa0-d2f226ea7051-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 921.702487] env[62066]: INFO nova.compute.manager [None req-0c3b1401-b067-4cd8-8015-7ffe20d4d016 tempest-ServersTestFqdnHostnames-1367628746 tempest-ServersTestFqdnHostnames-1367628746-project-member] [instance: afaca059-fc4e-4de2-8fa0-d2f226ea7051] Terminating instance [ 921.704447] env[62066]: DEBUG nova.compute.manager [None req-0c3b1401-b067-4cd8-8015-7ffe20d4d016 tempest-ServersTestFqdnHostnames-1367628746 tempest-ServersTestFqdnHostnames-1367628746-project-member] [instance: afaca059-fc4e-4de2-8fa0-d2f226ea7051] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 921.704646] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-0c3b1401-b067-4cd8-8015-7ffe20d4d016 tempest-ServersTestFqdnHostnames-1367628746 tempest-ServersTestFqdnHostnames-1367628746-project-member] [instance: afaca059-fc4e-4de2-8fa0-d2f226ea7051] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 921.705757] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2917326b-62d7-41be-9e91-a12d4fc14df5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.710648] env[62066]: DEBUG nova.network.neutron [None req-96e7ee9b-6f06-4817-a05d-ab2ef59667eb tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: c8e0d47c-4421-4e00-9183-206fceeabc40] Updating instance_info_cache with network_info: [{"id": "bb9924e0-8b18-43b2-bda0-3e409fa16718", "address": "fa:16:3e:b1:e4:41", "network": {"id": "5b89b2d4-71dd-4cd1-9edb-f4040be26540", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-283918837-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.165", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2252d0f6f18a47fd8a1ddad8020a6db7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "352165bb-004f-4180-9627-3a275dbe18af", "external-id": "nsx-vlan-transportzone-926", "segmentation_id": 926, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbb9924e0-8b", "ovs_interfaceid": "bb9924e0-8b18-43b2-bda0-3e409fa16718", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 921.717480] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-0c3b1401-b067-4cd8-8015-7ffe20d4d016 tempest-ServersTestFqdnHostnames-1367628746 tempest-ServersTestFqdnHostnames-1367628746-project-member] [instance: afaca059-fc4e-4de2-8fa0-d2f226ea7051] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 921.717778] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-20d21d6b-e32c-4fed-b62f-f0cd2618a469 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.719599] env[62066]: DEBUG nova.network.neutron [-] [instance: f4a84152-f222-46a2-9a35-8a0a0078fc12] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 921.721338] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0e878f04-940e-4d47-ad0a-15c5be73ba89 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 921.730278] env[62066]: DEBUG oslo_vmware.api [None req-0c3b1401-b067-4cd8-8015-7ffe20d4d016 tempest-ServersTestFqdnHostnames-1367628746 tempest-ServersTestFqdnHostnames-1367628746-project-member] Waiting for the task: (returnval){ [ 921.730278] env[62066]: value = "task-1341146" [ 921.730278] env[62066]: _type = "Task" [ 921.730278] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.745238] env[62066]: DEBUG oslo_vmware.api [None req-0c3b1401-b067-4cd8-8015-7ffe20d4d016 tempest-ServersTestFqdnHostnames-1367628746 tempest-ServersTestFqdnHostnames-1367628746-project-member] Task: {'id': task-1341146, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.814034] env[62066]: DEBUG oslo_vmware.api [None req-81c18bf6-1c8d-45a0-a66c-dcda0761f479 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Task: {'id': task-1341138, 'name': MoveVirtualDisk_Task} progress is 94%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.956867] env[62066]: DEBUG oslo_vmware.api [None req-ad6adfa4-e5e7-49bc-8c71-f4f4f133daea tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Task: {'id': task-1341145, 'name': ReconfigVM_Task, 'duration_secs': 0.464019} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.957436] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-ad6adfa4-e5e7-49bc-8c71-f4f4f133daea tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] [instance: 15afaea0-65de-4f32-851c-365003e5498f] Reconfigured VM instance instance-0000005a to attach disk [datastore1] 15afaea0-65de-4f32-851c-365003e5498f/15afaea0-65de-4f32-851c-365003e5498f.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 921.958565] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d560bd74-f3d7-46bb-9a76-30827960ef1a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.965394] env[62066]: DEBUG oslo_vmware.api [None req-ad6adfa4-e5e7-49bc-8c71-f4f4f133daea tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Waiting for the task: (returnval){ [ 921.965394] env[62066]: value = "task-1341147" [ 921.965394] env[62066]: _type = "Task" [ 921.965394] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.977287] env[62066]: DEBUG oslo_vmware.api [None req-ad6adfa4-e5e7-49bc-8c71-f4f4f133daea tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Task: {'id': task-1341147, 'name': Rename_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.166417] env[62066]: DEBUG nova.compute.utils [None req-e9681fdd-3744-4c1e-81e7-a2a0f833c385 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 922.170516] env[62066]: DEBUG nova.compute.manager [None req-e9681fdd-3744-4c1e-81e7-a2a0f833c385 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] [instance: a10a6dc0-7df7-40a1-888a-d1414dbeb1f7] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 922.170735] env[62066]: DEBUG nova.network.neutron [None req-e9681fdd-3744-4c1e-81e7-a2a0f833c385 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] [instance: a10a6dc0-7df7-40a1-888a-d1414dbeb1f7] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 922.214737] env[62066]: DEBUG oslo_concurrency.lockutils [None req-96e7ee9b-6f06-4817-a05d-ab2ef59667eb tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Releasing lock "refresh_cache-c8e0d47c-4421-4e00-9183-206fceeabc40" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 922.217496] env[62066]: DEBUG nova.policy [None req-e9681fdd-3744-4c1e-81e7-a2a0f833c385 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9d11ab2ea74442369fa1d93daf98392b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'dd30584efc254610b3939d1ba3806693', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 922.222926] env[62066]: INFO nova.compute.manager [-] [instance: f4a84152-f222-46a2-9a35-8a0a0078fc12] Took 1.61 seconds to deallocate network for instance. [ 922.241792] env[62066]: DEBUG oslo_vmware.api [None req-0c3b1401-b067-4cd8-8015-7ffe20d4d016 tempest-ServersTestFqdnHostnames-1367628746 tempest-ServersTestFqdnHostnames-1367628746-project-member] Task: {'id': task-1341146, 'name': PowerOffVM_Task, 'duration_secs': 0.22915} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.244348] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-0c3b1401-b067-4cd8-8015-7ffe20d4d016 tempest-ServersTestFqdnHostnames-1367628746 tempest-ServersTestFqdnHostnames-1367628746-project-member] [instance: afaca059-fc4e-4de2-8fa0-d2f226ea7051] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 922.244570] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-0c3b1401-b067-4cd8-8015-7ffe20d4d016 tempest-ServersTestFqdnHostnames-1367628746 tempest-ServersTestFqdnHostnames-1367628746-project-member] [instance: afaca059-fc4e-4de2-8fa0-d2f226ea7051] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 922.245254] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c9daa041-e088-4aba-af9d-0860ff30a847 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.291544] env[62066]: DEBUG nova.compute.manager [req-09c818c9-796d-4320-8971-fdff19507643 req-7b84baef-867e-42ae-9359-63a5a17657db service nova] [instance: f4a84152-f222-46a2-9a35-8a0a0078fc12] Received event network-vif-deleted-35ae652b-5fe5-468b-b4d4-75e727382145 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 922.305666] env[62066]: DEBUG oslo_vmware.api [None req-81c18bf6-1c8d-45a0-a66c-dcda0761f479 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Task: {'id': task-1341138, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.704656} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.306008] env[62066]: INFO nova.virt.vmwareapi.ds_util [None req-81c18bf6-1c8d-45a0-a66c-dcda0761f479 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Moved virtual disk from [datastore1] OSTACK_IMG_7d0cd4f2-d1f4-4ceb-9c76-75a77646fd93/OSTACK_IMG_7d0cd4f2-d1f4-4ceb-9c76-75a77646fd93.vmdk to [datastore1] devstack-image-cache_base/1ebd2ae1-0ae3-4486-822c-c75cafe27fe3/1ebd2ae1-0ae3-4486-822c-c75cafe27fe3.vmdk. [ 922.306171] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-81c18bf6-1c8d-45a0-a66c-dcda0761f479 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] [instance: 8ff6fbb9-c90f-498d-9a85-d220a8c2f794] Cleaning up location [datastore1] OSTACK_IMG_7d0cd4f2-d1f4-4ceb-9c76-75a77646fd93 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 922.306302] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-81c18bf6-1c8d-45a0-a66c-dcda0761f479 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Deleting the datastore file [datastore1] OSTACK_IMG_7d0cd4f2-d1f4-4ceb-9c76-75a77646fd93 {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 922.308804] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-140313ba-88cb-4189-b1e5-dc90d9a70c4d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.311995] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-0c3b1401-b067-4cd8-8015-7ffe20d4d016 tempest-ServersTestFqdnHostnames-1367628746 tempest-ServersTestFqdnHostnames-1367628746-project-member] [instance: afaca059-fc4e-4de2-8fa0-d2f226ea7051] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 922.312097] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-0c3b1401-b067-4cd8-8015-7ffe20d4d016 tempest-ServersTestFqdnHostnames-1367628746 tempest-ServersTestFqdnHostnames-1367628746-project-member] [instance: afaca059-fc4e-4de2-8fa0-d2f226ea7051] Deleting contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 922.312287] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-0c3b1401-b067-4cd8-8015-7ffe20d4d016 tempest-ServersTestFqdnHostnames-1367628746 tempest-ServersTestFqdnHostnames-1367628746-project-member] Deleting the datastore file [datastore2] afaca059-fc4e-4de2-8fa0-d2f226ea7051 {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 922.312931] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-efca9782-272f-4ddb-be73-3127e4dc6451 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.316780] env[62066]: DEBUG oslo_vmware.api [None req-81c18bf6-1c8d-45a0-a66c-dcda0761f479 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Waiting for the task: (returnval){ [ 922.316780] env[62066]: value = "task-1341149" [ 922.316780] env[62066]: _type = "Task" [ 922.316780] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.321111] env[62066]: DEBUG oslo_vmware.api [None req-0c3b1401-b067-4cd8-8015-7ffe20d4d016 tempest-ServersTestFqdnHostnames-1367628746 tempest-ServersTestFqdnHostnames-1367628746-project-member] Waiting for the task: (returnval){ [ 922.321111] env[62066]: value = "task-1341150" [ 922.321111] env[62066]: _type = "Task" [ 922.321111] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.327368] env[62066]: DEBUG oslo_vmware.api [None req-81c18bf6-1c8d-45a0-a66c-dcda0761f479 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Task: {'id': task-1341149, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.334721] env[62066]: DEBUG oslo_vmware.api [None req-0c3b1401-b067-4cd8-8015-7ffe20d4d016 tempest-ServersTestFqdnHostnames-1367628746 tempest-ServersTestFqdnHostnames-1367628746-project-member] Task: {'id': task-1341150, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.405799] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acfd8eb3-baf6-4023-bb01-cd8d6c3859b9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.413125] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3865dd0-d5c4-4839-ba2b-69d0b8f44e6f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.454178] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27034fe7-b2d4-45f1-8555-cc4a8b4f9cfa {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.462564] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ade4cb33-6da7-485a-be08-d3e612636e17 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.474875] env[62066]: DEBUG oslo_vmware.api [None req-ad6adfa4-e5e7-49bc-8c71-f4f4f133daea tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Task: {'id': task-1341147, 'name': Rename_Task, 'duration_secs': 0.373872} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.482533] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-ad6adfa4-e5e7-49bc-8c71-f4f4f133daea tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] [instance: 15afaea0-65de-4f32-851c-365003e5498f] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 922.483012] env[62066]: DEBUG nova.compute.provider_tree [None req-ab952be5-ba87-4d38-b219-8de60eb09bdf tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 922.484315] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-06c224e1-cae9-4d6a-8d5e-fd4cc85d0224 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.490177] env[62066]: DEBUG oslo_vmware.api [None req-ad6adfa4-e5e7-49bc-8c71-f4f4f133daea tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Waiting for the task: (returnval){ [ 922.490177] env[62066]: value = "task-1341151" [ 922.490177] env[62066]: _type = "Task" [ 922.490177] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.498605] env[62066]: DEBUG oslo_vmware.api [None req-ad6adfa4-e5e7-49bc-8c71-f4f4f133daea tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Task: {'id': task-1341151, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.648685] env[62066]: DEBUG nova.network.neutron [None req-e9681fdd-3744-4c1e-81e7-a2a0f833c385 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] [instance: a10a6dc0-7df7-40a1-888a-d1414dbeb1f7] Successfully created port: 26148c05-d41e-4564-a159-de3f90c20550 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 922.671588] env[62066]: DEBUG nova.compute.manager [None req-e9681fdd-3744-4c1e-81e7-a2a0f833c385 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] [instance: a10a6dc0-7df7-40a1-888a-d1414dbeb1f7] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 922.720711] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-96e7ee9b-6f06-4817-a05d-ab2ef59667eb tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: c8e0d47c-4421-4e00-9183-206fceeabc40] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 922.721281] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d2cbd272-d781-4a8d-99b9-db5e18a5853c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.728688] env[62066]: DEBUG oslo_vmware.api [None req-96e7ee9b-6f06-4817-a05d-ab2ef59667eb tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Waiting for the task: (returnval){ [ 922.728688] env[62066]: value = "task-1341152" [ 922.728688] env[62066]: _type = "Task" [ 922.728688] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.732318] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3bd520fa-2cd0-41e8-afc3-e67149da4105 tempest-ServerTagsTestJSON-2015212290 tempest-ServerTagsTestJSON-2015212290-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 922.741617] env[62066]: DEBUG oslo_vmware.api [None req-96e7ee9b-6f06-4817-a05d-ab2ef59667eb tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Task: {'id': task-1341152, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.832026] env[62066]: DEBUG oslo_vmware.api [None req-81c18bf6-1c8d-45a0-a66c-dcda0761f479 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Task: {'id': task-1341149, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.03399} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.832026] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-81c18bf6-1c8d-45a0-a66c-dcda0761f479 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 922.832026] env[62066]: DEBUG oslo_concurrency.lockutils [None req-81c18bf6-1c8d-45a0-a66c-dcda0761f479 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Releasing lock "[datastore1] devstack-image-cache_base/1ebd2ae1-0ae3-4486-822c-c75cafe27fe3/1ebd2ae1-0ae3-4486-822c-c75cafe27fe3.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 922.832026] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-81c18bf6-1c8d-45a0-a66c-dcda0761f479 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/1ebd2ae1-0ae3-4486-822c-c75cafe27fe3/1ebd2ae1-0ae3-4486-822c-c75cafe27fe3.vmdk to [datastore1] 8ff6fbb9-c90f-498d-9a85-d220a8c2f794/8ff6fbb9-c90f-498d-9a85-d220a8c2f794.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 922.832026] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-dc928416-e495-4573-8994-27ad88ee3219 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.843037] env[62066]: DEBUG oslo_vmware.api [None req-0c3b1401-b067-4cd8-8015-7ffe20d4d016 tempest-ServersTestFqdnHostnames-1367628746 tempest-ServersTestFqdnHostnames-1367628746-project-member] Task: {'id': task-1341150, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.147435} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.846344] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-0c3b1401-b067-4cd8-8015-7ffe20d4d016 tempest-ServersTestFqdnHostnames-1367628746 tempest-ServersTestFqdnHostnames-1367628746-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 922.847844] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-0c3b1401-b067-4cd8-8015-7ffe20d4d016 tempest-ServersTestFqdnHostnames-1367628746 tempest-ServersTestFqdnHostnames-1367628746-project-member] [instance: afaca059-fc4e-4de2-8fa0-d2f226ea7051] Deleted contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 922.848343] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-0c3b1401-b067-4cd8-8015-7ffe20d4d016 tempest-ServersTestFqdnHostnames-1367628746 tempest-ServersTestFqdnHostnames-1367628746-project-member] [instance: afaca059-fc4e-4de2-8fa0-d2f226ea7051] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 922.854465] env[62066]: INFO nova.compute.manager [None req-0c3b1401-b067-4cd8-8015-7ffe20d4d016 tempest-ServersTestFqdnHostnames-1367628746 tempest-ServersTestFqdnHostnames-1367628746-project-member] [instance: afaca059-fc4e-4de2-8fa0-d2f226ea7051] Took 1.15 seconds to destroy the instance on the hypervisor. [ 922.854465] env[62066]: DEBUG oslo.service.loopingcall [None req-0c3b1401-b067-4cd8-8015-7ffe20d4d016 tempest-ServersTestFqdnHostnames-1367628746 tempest-ServersTestFqdnHostnames-1367628746-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 922.855601] env[62066]: DEBUG nova.compute.manager [-] [instance: afaca059-fc4e-4de2-8fa0-d2f226ea7051] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 922.856171] env[62066]: DEBUG nova.network.neutron [-] [instance: afaca059-fc4e-4de2-8fa0-d2f226ea7051] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 922.858738] env[62066]: DEBUG oslo_vmware.api [None req-81c18bf6-1c8d-45a0-a66c-dcda0761f479 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Waiting for the task: (returnval){ [ 922.858738] env[62066]: value = "task-1341153" [ 922.858738] env[62066]: _type = "Task" [ 922.858738] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.874151] env[62066]: DEBUG oslo_vmware.api [None req-81c18bf6-1c8d-45a0-a66c-dcda0761f479 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Task: {'id': task-1341153, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.987527] env[62066]: DEBUG nova.scheduler.client.report [None req-ab952be5-ba87-4d38-b219-8de60eb09bdf tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 923.001198] env[62066]: DEBUG oslo_vmware.api [None req-ad6adfa4-e5e7-49bc-8c71-f4f4f133daea tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Task: {'id': task-1341151, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.103973] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-076268ef-b433-4366-94c4-0e0e4238849f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.124213] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-740993dc-fc35-4aa8-9a06-8f4c0e5f95e1 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 9c989152-ec69-478e-a0b0-62983852576c] Updating instance '9c989152-ec69-478e-a0b0-62983852576c' progress to 0 {{(pid=62066) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 923.241631] env[62066]: DEBUG oslo_vmware.api [None req-96e7ee9b-6f06-4817-a05d-ab2ef59667eb tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Task: {'id': task-1341152, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.374422] env[62066]: DEBUG oslo_vmware.api [None req-81c18bf6-1c8d-45a0-a66c-dcda0761f479 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Task: {'id': task-1341153, 'name': CopyVirtualDisk_Task} progress is 21%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.388919] env[62066]: DEBUG nova.compute.manager [req-0084a08d-ce22-4acd-8e95-3d8213fd47c1 req-0c6c133a-3203-4fd7-8ff2-01b859f426b8 service nova] [instance: afaca059-fc4e-4de2-8fa0-d2f226ea7051] Received event network-vif-deleted-725b6138-b39f-41c3-bf84-1a13a3e30fa9 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 923.389180] env[62066]: INFO nova.compute.manager [req-0084a08d-ce22-4acd-8e95-3d8213fd47c1 req-0c6c133a-3203-4fd7-8ff2-01b859f426b8 service nova] [instance: afaca059-fc4e-4de2-8fa0-d2f226ea7051] Neutron deleted interface 725b6138-b39f-41c3-bf84-1a13a3e30fa9; detaching it from the instance and deleting it from the info cache [ 923.389452] env[62066]: DEBUG nova.network.neutron [req-0084a08d-ce22-4acd-8e95-3d8213fd47c1 req-0c6c133a-3203-4fd7-8ff2-01b859f426b8 service nova] [instance: afaca059-fc4e-4de2-8fa0-d2f226ea7051] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 923.499030] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ab952be5-ba87-4d38-b219-8de60eb09bdf tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.833s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 923.500848] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b9bedc72-5782-477a-80b5-653c800704af tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.756s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 923.501598] env[62066]: DEBUG nova.objects.instance [None req-b9bedc72-5782-477a-80b5-653c800704af tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Lazy-loading 'resources' on Instance uuid 5e96def2-0cbd-4bd9-93f4-6a365a0142b0 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 923.510566] env[62066]: DEBUG oslo_vmware.api [None req-ad6adfa4-e5e7-49bc-8c71-f4f4f133daea tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Task: {'id': task-1341151, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.532083] env[62066]: INFO nova.scheduler.client.report [None req-ab952be5-ba87-4d38-b219-8de60eb09bdf tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Deleted allocations for instance 90c33434-c127-450b-9ff0-75181b4ac385 [ 923.631378] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-740993dc-fc35-4aa8-9a06-8f4c0e5f95e1 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 9c989152-ec69-478e-a0b0-62983852576c] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 923.631794] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4b7c2525-7697-4fe4-9fb4-29037a94a51c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.641910] env[62066]: DEBUG oslo_vmware.api [None req-740993dc-fc35-4aa8-9a06-8f4c0e5f95e1 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Waiting for the task: (returnval){ [ 923.641910] env[62066]: value = "task-1341154" [ 923.641910] env[62066]: _type = "Task" [ 923.641910] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.653903] env[62066]: DEBUG oslo_vmware.api [None req-740993dc-fc35-4aa8-9a06-8f4c0e5f95e1 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1341154, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.683020] env[62066]: DEBUG nova.compute.manager [None req-e9681fdd-3744-4c1e-81e7-a2a0f833c385 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] [instance: a10a6dc0-7df7-40a1-888a-d1414dbeb1f7] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 923.724137] env[62066]: DEBUG nova.virt.hardware [None req-e9681fdd-3744-4c1e-81e7-a2a0f833c385 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-23T13:40:41Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-23T13:40:23Z,direct_url=,disk_format='vmdk',id=50ff584c-3b50-4395-af07-3e66769bc9f7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='eb52f7069a374c61ae946f052007c6d9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-23T13:40:24Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 923.724419] env[62066]: DEBUG nova.virt.hardware [None req-e9681fdd-3744-4c1e-81e7-a2a0f833c385 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 923.724584] env[62066]: DEBUG nova.virt.hardware [None req-e9681fdd-3744-4c1e-81e7-a2a0f833c385 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 923.724777] env[62066]: DEBUG nova.virt.hardware [None req-e9681fdd-3744-4c1e-81e7-a2a0f833c385 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 923.725068] env[62066]: DEBUG nova.virt.hardware [None req-e9681fdd-3744-4c1e-81e7-a2a0f833c385 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 923.725259] env[62066]: DEBUG nova.virt.hardware [None req-e9681fdd-3744-4c1e-81e7-a2a0f833c385 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 923.725497] env[62066]: DEBUG nova.virt.hardware [None req-e9681fdd-3744-4c1e-81e7-a2a0f833c385 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 923.725663] env[62066]: DEBUG nova.virt.hardware [None req-e9681fdd-3744-4c1e-81e7-a2a0f833c385 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 923.725880] env[62066]: DEBUG nova.virt.hardware [None req-e9681fdd-3744-4c1e-81e7-a2a0f833c385 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 923.726093] env[62066]: DEBUG nova.virt.hardware [None req-e9681fdd-3744-4c1e-81e7-a2a0f833c385 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 923.726287] env[62066]: DEBUG nova.virt.hardware [None req-e9681fdd-3744-4c1e-81e7-a2a0f833c385 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 923.727621] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51a20232-d161-4324-83dc-bf6e87517f26 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.743529] env[62066]: DEBUG oslo_vmware.api [None req-96e7ee9b-6f06-4817-a05d-ab2ef59667eb tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Task: {'id': task-1341152, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.746935] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-daa853b8-a827-458b-ad11-5e10588c4836 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.866187] env[62066]: DEBUG nova.network.neutron [-] [instance: afaca059-fc4e-4de2-8fa0-d2f226ea7051] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 923.872050] env[62066]: DEBUG oslo_vmware.api [None req-81c18bf6-1c8d-45a0-a66c-dcda0761f479 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Task: {'id': task-1341153, 'name': CopyVirtualDisk_Task} progress is 40%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.892317] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f63c66c7-52d9-4794-a2e1-df6b51f38c3e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.904846] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d92555a2-561f-42a6-a18d-5e38988baf58 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.936147] env[62066]: DEBUG nova.compute.manager [req-0084a08d-ce22-4acd-8e95-3d8213fd47c1 req-0c6c133a-3203-4fd7-8ff2-01b859f426b8 service nova] [instance: afaca059-fc4e-4de2-8fa0-d2f226ea7051] Detach interface failed, port_id=725b6138-b39f-41c3-bf84-1a13a3e30fa9, reason: Instance afaca059-fc4e-4de2-8fa0-d2f226ea7051 could not be found. {{(pid=62066) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 924.009258] env[62066]: DEBUG oslo_vmware.api [None req-ad6adfa4-e5e7-49bc-8c71-f4f4f133daea tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Task: {'id': task-1341151, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.044218] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ab952be5-ba87-4d38-b219-8de60eb09bdf tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Lock "90c33434-c127-450b-9ff0-75181b4ac385" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.608s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 924.156330] env[62066]: DEBUG oslo_vmware.api [None req-740993dc-fc35-4aa8-9a06-8f4c0e5f95e1 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1341154, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.204503] env[62066]: DEBUG nova.network.neutron [None req-e9681fdd-3744-4c1e-81e7-a2a0f833c385 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] [instance: a10a6dc0-7df7-40a1-888a-d1414dbeb1f7] Successfully updated port: 26148c05-d41e-4564-a159-de3f90c20550 {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 924.248696] env[62066]: DEBUG oslo_vmware.api [None req-96e7ee9b-6f06-4817-a05d-ab2ef59667eb tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Task: {'id': task-1341152, 'name': PowerOnVM_Task, 'duration_secs': 1.468182} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.249022] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-96e7ee9b-6f06-4817-a05d-ab2ef59667eb tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: c8e0d47c-4421-4e00-9183-206fceeabc40] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 924.249202] env[62066]: DEBUG nova.compute.manager [None req-96e7ee9b-6f06-4817-a05d-ab2ef59667eb tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: c8e0d47c-4421-4e00-9183-206fceeabc40] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 924.250342] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4da1394f-2242-473d-8bc5-775d33c1630b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.261136] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe0e3bc4-ce2d-4ef3-99d0-202d0bc41e00 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.270506] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fdde8553-f9d1-4b41-a1e8-8e1f8ce1bb39 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.309250] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fe18710-ac91-4cd2-b81f-05ce42ed7dfd {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.318815] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4a49c51-7e97-44f8-808b-2b5ad61ef023 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.337127] env[62066]: DEBUG nova.compute.provider_tree [None req-b9bedc72-5782-477a-80b5-653c800704af tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 924.368789] env[62066]: INFO nova.compute.manager [-] [instance: afaca059-fc4e-4de2-8fa0-d2f226ea7051] Took 1.51 seconds to deallocate network for instance. [ 924.374400] env[62066]: DEBUG oslo_vmware.api [None req-81c18bf6-1c8d-45a0-a66c-dcda0761f479 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Task: {'id': task-1341153, 'name': CopyVirtualDisk_Task} progress is 63%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.408540] env[62066]: DEBUG nova.compute.manager [req-0473100f-3409-4878-9e71-5a27ccf218f8 req-3fdfd493-7db5-4d32-afd8-c4f42b1f41a8 service nova] [instance: a10a6dc0-7df7-40a1-888a-d1414dbeb1f7] Received event network-vif-plugged-26148c05-d41e-4564-a159-de3f90c20550 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 924.408782] env[62066]: DEBUG oslo_concurrency.lockutils [req-0473100f-3409-4878-9e71-5a27ccf218f8 req-3fdfd493-7db5-4d32-afd8-c4f42b1f41a8 service nova] Acquiring lock "a10a6dc0-7df7-40a1-888a-d1414dbeb1f7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 924.408946] env[62066]: DEBUG oslo_concurrency.lockutils [req-0473100f-3409-4878-9e71-5a27ccf218f8 req-3fdfd493-7db5-4d32-afd8-c4f42b1f41a8 service nova] Lock "a10a6dc0-7df7-40a1-888a-d1414dbeb1f7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 924.409564] env[62066]: DEBUG oslo_concurrency.lockutils [req-0473100f-3409-4878-9e71-5a27ccf218f8 req-3fdfd493-7db5-4d32-afd8-c4f42b1f41a8 service nova] Lock "a10a6dc0-7df7-40a1-888a-d1414dbeb1f7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 924.409564] env[62066]: DEBUG nova.compute.manager [req-0473100f-3409-4878-9e71-5a27ccf218f8 req-3fdfd493-7db5-4d32-afd8-c4f42b1f41a8 service nova] [instance: a10a6dc0-7df7-40a1-888a-d1414dbeb1f7] No waiting events found dispatching network-vif-plugged-26148c05-d41e-4564-a159-de3f90c20550 {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 924.409758] env[62066]: WARNING nova.compute.manager [req-0473100f-3409-4878-9e71-5a27ccf218f8 req-3fdfd493-7db5-4d32-afd8-c4f42b1f41a8 service nova] [instance: a10a6dc0-7df7-40a1-888a-d1414dbeb1f7] Received unexpected event network-vif-plugged-26148c05-d41e-4564-a159-de3f90c20550 for instance with vm_state building and task_state spawning. [ 924.410585] env[62066]: DEBUG nova.compute.manager [req-0473100f-3409-4878-9e71-5a27ccf218f8 req-3fdfd493-7db5-4d32-afd8-c4f42b1f41a8 service nova] [instance: a10a6dc0-7df7-40a1-888a-d1414dbeb1f7] Received event network-changed-26148c05-d41e-4564-a159-de3f90c20550 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 924.410585] env[62066]: DEBUG nova.compute.manager [req-0473100f-3409-4878-9e71-5a27ccf218f8 req-3fdfd493-7db5-4d32-afd8-c4f42b1f41a8 service nova] [instance: a10a6dc0-7df7-40a1-888a-d1414dbeb1f7] Refreshing instance network info cache due to event network-changed-26148c05-d41e-4564-a159-de3f90c20550. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 924.410585] env[62066]: DEBUG oslo_concurrency.lockutils [req-0473100f-3409-4878-9e71-5a27ccf218f8 req-3fdfd493-7db5-4d32-afd8-c4f42b1f41a8 service nova] Acquiring lock "refresh_cache-a10a6dc0-7df7-40a1-888a-d1414dbeb1f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 924.410585] env[62066]: DEBUG oslo_concurrency.lockutils [req-0473100f-3409-4878-9e71-5a27ccf218f8 req-3fdfd493-7db5-4d32-afd8-c4f42b1f41a8 service nova] Acquired lock "refresh_cache-a10a6dc0-7df7-40a1-888a-d1414dbeb1f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 924.410882] env[62066]: DEBUG nova.network.neutron [req-0473100f-3409-4878-9e71-5a27ccf218f8 req-3fdfd493-7db5-4d32-afd8-c4f42b1f41a8 service nova] [instance: a10a6dc0-7df7-40a1-888a-d1414dbeb1f7] Refreshing network info cache for port 26148c05-d41e-4564-a159-de3f90c20550 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 924.505627] env[62066]: DEBUG oslo_vmware.api [None req-ad6adfa4-e5e7-49bc-8c71-f4f4f133daea tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Task: {'id': task-1341151, 'name': PowerOnVM_Task, 'duration_secs': 1.705515} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.505627] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-ad6adfa4-e5e7-49bc-8c71-f4f4f133daea tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] [instance: 15afaea0-65de-4f32-851c-365003e5498f] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 924.506228] env[62066]: INFO nova.compute.manager [None req-ad6adfa4-e5e7-49bc-8c71-f4f4f133daea tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] [instance: 15afaea0-65de-4f32-851c-365003e5498f] Took 7.34 seconds to spawn the instance on the hypervisor. [ 924.506228] env[62066]: DEBUG nova.compute.manager [None req-ad6adfa4-e5e7-49bc-8c71-f4f4f133daea tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] [instance: 15afaea0-65de-4f32-851c-365003e5498f] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 924.507255] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-690eab48-8b57-4350-8b93-a24bcfc9de30 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.654987] env[62066]: DEBUG oslo_vmware.api [None req-740993dc-fc35-4aa8-9a06-8f4c0e5f95e1 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1341154, 'name': PowerOffVM_Task, 'duration_secs': 0.710989} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.654987] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-740993dc-fc35-4aa8-9a06-8f4c0e5f95e1 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 9c989152-ec69-478e-a0b0-62983852576c] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 924.655233] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-740993dc-fc35-4aa8-9a06-8f4c0e5f95e1 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 9c989152-ec69-478e-a0b0-62983852576c] Updating instance '9c989152-ec69-478e-a0b0-62983852576c' progress to 17 {{(pid=62066) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 924.706553] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e9681fdd-3744-4c1e-81e7-a2a0f833c385 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Acquiring lock "refresh_cache-a10a6dc0-7df7-40a1-888a-d1414dbeb1f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 924.841266] env[62066]: DEBUG nova.scheduler.client.report [None req-b9bedc72-5782-477a-80b5-653c800704af tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 924.875843] env[62066]: DEBUG oslo_vmware.api [None req-81c18bf6-1c8d-45a0-a66c-dcda0761f479 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Task: {'id': task-1341153, 'name': CopyVirtualDisk_Task} progress is 83%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.879063] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0c3b1401-b067-4cd8-8015-7ffe20d4d016 tempest-ServersTestFqdnHostnames-1367628746 tempest-ServersTestFqdnHostnames-1367628746-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 924.965368] env[62066]: DEBUG nova.network.neutron [req-0473100f-3409-4878-9e71-5a27ccf218f8 req-3fdfd493-7db5-4d32-afd8-c4f42b1f41a8 service nova] [instance: a10a6dc0-7df7-40a1-888a-d1414dbeb1f7] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 925.026160] env[62066]: INFO nova.compute.manager [None req-ad6adfa4-e5e7-49bc-8c71-f4f4f133daea tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] [instance: 15afaea0-65de-4f32-851c-365003e5498f] Took 26.81 seconds to build instance. [ 925.057097] env[62066]: DEBUG nova.network.neutron [req-0473100f-3409-4878-9e71-5a27ccf218f8 req-3fdfd493-7db5-4d32-afd8-c4f42b1f41a8 service nova] [instance: a10a6dc0-7df7-40a1-888a-d1414dbeb1f7] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 925.162983] env[62066]: DEBUG nova.virt.hardware [None req-740993dc-fc35-4aa8-9a06-8f4c0e5f95e1 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-23T13:40:43Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=50ff584c-3b50-4395-af07-3e66769bc9f7,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 925.163281] env[62066]: DEBUG nova.virt.hardware [None req-740993dc-fc35-4aa8-9a06-8f4c0e5f95e1 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 925.163357] env[62066]: DEBUG nova.virt.hardware [None req-740993dc-fc35-4aa8-9a06-8f4c0e5f95e1 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 925.163540] env[62066]: DEBUG nova.virt.hardware [None req-740993dc-fc35-4aa8-9a06-8f4c0e5f95e1 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 925.163693] env[62066]: DEBUG nova.virt.hardware [None req-740993dc-fc35-4aa8-9a06-8f4c0e5f95e1 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 925.163846] env[62066]: DEBUG nova.virt.hardware [None req-740993dc-fc35-4aa8-9a06-8f4c0e5f95e1 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 925.164153] env[62066]: DEBUG nova.virt.hardware [None req-740993dc-fc35-4aa8-9a06-8f4c0e5f95e1 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 925.164264] env[62066]: DEBUG nova.virt.hardware [None req-740993dc-fc35-4aa8-9a06-8f4c0e5f95e1 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 925.164423] env[62066]: DEBUG nova.virt.hardware [None req-740993dc-fc35-4aa8-9a06-8f4c0e5f95e1 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 925.164577] env[62066]: DEBUG nova.virt.hardware [None req-740993dc-fc35-4aa8-9a06-8f4c0e5f95e1 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 925.164752] env[62066]: DEBUG nova.virt.hardware [None req-740993dc-fc35-4aa8-9a06-8f4c0e5f95e1 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 925.169829] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-db82617a-b1c2-43cc-bca8-2cba91e1424a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.187864] env[62066]: DEBUG oslo_vmware.api [None req-740993dc-fc35-4aa8-9a06-8f4c0e5f95e1 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Waiting for the task: (returnval){ [ 925.187864] env[62066]: value = "task-1341155" [ 925.187864] env[62066]: _type = "Task" [ 925.187864] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.198713] env[62066]: DEBUG oslo_vmware.api [None req-740993dc-fc35-4aa8-9a06-8f4c0e5f95e1 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1341155, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.355362] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b9bedc72-5782-477a-80b5-653c800704af tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.854s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 925.358380] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 8.328s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 925.372479] env[62066]: DEBUG oslo_vmware.api [None req-81c18bf6-1c8d-45a0-a66c-dcda0761f479 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Task: {'id': task-1341153, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.368461} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.372479] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-81c18bf6-1c8d-45a0-a66c-dcda0761f479 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/1ebd2ae1-0ae3-4486-822c-c75cafe27fe3/1ebd2ae1-0ae3-4486-822c-c75cafe27fe3.vmdk to [datastore1] 8ff6fbb9-c90f-498d-9a85-d220a8c2f794/8ff6fbb9-c90f-498d-9a85-d220a8c2f794.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 925.373289] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53e6cd1c-5e65-4a6a-990a-36930d051a82 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.376807] env[62066]: INFO nova.scheduler.client.report [None req-b9bedc72-5782-477a-80b5-653c800704af tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Deleted allocations for instance 5e96def2-0cbd-4bd9-93f4-6a365a0142b0 [ 925.402694] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-81c18bf6-1c8d-45a0-a66c-dcda0761f479 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] [instance: 8ff6fbb9-c90f-498d-9a85-d220a8c2f794] Reconfiguring VM instance instance-00000045 to attach disk [datastore1] 8ff6fbb9-c90f-498d-9a85-d220a8c2f794/8ff6fbb9-c90f-498d-9a85-d220a8c2f794.vmdk or device None with type streamOptimized {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 925.403056] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-96ba02eb-ce43-481b-aa75-40c49b6926d6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.425775] env[62066]: DEBUG oslo_vmware.api [None req-81c18bf6-1c8d-45a0-a66c-dcda0761f479 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Waiting for the task: (returnval){ [ 925.425775] env[62066]: value = "task-1341156" [ 925.425775] env[62066]: _type = "Task" [ 925.425775] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.437408] env[62066]: DEBUG oslo_concurrency.lockutils [None req-bb369cbc-f881-4f95-b841-517845c5cb59 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Acquiring lock "5693eea2-1cac-45d5-aa27-58526b65136e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 925.437648] env[62066]: DEBUG oslo_concurrency.lockutils [None req-bb369cbc-f881-4f95-b841-517845c5cb59 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Lock "5693eea2-1cac-45d5-aa27-58526b65136e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 925.438709] env[62066]: DEBUG oslo_vmware.api [None req-81c18bf6-1c8d-45a0-a66c-dcda0761f479 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Task: {'id': task-1341156, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.528767] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ad6adfa4-e5e7-49bc-8c71-f4f4f133daea tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Lock "15afaea0-65de-4f32-851c-365003e5498f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 28.321s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 925.559722] env[62066]: DEBUG oslo_concurrency.lockutils [req-0473100f-3409-4878-9e71-5a27ccf218f8 req-3fdfd493-7db5-4d32-afd8-c4f42b1f41a8 service nova] Releasing lock "refresh_cache-a10a6dc0-7df7-40a1-888a-d1414dbeb1f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 925.560286] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e9681fdd-3744-4c1e-81e7-a2a0f833c385 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Acquired lock "refresh_cache-a10a6dc0-7df7-40a1-888a-d1414dbeb1f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 925.560400] env[62066]: DEBUG nova.network.neutron [None req-e9681fdd-3744-4c1e-81e7-a2a0f833c385 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] [instance: a10a6dc0-7df7-40a1-888a-d1414dbeb1f7] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 925.699363] env[62066]: DEBUG oslo_vmware.api [None req-740993dc-fc35-4aa8-9a06-8f4c0e5f95e1 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1341155, 'name': ReconfigVM_Task, 'duration_secs': 0.457789} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.699592] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-740993dc-fc35-4aa8-9a06-8f4c0e5f95e1 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 9c989152-ec69-478e-a0b0-62983852576c] Updating instance '9c989152-ec69-478e-a0b0-62983852576c' progress to 33 {{(pid=62066) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 925.884819] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b9bedc72-5782-477a-80b5-653c800704af tempest-ServerRescueTestJSON-1613145410 tempest-ServerRescueTestJSON-1613145410-project-member] Lock "5e96def2-0cbd-4bd9-93f4-6a365a0142b0" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.988s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 925.936270] env[62066]: DEBUG oslo_vmware.api [None req-81c18bf6-1c8d-45a0-a66c-dcda0761f479 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Task: {'id': task-1341156, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.940408] env[62066]: DEBUG nova.compute.manager [None req-bb369cbc-f881-4f95-b841-517845c5cb59 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: 5693eea2-1cac-45d5-aa27-58526b65136e] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 926.094827] env[62066]: DEBUG nova.network.neutron [None req-e9681fdd-3744-4c1e-81e7-a2a0f833c385 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] [instance: a10a6dc0-7df7-40a1-888a-d1414dbeb1f7] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 926.206144] env[62066]: DEBUG nova.virt.hardware [None req-740993dc-fc35-4aa8-9a06-8f4c0e5f95e1 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-23T13:40:41Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=50ff584c-3b50-4395-af07-3e66769bc9f7,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 926.206585] env[62066]: DEBUG nova.virt.hardware [None req-740993dc-fc35-4aa8-9a06-8f4c0e5f95e1 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 926.206857] env[62066]: DEBUG nova.virt.hardware [None req-740993dc-fc35-4aa8-9a06-8f4c0e5f95e1 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 926.207138] env[62066]: DEBUG nova.virt.hardware [None req-740993dc-fc35-4aa8-9a06-8f4c0e5f95e1 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 926.207364] env[62066]: DEBUG nova.virt.hardware [None req-740993dc-fc35-4aa8-9a06-8f4c0e5f95e1 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 926.207582] env[62066]: DEBUG nova.virt.hardware [None req-740993dc-fc35-4aa8-9a06-8f4c0e5f95e1 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 926.207867] env[62066]: DEBUG nova.virt.hardware [None req-740993dc-fc35-4aa8-9a06-8f4c0e5f95e1 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 926.208129] env[62066]: DEBUG nova.virt.hardware [None req-740993dc-fc35-4aa8-9a06-8f4c0e5f95e1 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 926.208401] env[62066]: DEBUG nova.virt.hardware [None req-740993dc-fc35-4aa8-9a06-8f4c0e5f95e1 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 926.208648] env[62066]: DEBUG nova.virt.hardware [None req-740993dc-fc35-4aa8-9a06-8f4c0e5f95e1 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 926.208996] env[62066]: DEBUG nova.virt.hardware [None req-740993dc-fc35-4aa8-9a06-8f4c0e5f95e1 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 926.214487] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-740993dc-fc35-4aa8-9a06-8f4c0e5f95e1 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 9c989152-ec69-478e-a0b0-62983852576c] Reconfiguring VM instance instance-00000056 to detach disk 2000 {{(pid=62066) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 926.214848] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5ced5852-9df4-4ea5-af1b-f6c8b595790e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.229631] env[62066]: DEBUG nova.network.neutron [None req-e9681fdd-3744-4c1e-81e7-a2a0f833c385 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] [instance: a10a6dc0-7df7-40a1-888a-d1414dbeb1f7] Updating instance_info_cache with network_info: [{"id": "26148c05-d41e-4564-a159-de3f90c20550", "address": "fa:16:3e:c8:14:c9", "network": {"id": "4114ddfe-818f-47a5-bb6f-44a4c22517f6", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-676131672-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dd30584efc254610b3939d1ba3806693", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a485857d-7086-4dcf-9d65-d0dcd177fcb0", "external-id": "nsx-vlan-transportzone-232", "segmentation_id": 232, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap26148c05-d4", "ovs_interfaceid": "26148c05-d41e-4564-a159-de3f90c20550", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 926.235903] env[62066]: DEBUG oslo_vmware.api [None req-740993dc-fc35-4aa8-9a06-8f4c0e5f95e1 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Waiting for the task: (returnval){ [ 926.235903] env[62066]: value = "task-1341157" [ 926.235903] env[62066]: _type = "Task" [ 926.235903] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.245083] env[62066]: INFO nova.compute.manager [None req-9f08ab51-33e2-4c97-838b-004b2127a87f tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] [instance: 15afaea0-65de-4f32-851c-365003e5498f] Rebuilding instance [ 926.247231] env[62066]: DEBUG oslo_vmware.api [None req-740993dc-fc35-4aa8-9a06-8f4c0e5f95e1 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1341157, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.291378] env[62066]: DEBUG nova.compute.manager [None req-9f08ab51-33e2-4c97-838b-004b2127a87f tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] [instance: 15afaea0-65de-4f32-851c-365003e5498f] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 926.292394] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34d271d4-b993-43cc-b3a8-9339d7108e5d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.368865] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Applying migration context for instance 9c989152-ec69-478e-a0b0-62983852576c as it has an incoming, in-progress migration aba80e34-5809-409a-9dcc-c487ec0ebe3f. Migration status is migrating {{(pid=62066) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 926.370396] env[62066]: INFO nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] [instance: 9c989152-ec69-478e-a0b0-62983852576c] Updating resource usage from migration aba80e34-5809-409a-9dcc-c487ec0ebe3f [ 926.390844] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Instance 02fb3e92-5dd6-4b1e-a6e4-d60d3fc85b07 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 926.390844] env[62066]: WARNING nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Instance b08f27b1-429e-4d05-9a95-f45fd0cb2904 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 926.390844] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Instance c8e0d47c-4421-4e00-9183-206fceeabc40 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 926.390844] env[62066]: WARNING nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Instance afaca059-fc4e-4de2-8fa0-d2f226ea7051 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 926.390844] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Instance b718b09b-a7fa-445b-8be3-5b3ebca210a2 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 926.390844] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Instance 8ff6fbb9-c90f-498d-9a85-d220a8c2f794 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 926.391666] env[62066]: WARNING nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Instance f4a84152-f222-46a2-9a35-8a0a0078fc12 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 926.391666] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Instance 4e39b9b0-4ab0-4f48-8a3f-6bc1895597eb actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 926.391666] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Instance 15afaea0-65de-4f32-851c-365003e5498f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 926.391666] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Migration aba80e34-5809-409a-9dcc-c487ec0ebe3f is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 926.391666] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Instance 9c989152-ec69-478e-a0b0-62983852576c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 926.391666] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Instance a10a6dc0-7df7-40a1-888a-d1414dbeb1f7 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 926.436829] env[62066]: DEBUG oslo_vmware.api [None req-81c18bf6-1c8d-45a0-a66c-dcda0761f479 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Task: {'id': task-1341156, 'name': ReconfigVM_Task, 'duration_secs': 0.62542} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.437214] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-81c18bf6-1c8d-45a0-a66c-dcda0761f479 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] [instance: 8ff6fbb9-c90f-498d-9a85-d220a8c2f794] Reconfigured VM instance instance-00000045 to attach disk [datastore1] 8ff6fbb9-c90f-498d-9a85-d220a8c2f794/8ff6fbb9-c90f-498d-9a85-d220a8c2f794.vmdk or device None with type streamOptimized {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 926.437806] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3ef79c3d-380c-4e6a-a64d-f49098ebb2fe {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.444152] env[62066]: DEBUG oslo_vmware.api [None req-81c18bf6-1c8d-45a0-a66c-dcda0761f479 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Waiting for the task: (returnval){ [ 926.444152] env[62066]: value = "task-1341158" [ 926.444152] env[62066]: _type = "Task" [ 926.444152] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.455547] env[62066]: DEBUG oslo_vmware.api [None req-81c18bf6-1c8d-45a0-a66c-dcda0761f479 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Task: {'id': task-1341158, 'name': Rename_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.461535] env[62066]: DEBUG oslo_concurrency.lockutils [None req-bb369cbc-f881-4f95-b841-517845c5cb59 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 926.732229] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e9681fdd-3744-4c1e-81e7-a2a0f833c385 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Releasing lock "refresh_cache-a10a6dc0-7df7-40a1-888a-d1414dbeb1f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 926.732585] env[62066]: DEBUG nova.compute.manager [None req-e9681fdd-3744-4c1e-81e7-a2a0f833c385 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] [instance: a10a6dc0-7df7-40a1-888a-d1414dbeb1f7] Instance network_info: |[{"id": "26148c05-d41e-4564-a159-de3f90c20550", "address": "fa:16:3e:c8:14:c9", "network": {"id": "4114ddfe-818f-47a5-bb6f-44a4c22517f6", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-676131672-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dd30584efc254610b3939d1ba3806693", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a485857d-7086-4dcf-9d65-d0dcd177fcb0", "external-id": "nsx-vlan-transportzone-232", "segmentation_id": 232, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap26148c05-d4", "ovs_interfaceid": "26148c05-d41e-4564-a159-de3f90c20550", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 926.733126] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-e9681fdd-3744-4c1e-81e7-a2a0f833c385 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] [instance: a10a6dc0-7df7-40a1-888a-d1414dbeb1f7] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c8:14:c9', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a485857d-7086-4dcf-9d65-d0dcd177fcb0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '26148c05-d41e-4564-a159-de3f90c20550', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 926.741233] env[62066]: DEBUG oslo.service.loopingcall [None req-e9681fdd-3744-4c1e-81e7-a2a0f833c385 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 926.741562] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a10a6dc0-7df7-40a1-888a-d1414dbeb1f7] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 926.744742] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-842aa02e-57e6-4709-a4b4-43d2433c82fc {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.766192] env[62066]: DEBUG oslo_vmware.api [None req-740993dc-fc35-4aa8-9a06-8f4c0e5f95e1 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1341157, 'name': ReconfigVM_Task, 'duration_secs': 0.400493} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.767362] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-740993dc-fc35-4aa8-9a06-8f4c0e5f95e1 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 9c989152-ec69-478e-a0b0-62983852576c] Reconfigured VM instance instance-00000056 to detach disk 2000 {{(pid=62066) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 926.767628] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 926.767628] env[62066]: value = "task-1341159" [ 926.767628] env[62066]: _type = "Task" [ 926.767628] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.768379] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8faa6b68-98c3-4e46-afd3-3a2c6cb865e5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.797098] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-740993dc-fc35-4aa8-9a06-8f4c0e5f95e1 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 9c989152-ec69-478e-a0b0-62983852576c] Reconfiguring VM instance instance-00000056 to attach disk [datastore2] 9c989152-ec69-478e-a0b0-62983852576c/9c989152-ec69-478e-a0b0-62983852576c.vmdk or device None with type thin {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 926.800848] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e6c89c9f-94e4-49ac-8878-e93a2dcd921e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.814738] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1341159, 'name': CreateVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.816048] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-9f08ab51-33e2-4c97-838b-004b2127a87f tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] [instance: 15afaea0-65de-4f32-851c-365003e5498f] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 926.816342] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-031ab8ec-0194-49da-b9da-e92f960c9cf3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.826881] env[62066]: DEBUG oslo_vmware.api [None req-9f08ab51-33e2-4c97-838b-004b2127a87f tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Waiting for the task: (returnval){ [ 926.826881] env[62066]: value = "task-1341160" [ 926.826881] env[62066]: _type = "Task" [ 926.826881] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.827162] env[62066]: DEBUG oslo_vmware.api [None req-740993dc-fc35-4aa8-9a06-8f4c0e5f95e1 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Waiting for the task: (returnval){ [ 926.827162] env[62066]: value = "task-1341161" [ 926.827162] env[62066]: _type = "Task" [ 926.827162] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.841857] env[62066]: DEBUG oslo_vmware.api [None req-740993dc-fc35-4aa8-9a06-8f4c0e5f95e1 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1341161, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.845269] env[62066]: DEBUG oslo_vmware.api [None req-9f08ab51-33e2-4c97-838b-004b2127a87f tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Task: {'id': task-1341160, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.894432] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Instance 20987cb6-e12c-48c5-8fae-d990b1bb6b8a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 926.955636] env[62066]: DEBUG oslo_vmware.api [None req-81c18bf6-1c8d-45a0-a66c-dcda0761f479 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Task: {'id': task-1341158, 'name': Rename_Task, 'duration_secs': 0.349495} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.956144] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-81c18bf6-1c8d-45a0-a66c-dcda0761f479 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] [instance: 8ff6fbb9-c90f-498d-9a85-d220a8c2f794] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 926.956557] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-970a555f-16ba-4419-a24c-93f99622a4e2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.963867] env[62066]: DEBUG oslo_vmware.api [None req-81c18bf6-1c8d-45a0-a66c-dcda0761f479 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Waiting for the task: (returnval){ [ 926.963867] env[62066]: value = "task-1341162" [ 926.963867] env[62066]: _type = "Task" [ 926.963867] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.973250] env[62066]: DEBUG oslo_vmware.api [None req-81c18bf6-1c8d-45a0-a66c-dcda0761f479 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Task: {'id': task-1341162, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.280923] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1341159, 'name': CreateVM_Task, 'duration_secs': 0.369196} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.281104] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a10a6dc0-7df7-40a1-888a-d1414dbeb1f7] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 927.281808] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e9681fdd-3744-4c1e-81e7-a2a0f833c385 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 927.281978] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e9681fdd-3744-4c1e-81e7-a2a0f833c385 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Acquired lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 927.282603] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e9681fdd-3744-4c1e-81e7-a2a0f833c385 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 927.282859] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f0aa4e66-9e89-483e-a89d-60d2c6d5f149 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.287220] env[62066]: DEBUG oslo_vmware.api [None req-e9681fdd-3744-4c1e-81e7-a2a0f833c385 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Waiting for the task: (returnval){ [ 927.287220] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]52dd2c5a-2774-3dcd-67f2-92dd93a12d2f" [ 927.287220] env[62066]: _type = "Task" [ 927.287220] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.294631] env[62066]: DEBUG oslo_vmware.api [None req-e9681fdd-3744-4c1e-81e7-a2a0f833c385 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52dd2c5a-2774-3dcd-67f2-92dd93a12d2f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.339697] env[62066]: DEBUG oslo_vmware.api [None req-9f08ab51-33e2-4c97-838b-004b2127a87f tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Task: {'id': task-1341160, 'name': PowerOffVM_Task, 'duration_secs': 0.127325} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.343140] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-9f08ab51-33e2-4c97-838b-004b2127a87f tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] [instance: 15afaea0-65de-4f32-851c-365003e5498f] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 927.343442] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-9f08ab51-33e2-4c97-838b-004b2127a87f tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] [instance: 15afaea0-65de-4f32-851c-365003e5498f] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 927.343756] env[62066]: DEBUG oslo_vmware.api [None req-740993dc-fc35-4aa8-9a06-8f4c0e5f95e1 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1341161, 'name': ReconfigVM_Task, 'duration_secs': 0.388384} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.344478] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86883119-12b4-433f-88e7-31f34b330950 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.347161] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-740993dc-fc35-4aa8-9a06-8f4c0e5f95e1 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 9c989152-ec69-478e-a0b0-62983852576c] Reconfigured VM instance instance-00000056 to attach disk [datastore2] 9c989152-ec69-478e-a0b0-62983852576c/9c989152-ec69-478e-a0b0-62983852576c.vmdk or device None with type thin {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 927.347438] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-740993dc-fc35-4aa8-9a06-8f4c0e5f95e1 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 9c989152-ec69-478e-a0b0-62983852576c] Updating instance '9c989152-ec69-478e-a0b0-62983852576c' progress to 50 {{(pid=62066) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 927.354976] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-9f08ab51-33e2-4c97-838b-004b2127a87f tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] [instance: 15afaea0-65de-4f32-851c-365003e5498f] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 927.355241] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f39855a0-c605-4d25-99c8-a617aa2f3c7c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.379775] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-9f08ab51-33e2-4c97-838b-004b2127a87f tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] [instance: 15afaea0-65de-4f32-851c-365003e5498f] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 927.379983] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-9f08ab51-33e2-4c97-838b-004b2127a87f tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] [instance: 15afaea0-65de-4f32-851c-365003e5498f] Deleting contents of the VM from datastore datastore1 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 927.380218] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-9f08ab51-33e2-4c97-838b-004b2127a87f tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Deleting the datastore file [datastore1] 15afaea0-65de-4f32-851c-365003e5498f {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 927.380502] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9684983c-d1ff-41d7-9606-bbca79be9d6b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.387205] env[62066]: DEBUG oslo_vmware.api [None req-9f08ab51-33e2-4c97-838b-004b2127a87f tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Waiting for the task: (returnval){ [ 927.387205] env[62066]: value = "task-1341164" [ 927.387205] env[62066]: _type = "Task" [ 927.387205] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.395984] env[62066]: DEBUG oslo_vmware.api [None req-9f08ab51-33e2-4c97-838b-004b2127a87f tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Task: {'id': task-1341164, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.397759] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Instance 651e5142-7ab0-4040-97b4-4599bd4e419a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 927.474832] env[62066]: DEBUG oslo_vmware.api [None req-81c18bf6-1c8d-45a0-a66c-dcda0761f479 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Task: {'id': task-1341162, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.797283] env[62066]: DEBUG oslo_vmware.api [None req-e9681fdd-3744-4c1e-81e7-a2a0f833c385 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52dd2c5a-2774-3dcd-67f2-92dd93a12d2f, 'name': SearchDatastore_Task, 'duration_secs': 0.011763} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.797606] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e9681fdd-3744-4c1e-81e7-a2a0f833c385 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Releasing lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 927.797847] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-e9681fdd-3744-4c1e-81e7-a2a0f833c385 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] [instance: a10a6dc0-7df7-40a1-888a-d1414dbeb1f7] Processing image 50ff584c-3b50-4395-af07-3e66769bc9f7 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 927.798128] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e9681fdd-3744-4c1e-81e7-a2a0f833c385 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 927.798288] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e9681fdd-3744-4c1e-81e7-a2a0f833c385 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Acquired lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 927.798448] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-e9681fdd-3744-4c1e-81e7-a2a0f833c385 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 927.799060] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3dece732-ee40-40b1-a0c1-43ce634c55d9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.807732] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-e9681fdd-3744-4c1e-81e7-a2a0f833c385 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 927.807914] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-e9681fdd-3744-4c1e-81e7-a2a0f833c385 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 927.808639] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cbbe560e-a748-4fbf-9558-aa16e7f411d2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.813596] env[62066]: DEBUG oslo_vmware.api [None req-e9681fdd-3744-4c1e-81e7-a2a0f833c385 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Waiting for the task: (returnval){ [ 927.813596] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]523a2b77-19b1-bd7a-a008-dc0bb3e112df" [ 927.813596] env[62066]: _type = "Task" [ 927.813596] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.820576] env[62066]: DEBUG oslo_vmware.api [None req-e9681fdd-3744-4c1e-81e7-a2a0f833c385 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]523a2b77-19b1-bd7a-a008-dc0bb3e112df, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.855466] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eab2764f-6791-4828-bc7e-1fb5e6e92fcb {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.876503] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1d21495-dfd5-4e61-a457-7f2ec6458aaf {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.895942] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-740993dc-fc35-4aa8-9a06-8f4c0e5f95e1 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 9c989152-ec69-478e-a0b0-62983852576c] Updating instance '9c989152-ec69-478e-a0b0-62983852576c' progress to 67 {{(pid=62066) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 927.902788] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Instance 5693eea2-1cac-45d5-aa27-58526b65136e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 927.903045] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Total usable vcpus: 48, total allocated vcpus: 9 {{(pid=62066) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 927.903199] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2304MB phys_disk=200GB used_disk=9GB total_vcpus=48 used_vcpus=9 pci_stats=[] {{(pid=62066) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 927.910661] env[62066]: DEBUG oslo_vmware.api [None req-9f08ab51-33e2-4c97-838b-004b2127a87f tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Task: {'id': task-1341164, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.089754} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.911272] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-9f08ab51-33e2-4c97-838b-004b2127a87f tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 927.911272] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-9f08ab51-33e2-4c97-838b-004b2127a87f tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] [instance: 15afaea0-65de-4f32-851c-365003e5498f] Deleted contents of the VM from datastore datastore1 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 927.911272] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-9f08ab51-33e2-4c97-838b-004b2127a87f tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] [instance: 15afaea0-65de-4f32-851c-365003e5498f] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 927.976591] env[62066]: DEBUG oslo_vmware.api [None req-81c18bf6-1c8d-45a0-a66c-dcda0761f479 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Task: {'id': task-1341162, 'name': PowerOnVM_Task, 'duration_secs': 0.629474} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.978905] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-81c18bf6-1c8d-45a0-a66c-dcda0761f479 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] [instance: 8ff6fbb9-c90f-498d-9a85-d220a8c2f794] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 928.087809] env[62066]: DEBUG nova.compute.manager [None req-81c18bf6-1c8d-45a0-a66c-dcda0761f479 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] [instance: 8ff6fbb9-c90f-498d-9a85-d220a8c2f794] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 928.088816] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb82865a-8d8e-41e2-84e1-438723b749cf {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.099712] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fdb3727d-caf7-47db-adb5-1e23fe9cc109 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.106788] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5803acb4-8167-4df9-9b6f-1b5601b85432 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.140386] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a5bc8b5-d5d3-4706-9093-74768ad09786 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.148883] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83639054-dde5-4e9c-8803-6c5054a8fef9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.162812] env[62066]: DEBUG nova.compute.provider_tree [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 928.324297] env[62066]: DEBUG oslo_vmware.api [None req-e9681fdd-3744-4c1e-81e7-a2a0f833c385 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]523a2b77-19b1-bd7a-a008-dc0bb3e112df, 'name': SearchDatastore_Task, 'duration_secs': 0.014248} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.325092] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7447edc5-ac5b-463b-a569-a1d6d77c311d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.330509] env[62066]: DEBUG oslo_vmware.api [None req-e9681fdd-3744-4c1e-81e7-a2a0f833c385 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Waiting for the task: (returnval){ [ 928.330509] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]52296e63-3346-efee-bf9f-6e187950cef8" [ 928.330509] env[62066]: _type = "Task" [ 928.330509] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.337904] env[62066]: DEBUG oslo_vmware.api [None req-e9681fdd-3744-4c1e-81e7-a2a0f833c385 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52296e63-3346-efee-bf9f-6e187950cef8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.440147] env[62066]: DEBUG nova.network.neutron [None req-740993dc-fc35-4aa8-9a06-8f4c0e5f95e1 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 9c989152-ec69-478e-a0b0-62983852576c] Port 1c0d0075-7f22-441e-a283-21e9b5c5aecc binding to destination host cpu-1 is already ACTIVE {{(pid=62066) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 928.608473] env[62066]: DEBUG oslo_concurrency.lockutils [None req-81c18bf6-1c8d-45a0-a66c-dcda0761f479 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Lock "8ff6fbb9-c90f-498d-9a85-d220a8c2f794" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 34.959s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 928.666058] env[62066]: DEBUG nova.scheduler.client.report [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 928.841823] env[62066]: DEBUG oslo_vmware.api [None req-e9681fdd-3744-4c1e-81e7-a2a0f833c385 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52296e63-3346-efee-bf9f-6e187950cef8, 'name': SearchDatastore_Task, 'duration_secs': 0.032062} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.842107] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e9681fdd-3744-4c1e-81e7-a2a0f833c385 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Releasing lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 928.842449] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-e9681fdd-3744-4c1e-81e7-a2a0f833c385 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk to [datastore2] a10a6dc0-7df7-40a1-888a-d1414dbeb1f7/a10a6dc0-7df7-40a1-888a-d1414dbeb1f7.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 928.842715] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-bf2132fb-e168-413e-8b7a-ce5c2bac2709 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.849182] env[62066]: DEBUG oslo_vmware.api [None req-e9681fdd-3744-4c1e-81e7-a2a0f833c385 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Waiting for the task: (returnval){ [ 928.849182] env[62066]: value = "task-1341165" [ 928.849182] env[62066]: _type = "Task" [ 928.849182] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.857501] env[62066]: DEBUG oslo_vmware.api [None req-e9681fdd-3744-4c1e-81e7-a2a0f833c385 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Task: {'id': task-1341165, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.942068] env[62066]: DEBUG nova.virt.hardware [None req-9f08ab51-33e2-4c97-838b-004b2127a87f tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-23T13:40:41Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-23T13:40:23Z,direct_url=,disk_format='vmdk',id=50ff584c-3b50-4395-af07-3e66769bc9f7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='eb52f7069a374c61ae946f052007c6d9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-23T13:40:24Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 928.942332] env[62066]: DEBUG nova.virt.hardware [None req-9f08ab51-33e2-4c97-838b-004b2127a87f tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 928.942533] env[62066]: DEBUG nova.virt.hardware [None req-9f08ab51-33e2-4c97-838b-004b2127a87f tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 928.942752] env[62066]: DEBUG nova.virt.hardware [None req-9f08ab51-33e2-4c97-838b-004b2127a87f tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 928.942904] env[62066]: DEBUG nova.virt.hardware [None req-9f08ab51-33e2-4c97-838b-004b2127a87f tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 928.943075] env[62066]: DEBUG nova.virt.hardware [None req-9f08ab51-33e2-4c97-838b-004b2127a87f tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 928.943291] env[62066]: DEBUG nova.virt.hardware [None req-9f08ab51-33e2-4c97-838b-004b2127a87f tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 928.943456] env[62066]: DEBUG nova.virt.hardware [None req-9f08ab51-33e2-4c97-838b-004b2127a87f tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 928.943624] env[62066]: DEBUG nova.virt.hardware [None req-9f08ab51-33e2-4c97-838b-004b2127a87f tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 928.943792] env[62066]: DEBUG nova.virt.hardware [None req-9f08ab51-33e2-4c97-838b-004b2127a87f tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 928.943968] env[62066]: DEBUG nova.virt.hardware [None req-9f08ab51-33e2-4c97-838b-004b2127a87f tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 928.948427] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26d0d94a-0004-44cf-bb0f-c1982e3a03e8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.957269] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff5c9f84-61eb-4fe2-a1bf-946f7f9b18b0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.971550] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-9f08ab51-33e2-4c97-838b-004b2127a87f tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] [instance: 15afaea0-65de-4f32-851c-365003e5498f] Instance VIF info [] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 928.977461] env[62066]: DEBUG oslo.service.loopingcall [None req-9f08ab51-33e2-4c97-838b-004b2127a87f tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 928.977587] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 15afaea0-65de-4f32-851c-365003e5498f] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 928.977794] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ca4323fb-8d6e-4f84-b05a-c646e88b5910 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.997100] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 928.997100] env[62066]: value = "task-1341166" [ 928.997100] env[62066]: _type = "Task" [ 928.997100] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.008278] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1341166, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.171624] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62066) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 929.172093] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.814s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 929.172494] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d267d329-25e8-414c-9540-a1c71897bff8 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.209s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 929.174944] env[62066]: INFO nova.compute.claims [None req-d267d329-25e8-414c-9540-a1c71897bff8 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 20987cb6-e12c-48c5-8fae-d990b1bb6b8a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 929.360230] env[62066]: DEBUG oslo_vmware.api [None req-e9681fdd-3744-4c1e-81e7-a2a0f833c385 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Task: {'id': task-1341165, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.489583} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.360230] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-e9681fdd-3744-4c1e-81e7-a2a0f833c385 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk to [datastore2] a10a6dc0-7df7-40a1-888a-d1414dbeb1f7/a10a6dc0-7df7-40a1-888a-d1414dbeb1f7.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 929.360230] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-e9681fdd-3744-4c1e-81e7-a2a0f833c385 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] [instance: a10a6dc0-7df7-40a1-888a-d1414dbeb1f7] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 929.360487] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8c0376cf-3022-4f36-8f6e-6c886e2be11b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.366636] env[62066]: DEBUG oslo_vmware.api [None req-e9681fdd-3744-4c1e-81e7-a2a0f833c385 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Waiting for the task: (returnval){ [ 929.366636] env[62066]: value = "task-1341167" [ 929.366636] env[62066]: _type = "Task" [ 929.366636] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.376348] env[62066]: DEBUG oslo_vmware.api [None req-e9681fdd-3744-4c1e-81e7-a2a0f833c385 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Task: {'id': task-1341167, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.465769] env[62066]: DEBUG oslo_concurrency.lockutils [None req-740993dc-fc35-4aa8-9a06-8f4c0e5f95e1 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Acquiring lock "9c989152-ec69-478e-a0b0-62983852576c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 929.465769] env[62066]: DEBUG oslo_concurrency.lockutils [None req-740993dc-fc35-4aa8-9a06-8f4c0e5f95e1 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Lock "9c989152-ec69-478e-a0b0-62983852576c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 929.465964] env[62066]: DEBUG oslo_concurrency.lockutils [None req-740993dc-fc35-4aa8-9a06-8f4c0e5f95e1 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Lock "9c989152-ec69-478e-a0b0-62983852576c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 929.507308] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1341166, 'name': CreateVM_Task, 'duration_secs': 0.279752} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.507520] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 15afaea0-65de-4f32-851c-365003e5498f] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 929.507982] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9f08ab51-33e2-4c97-838b-004b2127a87f tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 929.508214] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9f08ab51-33e2-4c97-838b-004b2127a87f tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Acquired lock "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 929.508540] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9f08ab51-33e2-4c97-838b-004b2127a87f tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 929.508784] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3e813df5-0d3e-486b-87ca-db4482d1f137 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.513413] env[62066]: DEBUG oslo_vmware.api [None req-9f08ab51-33e2-4c97-838b-004b2127a87f tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Waiting for the task: (returnval){ [ 929.513413] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]52d6e260-399d-74ff-f234-ff6adbf12c96" [ 929.513413] env[62066]: _type = "Task" [ 929.513413] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.520778] env[62066]: DEBUG oslo_vmware.api [None req-9f08ab51-33e2-4c97-838b-004b2127a87f tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52d6e260-399d-74ff-f234-ff6adbf12c96, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.876584] env[62066]: DEBUG oslo_vmware.api [None req-e9681fdd-3744-4c1e-81e7-a2a0f833c385 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Task: {'id': task-1341167, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062862} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.876907] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-e9681fdd-3744-4c1e-81e7-a2a0f833c385 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] [instance: a10a6dc0-7df7-40a1-888a-d1414dbeb1f7] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 929.877700] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4cb5656e-5478-48b3-a1ce-eb0bf5921f04 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.899585] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-e9681fdd-3744-4c1e-81e7-a2a0f833c385 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] [instance: a10a6dc0-7df7-40a1-888a-d1414dbeb1f7] Reconfiguring VM instance instance-0000005b to attach disk [datastore2] a10a6dc0-7df7-40a1-888a-d1414dbeb1f7/a10a6dc0-7df7-40a1-888a-d1414dbeb1f7.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 929.899852] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d942f6fa-5525-4354-b0fb-e92da04e2a41 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.919084] env[62066]: DEBUG oslo_vmware.api [None req-e9681fdd-3744-4c1e-81e7-a2a0f833c385 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Waiting for the task: (returnval){ [ 929.919084] env[62066]: value = "task-1341168" [ 929.919084] env[62066]: _type = "Task" [ 929.919084] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.926341] env[62066]: DEBUG oslo_vmware.api [None req-e9681fdd-3744-4c1e-81e7-a2a0f833c385 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Task: {'id': task-1341168, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.024212] env[62066]: DEBUG oslo_vmware.api [None req-9f08ab51-33e2-4c97-838b-004b2127a87f tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52d6e260-399d-74ff-f234-ff6adbf12c96, 'name': SearchDatastore_Task, 'duration_secs': 0.008468} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.024540] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9f08ab51-33e2-4c97-838b-004b2127a87f tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Releasing lock "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 930.024783] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-9f08ab51-33e2-4c97-838b-004b2127a87f tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] [instance: 15afaea0-65de-4f32-851c-365003e5498f] Processing image 50ff584c-3b50-4395-af07-3e66769bc9f7 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 930.025032] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9f08ab51-33e2-4c97-838b-004b2127a87f tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 930.025191] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9f08ab51-33e2-4c97-838b-004b2127a87f tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Acquired lock "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 930.025377] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-9f08ab51-33e2-4c97-838b-004b2127a87f tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 930.025648] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-674757a0-5e40-454f-b316-d094694bc0f9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.033114] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-9f08ab51-33e2-4c97-838b-004b2127a87f tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 930.033297] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-9f08ab51-33e2-4c97-838b-004b2127a87f tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 930.033990] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0e5092da-8396-437c-9acf-d3a73f1bc4d8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.038902] env[62066]: DEBUG oslo_vmware.api [None req-9f08ab51-33e2-4c97-838b-004b2127a87f tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Waiting for the task: (returnval){ [ 930.038902] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]52806054-2d95-1b2c-4821-a3fb33252a48" [ 930.038902] env[62066]: _type = "Task" [ 930.038902] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.046808] env[62066]: DEBUG oslo_vmware.api [None req-9f08ab51-33e2-4c97-838b-004b2127a87f tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52806054-2d95-1b2c-4821-a3fb33252a48, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.376711] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ef24cfc-f81e-455f-b6e1-03bb8d4d5531 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.384219] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-383f04bb-9038-4109-99a1-156c1e1137fc {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.416017] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41a08cc3-d187-49c2-8f90-66db6732e2ab {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.426428] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42efa13f-1b54-4c20-a96b-bed7bf2578b3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.433200] env[62066]: DEBUG oslo_vmware.api [None req-e9681fdd-3744-4c1e-81e7-a2a0f833c385 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Task: {'id': task-1341168, 'name': ReconfigVM_Task, 'duration_secs': 0.388522} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.433861] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-e9681fdd-3744-4c1e-81e7-a2a0f833c385 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] [instance: a10a6dc0-7df7-40a1-888a-d1414dbeb1f7] Reconfigured VM instance instance-0000005b to attach disk [datastore2] a10a6dc0-7df7-40a1-888a-d1414dbeb1f7/a10a6dc0-7df7-40a1-888a-d1414dbeb1f7.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 930.434479] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0cc03631-6009-4887-a0a4-ff8f4fcc3394 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.443865] env[62066]: DEBUG nova.compute.provider_tree [None req-d267d329-25e8-414c-9540-a1c71897bff8 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 930.446520] env[62066]: DEBUG oslo_vmware.api [None req-e9681fdd-3744-4c1e-81e7-a2a0f833c385 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Waiting for the task: (returnval){ [ 930.446520] env[62066]: value = "task-1341169" [ 930.446520] env[62066]: _type = "Task" [ 930.446520] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.454717] env[62066]: DEBUG oslo_vmware.api [None req-e9681fdd-3744-4c1e-81e7-a2a0f833c385 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Task: {'id': task-1341169, 'name': Rename_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.522954] env[62066]: DEBUG oslo_concurrency.lockutils [None req-740993dc-fc35-4aa8-9a06-8f4c0e5f95e1 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Acquiring lock "refresh_cache-9c989152-ec69-478e-a0b0-62983852576c" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 930.523182] env[62066]: DEBUG oslo_concurrency.lockutils [None req-740993dc-fc35-4aa8-9a06-8f4c0e5f95e1 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Acquired lock "refresh_cache-9c989152-ec69-478e-a0b0-62983852576c" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 930.523371] env[62066]: DEBUG nova.network.neutron [None req-740993dc-fc35-4aa8-9a06-8f4c0e5f95e1 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 9c989152-ec69-478e-a0b0-62983852576c] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 930.549130] env[62066]: DEBUG oslo_vmware.api [None req-9f08ab51-33e2-4c97-838b-004b2127a87f tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52806054-2d95-1b2c-4821-a3fb33252a48, 'name': SearchDatastore_Task, 'duration_secs': 0.010079} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.550019] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-492a905d-ed3b-42f8-8c05-64a1c510f552 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.556023] env[62066]: DEBUG oslo_vmware.api [None req-9f08ab51-33e2-4c97-838b-004b2127a87f tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Waiting for the task: (returnval){ [ 930.556023] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]52b09d4f-1cc4-5efc-b1f4-300e851c8cc1" [ 930.556023] env[62066]: _type = "Task" [ 930.556023] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.563863] env[62066]: DEBUG oslo_vmware.api [None req-9f08ab51-33e2-4c97-838b-004b2127a87f tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52b09d4f-1cc4-5efc-b1f4-300e851c8cc1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.948721] env[62066]: DEBUG nova.scheduler.client.report [None req-d267d329-25e8-414c-9540-a1c71897bff8 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 930.960819] env[62066]: DEBUG oslo_vmware.api [None req-e9681fdd-3744-4c1e-81e7-a2a0f833c385 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Task: {'id': task-1341169, 'name': Rename_Task, 'duration_secs': 0.146549} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.961058] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-e9681fdd-3744-4c1e-81e7-a2a0f833c385 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] [instance: a10a6dc0-7df7-40a1-888a-d1414dbeb1f7] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 930.961312] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-dc451d6b-1cd7-4204-8102-f12214524f73 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.967813] env[62066]: DEBUG oslo_vmware.api [None req-e9681fdd-3744-4c1e-81e7-a2a0f833c385 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Waiting for the task: (returnval){ [ 930.967813] env[62066]: value = "task-1341170" [ 930.967813] env[62066]: _type = "Task" [ 930.967813] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.975856] env[62066]: DEBUG oslo_vmware.api [None req-e9681fdd-3744-4c1e-81e7-a2a0f833c385 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Task: {'id': task-1341170, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.066971] env[62066]: DEBUG oslo_vmware.api [None req-9f08ab51-33e2-4c97-838b-004b2127a87f tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52b09d4f-1cc4-5efc-b1f4-300e851c8cc1, 'name': SearchDatastore_Task, 'duration_secs': 0.012531} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.067285] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9f08ab51-33e2-4c97-838b-004b2127a87f tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Releasing lock "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 931.067580] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-9f08ab51-33e2-4c97-838b-004b2127a87f tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk to [datastore1] 15afaea0-65de-4f32-851c-365003e5498f/15afaea0-65de-4f32-851c-365003e5498f.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 931.067847] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-46f783be-2ff8-4daa-814b-a7ac15a85de7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.074103] env[62066]: DEBUG oslo_vmware.api [None req-9f08ab51-33e2-4c97-838b-004b2127a87f tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Waiting for the task: (returnval){ [ 931.074103] env[62066]: value = "task-1341171" [ 931.074103] env[62066]: _type = "Task" [ 931.074103] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.083090] env[62066]: DEBUG oslo_vmware.api [None req-9f08ab51-33e2-4c97-838b-004b2127a87f tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Task: {'id': task-1341171, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.239800] env[62066]: DEBUG nova.network.neutron [None req-740993dc-fc35-4aa8-9a06-8f4c0e5f95e1 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 9c989152-ec69-478e-a0b0-62983852576c] Updating instance_info_cache with network_info: [{"id": "1c0d0075-7f22-441e-a283-21e9b5c5aecc", "address": "fa:16:3e:c5:c3:07", "network": {"id": "1c6c51e4-96f8-4175-b7f5-96b8a42fcb13", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-2006089809-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d609babc987a42e2a8ddb4bfb9c3b103", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a965790c-2d2f-4c2a-9ee7-745f4d53039b", "external-id": "nsx-vlan-transportzone-708", "segmentation_id": 708, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1c0d0075-7f", "ovs_interfaceid": "1c0d0075-7f22-441e-a283-21e9b5c5aecc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 931.456111] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d267d329-25e8-414c-9540-a1c71897bff8 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.283s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 931.456654] env[62066]: DEBUG nova.compute.manager [None req-d267d329-25e8-414c-9540-a1c71897bff8 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 20987cb6-e12c-48c5-8fae-d990b1bb6b8a] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 931.460240] env[62066]: DEBUG oslo_concurrency.lockutils [None req-54702e84-4f3e-4046-b83c-40cd5c6f41f7 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.176s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 931.461632] env[62066]: INFO nova.compute.claims [None req-54702e84-4f3e-4046-b83c-40cd5c6f41f7 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 651e5142-7ab0-4040-97b4-4599bd4e419a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 931.477318] env[62066]: DEBUG oslo_vmware.api [None req-e9681fdd-3744-4c1e-81e7-a2a0f833c385 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Task: {'id': task-1341170, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.584176] env[62066]: DEBUG oslo_vmware.api [None req-9f08ab51-33e2-4c97-838b-004b2127a87f tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Task: {'id': task-1341171, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.743267] env[62066]: DEBUG oslo_concurrency.lockutils [None req-740993dc-fc35-4aa8-9a06-8f4c0e5f95e1 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Releasing lock "refresh_cache-9c989152-ec69-478e-a0b0-62983852576c" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 931.965936] env[62066]: DEBUG nova.compute.utils [None req-d267d329-25e8-414c-9540-a1c71897bff8 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 931.969170] env[62066]: DEBUG nova.compute.manager [None req-d267d329-25e8-414c-9540-a1c71897bff8 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 20987cb6-e12c-48c5-8fae-d990b1bb6b8a] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 931.969361] env[62066]: DEBUG nova.network.neutron [None req-d267d329-25e8-414c-9540-a1c71897bff8 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 20987cb6-e12c-48c5-8fae-d990b1bb6b8a] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 931.981630] env[62066]: DEBUG oslo_vmware.api [None req-e9681fdd-3744-4c1e-81e7-a2a0f833c385 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Task: {'id': task-1341170, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.042565] env[62066]: DEBUG nova.policy [None req-d267d329-25e8-414c-9540-a1c71897bff8 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8910a229218b4ec5ad72c893badfc598', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6497ab02f327476d8ff81c2ecc0371e5', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 932.086162] env[62066]: DEBUG oslo_vmware.api [None req-9f08ab51-33e2-4c97-838b-004b2127a87f tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Task: {'id': task-1341171, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.904563} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.086499] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-9f08ab51-33e2-4c97-838b-004b2127a87f tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk to [datastore1] 15afaea0-65de-4f32-851c-365003e5498f/15afaea0-65de-4f32-851c-365003e5498f.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 932.086725] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-9f08ab51-33e2-4c97-838b-004b2127a87f tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] [instance: 15afaea0-65de-4f32-851c-365003e5498f] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 932.086990] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e187fdac-9c56-4ee6-b9e2-0f52e48f82ae {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.093405] env[62066]: DEBUG oslo_vmware.api [None req-9f08ab51-33e2-4c97-838b-004b2127a87f tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Waiting for the task: (returnval){ [ 932.093405] env[62066]: value = "task-1341172" [ 932.093405] env[62066]: _type = "Task" [ 932.093405] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.102635] env[62066]: DEBUG oslo_vmware.api [None req-9f08ab51-33e2-4c97-838b-004b2127a87f tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Task: {'id': task-1341172, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.267074] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11925b00-4df6-4cab-8cd6-13655d067d3b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.285742] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-293fa109-4d5f-40e3-9697-4d4d4cdc12e5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.294262] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-740993dc-fc35-4aa8-9a06-8f4c0e5f95e1 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 9c989152-ec69-478e-a0b0-62983852576c] Updating instance '9c989152-ec69-478e-a0b0-62983852576c' progress to 83 {{(pid=62066) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 932.305994] env[62066]: DEBUG nova.network.neutron [None req-d267d329-25e8-414c-9540-a1c71897bff8 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 20987cb6-e12c-48c5-8fae-d990b1bb6b8a] Successfully created port: f24a3aad-eb9f-4bd9-aa9d-e7cc51421b37 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 932.470390] env[62066]: DEBUG nova.compute.manager [None req-d267d329-25e8-414c-9540-a1c71897bff8 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 20987cb6-e12c-48c5-8fae-d990b1bb6b8a] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 932.484787] env[62066]: DEBUG oslo_vmware.api [None req-e9681fdd-3744-4c1e-81e7-a2a0f833c385 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Task: {'id': task-1341170, 'name': PowerOnVM_Task, 'duration_secs': 1.183334} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.485084] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-e9681fdd-3744-4c1e-81e7-a2a0f833c385 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] [instance: a10a6dc0-7df7-40a1-888a-d1414dbeb1f7] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 932.485295] env[62066]: INFO nova.compute.manager [None req-e9681fdd-3744-4c1e-81e7-a2a0f833c385 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] [instance: a10a6dc0-7df7-40a1-888a-d1414dbeb1f7] Took 8.80 seconds to spawn the instance on the hypervisor. [ 932.485479] env[62066]: DEBUG nova.compute.manager [None req-e9681fdd-3744-4c1e-81e7-a2a0f833c385 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] [instance: a10a6dc0-7df7-40a1-888a-d1414dbeb1f7] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 932.486881] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e208a078-498c-4617-9334-f959c1696f46 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.604891] env[62066]: DEBUG oslo_vmware.api [None req-9f08ab51-33e2-4c97-838b-004b2127a87f tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Task: {'id': task-1341172, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071081} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.605189] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-9f08ab51-33e2-4c97-838b-004b2127a87f tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] [instance: 15afaea0-65de-4f32-851c-365003e5498f] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 932.605960] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-238999e3-ef31-4127-a087-21e837840cf5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.627032] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-9f08ab51-33e2-4c97-838b-004b2127a87f tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] [instance: 15afaea0-65de-4f32-851c-365003e5498f] Reconfiguring VM instance instance-0000005a to attach disk [datastore1] 15afaea0-65de-4f32-851c-365003e5498f/15afaea0-65de-4f32-851c-365003e5498f.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 932.629478] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3f5ee48d-493c-4733-b037-029db81aefee {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.648451] env[62066]: DEBUG oslo_vmware.api [None req-9f08ab51-33e2-4c97-838b-004b2127a87f tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Waiting for the task: (returnval){ [ 932.648451] env[62066]: value = "task-1341173" [ 932.648451] env[62066]: _type = "Task" [ 932.648451] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.657371] env[62066]: DEBUG oslo_vmware.api [None req-9f08ab51-33e2-4c97-838b-004b2127a87f tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Task: {'id': task-1341173, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.765552] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-503ff485-7f8c-4868-865a-2c7ad1884166 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.772828] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1047d1ce-c036-40da-9a4c-34a6a8c98758 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.803325] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-740993dc-fc35-4aa8-9a06-8f4c0e5f95e1 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 9c989152-ec69-478e-a0b0-62983852576c] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 932.803773] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8bb9056f-e726-4af6-aa1b-43126459550c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.805864] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8660e7f-d282-4fed-a317-30418f0437dd {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.814655] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afa64a11-28fa-4fd8-baa2-cd5f9443cad1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.819775] env[62066]: DEBUG oslo_vmware.api [None req-740993dc-fc35-4aa8-9a06-8f4c0e5f95e1 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Waiting for the task: (returnval){ [ 932.819775] env[62066]: value = "task-1341174" [ 932.819775] env[62066]: _type = "Task" [ 932.819775] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.830650] env[62066]: DEBUG nova.compute.provider_tree [None req-54702e84-4f3e-4046-b83c-40cd5c6f41f7 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 932.837066] env[62066]: DEBUG oslo_vmware.api [None req-740993dc-fc35-4aa8-9a06-8f4c0e5f95e1 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1341174, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.012232] env[62066]: INFO nova.compute.manager [None req-e9681fdd-3744-4c1e-81e7-a2a0f833c385 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] [instance: a10a6dc0-7df7-40a1-888a-d1414dbeb1f7] Took 20.93 seconds to build instance. [ 933.159544] env[62066]: DEBUG oslo_vmware.api [None req-9f08ab51-33e2-4c97-838b-004b2127a87f tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Task: {'id': task-1341173, 'name': ReconfigVM_Task, 'duration_secs': 0.319984} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.159827] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-9f08ab51-33e2-4c97-838b-004b2127a87f tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] [instance: 15afaea0-65de-4f32-851c-365003e5498f] Reconfigured VM instance instance-0000005a to attach disk [datastore1] 15afaea0-65de-4f32-851c-365003e5498f/15afaea0-65de-4f32-851c-365003e5498f.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 933.160803] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-948031f9-1d99-466d-ac88-019cc87f6983 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.167256] env[62066]: DEBUG oslo_vmware.api [None req-9f08ab51-33e2-4c97-838b-004b2127a87f tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Waiting for the task: (returnval){ [ 933.167256] env[62066]: value = "task-1341175" [ 933.167256] env[62066]: _type = "Task" [ 933.167256] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.174846] env[62066]: DEBUG oslo_vmware.api [None req-9f08ab51-33e2-4c97-838b-004b2127a87f tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Task: {'id': task-1341175, 'name': Rename_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.328459] env[62066]: DEBUG oslo_vmware.api [None req-740993dc-fc35-4aa8-9a06-8f4c0e5f95e1 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1341174, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.333732] env[62066]: DEBUG nova.scheduler.client.report [None req-54702e84-4f3e-4046-b83c-40cd5c6f41f7 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 933.483500] env[62066]: DEBUG nova.compute.manager [None req-d267d329-25e8-414c-9540-a1c71897bff8 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 20987cb6-e12c-48c5-8fae-d990b1bb6b8a] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 933.507725] env[62066]: DEBUG nova.virt.hardware [None req-d267d329-25e8-414c-9540-a1c71897bff8 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-23T13:40:41Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-23T13:40:23Z,direct_url=,disk_format='vmdk',id=50ff584c-3b50-4395-af07-3e66769bc9f7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='eb52f7069a374c61ae946f052007c6d9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-23T13:40:24Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 933.507964] env[62066]: DEBUG nova.virt.hardware [None req-d267d329-25e8-414c-9540-a1c71897bff8 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 933.508174] env[62066]: DEBUG nova.virt.hardware [None req-d267d329-25e8-414c-9540-a1c71897bff8 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 933.508371] env[62066]: DEBUG nova.virt.hardware [None req-d267d329-25e8-414c-9540-a1c71897bff8 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 933.508522] env[62066]: DEBUG nova.virt.hardware [None req-d267d329-25e8-414c-9540-a1c71897bff8 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 933.508673] env[62066]: DEBUG nova.virt.hardware [None req-d267d329-25e8-414c-9540-a1c71897bff8 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 933.508883] env[62066]: DEBUG nova.virt.hardware [None req-d267d329-25e8-414c-9540-a1c71897bff8 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 933.509057] env[62066]: DEBUG nova.virt.hardware [None req-d267d329-25e8-414c-9540-a1c71897bff8 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 933.509236] env[62066]: DEBUG nova.virt.hardware [None req-d267d329-25e8-414c-9540-a1c71897bff8 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 933.509408] env[62066]: DEBUG nova.virt.hardware [None req-d267d329-25e8-414c-9540-a1c71897bff8 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 933.509584] env[62066]: DEBUG nova.virt.hardware [None req-d267d329-25e8-414c-9540-a1c71897bff8 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 933.510501] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc44c546-d9e5-45ef-8c9c-8e1f6209ad6f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.519678] env[62066]: DEBUG oslo_concurrency.lockutils [None req-e9681fdd-3744-4c1e-81e7-a2a0f833c385 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Lock "a10a6dc0-7df7-40a1-888a-d1414dbeb1f7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.446s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 933.520995] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6012715a-18e2-4f57-a5e1-09030d5be337 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.571364] env[62066]: DEBUG nova.compute.manager [req-58eda0e7-57b7-4c18-956a-449b71e34d1b req-97073218-2226-4274-a326-bd8bc76bc688 service nova] [instance: a10a6dc0-7df7-40a1-888a-d1414dbeb1f7] Received event network-changed-26148c05-d41e-4564-a159-de3f90c20550 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 933.571582] env[62066]: DEBUG nova.compute.manager [req-58eda0e7-57b7-4c18-956a-449b71e34d1b req-97073218-2226-4274-a326-bd8bc76bc688 service nova] [instance: a10a6dc0-7df7-40a1-888a-d1414dbeb1f7] Refreshing instance network info cache due to event network-changed-26148c05-d41e-4564-a159-de3f90c20550. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 933.571801] env[62066]: DEBUG oslo_concurrency.lockutils [req-58eda0e7-57b7-4c18-956a-449b71e34d1b req-97073218-2226-4274-a326-bd8bc76bc688 service nova] Acquiring lock "refresh_cache-a10a6dc0-7df7-40a1-888a-d1414dbeb1f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 933.571947] env[62066]: DEBUG oslo_concurrency.lockutils [req-58eda0e7-57b7-4c18-956a-449b71e34d1b req-97073218-2226-4274-a326-bd8bc76bc688 service nova] Acquired lock "refresh_cache-a10a6dc0-7df7-40a1-888a-d1414dbeb1f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 933.572367] env[62066]: DEBUG nova.network.neutron [req-58eda0e7-57b7-4c18-956a-449b71e34d1b req-97073218-2226-4274-a326-bd8bc76bc688 service nova] [instance: a10a6dc0-7df7-40a1-888a-d1414dbeb1f7] Refreshing network info cache for port 26148c05-d41e-4564-a159-de3f90c20550 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 933.677468] env[62066]: DEBUG oslo_vmware.api [None req-9f08ab51-33e2-4c97-838b-004b2127a87f tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Task: {'id': task-1341175, 'name': Rename_Task, 'duration_secs': 0.141525} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.677740] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-9f08ab51-33e2-4c97-838b-004b2127a87f tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] [instance: 15afaea0-65de-4f32-851c-365003e5498f] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 933.677958] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3f508d75-2f76-4d27-aba5-3fa6bc826cda {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.684534] env[62066]: DEBUG oslo_vmware.api [None req-9f08ab51-33e2-4c97-838b-004b2127a87f tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Waiting for the task: (returnval){ [ 933.684534] env[62066]: value = "task-1341176" [ 933.684534] env[62066]: _type = "Task" [ 933.684534] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.694352] env[62066]: DEBUG oslo_vmware.api [None req-9f08ab51-33e2-4c97-838b-004b2127a87f tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Task: {'id': task-1341176, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.812385] env[62066]: DEBUG nova.network.neutron [None req-d267d329-25e8-414c-9540-a1c71897bff8 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 20987cb6-e12c-48c5-8fae-d990b1bb6b8a] Successfully updated port: f24a3aad-eb9f-4bd9-aa9d-e7cc51421b37 {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 933.832695] env[62066]: DEBUG oslo_vmware.api [None req-740993dc-fc35-4aa8-9a06-8f4c0e5f95e1 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1341174, 'name': PowerOnVM_Task, 'duration_secs': 0.573055} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.832695] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-740993dc-fc35-4aa8-9a06-8f4c0e5f95e1 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 9c989152-ec69-478e-a0b0-62983852576c] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 933.832882] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-740993dc-fc35-4aa8-9a06-8f4c0e5f95e1 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 9c989152-ec69-478e-a0b0-62983852576c] Updating instance '9c989152-ec69-478e-a0b0-62983852576c' progress to 100 {{(pid=62066) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 933.840755] env[62066]: DEBUG oslo_concurrency.lockutils [None req-54702e84-4f3e-4046-b83c-40cd5c6f41f7 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.381s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 933.841280] env[62066]: DEBUG nova.compute.manager [None req-54702e84-4f3e-4046-b83c-40cd5c6f41f7 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 651e5142-7ab0-4040-97b4-4599bd4e419a] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 933.843648] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0e878f04-940e-4d47-ad0a-15c5be73ba89 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 12.122s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 933.843846] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0e878f04-940e-4d47-ad0a-15c5be73ba89 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 933.849088] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3bd520fa-2cd0-41e8-afc3-e67149da4105 tempest-ServerTagsTestJSON-2015212290 tempest-ServerTagsTestJSON-2015212290-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 11.117s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 933.849305] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3bd520fa-2cd0-41e8-afc3-e67149da4105 tempest-ServerTagsTestJSON-2015212290 tempest-ServerTagsTestJSON-2015212290-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 933.855146] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0c3b1401-b067-4cd8-8015-7ffe20d4d016 tempest-ServersTestFqdnHostnames-1367628746 tempest-ServersTestFqdnHostnames-1367628746-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 8.972s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 933.855415] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0c3b1401-b067-4cd8-8015-7ffe20d4d016 tempest-ServersTestFqdnHostnames-1367628746 tempest-ServersTestFqdnHostnames-1367628746-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.004s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 933.857555] env[62066]: DEBUG oslo_concurrency.lockutils [None req-bb369cbc-f881-4f95-b841-517845c5cb59 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.396s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 933.859520] env[62066]: INFO nova.compute.claims [None req-bb369cbc-f881-4f95-b841-517845c5cb59 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: 5693eea2-1cac-45d5-aa27-58526b65136e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 933.888988] env[62066]: INFO nova.scheduler.client.report [None req-0e878f04-940e-4d47-ad0a-15c5be73ba89 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Deleted allocations for instance b08f27b1-429e-4d05-9a95-f45fd0cb2904 [ 933.891027] env[62066]: INFO nova.scheduler.client.report [None req-3bd520fa-2cd0-41e8-afc3-e67149da4105 tempest-ServerTagsTestJSON-2015212290 tempest-ServerTagsTestJSON-2015212290-project-member] Deleted allocations for instance f4a84152-f222-46a2-9a35-8a0a0078fc12 [ 933.902815] env[62066]: INFO nova.scheduler.client.report [None req-0c3b1401-b067-4cd8-8015-7ffe20d4d016 tempest-ServersTestFqdnHostnames-1367628746 tempest-ServersTestFqdnHostnames-1367628746-project-member] Deleted allocations for instance afaca059-fc4e-4de2-8fa0-d2f226ea7051 [ 934.195618] env[62066]: DEBUG oslo_vmware.api [None req-9f08ab51-33e2-4c97-838b-004b2127a87f tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Task: {'id': task-1341176, 'name': PowerOnVM_Task, 'duration_secs': 0.426762} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.198310] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-9f08ab51-33e2-4c97-838b-004b2127a87f tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] [instance: 15afaea0-65de-4f32-851c-365003e5498f] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 934.198540] env[62066]: DEBUG nova.compute.manager [None req-9f08ab51-33e2-4c97-838b-004b2127a87f tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] [instance: 15afaea0-65de-4f32-851c-365003e5498f] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 934.199382] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a789cd7-77d2-4619-8da9-22b5c371f72b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.317959] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d267d329-25e8-414c-9540-a1c71897bff8 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Acquiring lock "refresh_cache-20987cb6-e12c-48c5-8fae-d990b1bb6b8a" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 934.318196] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d267d329-25e8-414c-9540-a1c71897bff8 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Acquired lock "refresh_cache-20987cb6-e12c-48c5-8fae-d990b1bb6b8a" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 934.318365] env[62066]: DEBUG nova.network.neutron [None req-d267d329-25e8-414c-9540-a1c71897bff8 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 20987cb6-e12c-48c5-8fae-d990b1bb6b8a] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 934.356177] env[62066]: DEBUG nova.network.neutron [req-58eda0e7-57b7-4c18-956a-449b71e34d1b req-97073218-2226-4274-a326-bd8bc76bc688 service nova] [instance: a10a6dc0-7df7-40a1-888a-d1414dbeb1f7] Updated VIF entry in instance network info cache for port 26148c05-d41e-4564-a159-de3f90c20550. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 934.356554] env[62066]: DEBUG nova.network.neutron [req-58eda0e7-57b7-4c18-956a-449b71e34d1b req-97073218-2226-4274-a326-bd8bc76bc688 service nova] [instance: a10a6dc0-7df7-40a1-888a-d1414dbeb1f7] Updating instance_info_cache with network_info: [{"id": "26148c05-d41e-4564-a159-de3f90c20550", "address": "fa:16:3e:c8:14:c9", "network": {"id": "4114ddfe-818f-47a5-bb6f-44a4c22517f6", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-676131672-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.160", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dd30584efc254610b3939d1ba3806693", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a485857d-7086-4dcf-9d65-d0dcd177fcb0", "external-id": "nsx-vlan-transportzone-232", "segmentation_id": 232, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap26148c05-d4", "ovs_interfaceid": "26148c05-d41e-4564-a159-de3f90c20550", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 934.364456] env[62066]: DEBUG nova.compute.utils [None req-54702e84-4f3e-4046-b83c-40cd5c6f41f7 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 934.368165] env[62066]: DEBUG nova.compute.manager [None req-54702e84-4f3e-4046-b83c-40cd5c6f41f7 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 651e5142-7ab0-4040-97b4-4599bd4e419a] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 934.368332] env[62066]: DEBUG nova.network.neutron [None req-54702e84-4f3e-4046-b83c-40cd5c6f41f7 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 651e5142-7ab0-4040-97b4-4599bd4e419a] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 934.402336] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3bd520fa-2cd0-41e8-afc3-e67149da4105 tempest-ServerTagsTestJSON-2015212290 tempest-ServerTagsTestJSON-2015212290-project-member] Lock "f4a84152-f222-46a2-9a35-8a0a0078fc12" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 14.968s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 934.403741] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0e878f04-940e-4d47-ad0a-15c5be73ba89 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Lock "b08f27b1-429e-4d05-9a95-f45fd0cb2904" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 15.699s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 934.406344] env[62066]: DEBUG nova.policy [None req-54702e84-4f3e-4046-b83c-40cd5c6f41f7 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '60e2bc61839b4299912c53137bc23fbc', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '75f0352c852947369474c9e05766c584', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 934.410077] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0c3b1401-b067-4cd8-8015-7ffe20d4d016 tempest-ServersTestFqdnHostnames-1367628746 tempest-ServersTestFqdnHostnames-1367628746-project-member] Lock "afaca059-fc4e-4de2-8fa0-d2f226ea7051" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 12.711s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 934.721610] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9f08ab51-33e2-4c97-838b-004b2127a87f tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 934.727366] env[62066]: DEBUG nova.network.neutron [None req-54702e84-4f3e-4046-b83c-40cd5c6f41f7 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 651e5142-7ab0-4040-97b4-4599bd4e419a] Successfully created port: 66054023-ce27-4e2d-92f5-18f1dde734d0 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 934.861193] env[62066]: DEBUG oslo_concurrency.lockutils [req-58eda0e7-57b7-4c18-956a-449b71e34d1b req-97073218-2226-4274-a326-bd8bc76bc688 service nova] Releasing lock "refresh_cache-a10a6dc0-7df7-40a1-888a-d1414dbeb1f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 934.876346] env[62066]: DEBUG nova.compute.manager [None req-54702e84-4f3e-4046-b83c-40cd5c6f41f7 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 651e5142-7ab0-4040-97b4-4599bd4e419a] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 934.897791] env[62066]: DEBUG nova.network.neutron [None req-d267d329-25e8-414c-9540-a1c71897bff8 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 20987cb6-e12c-48c5-8fae-d990b1bb6b8a] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 935.071277] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72e426c5-25da-41ab-9850-03c0dd3190a7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.082071] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d02baf2f-98d4-49b9-9f46-e1a2d66d179e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.124402] env[62066]: DEBUG oslo_concurrency.lockutils [None req-594ce8c1-1ab1-41b6-bcf3-6a5287350f9c tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Acquiring lock "15afaea0-65de-4f32-851c-365003e5498f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 935.124757] env[62066]: DEBUG oslo_concurrency.lockutils [None req-594ce8c1-1ab1-41b6-bcf3-6a5287350f9c tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Lock "15afaea0-65de-4f32-851c-365003e5498f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 935.124991] env[62066]: DEBUG oslo_concurrency.lockutils [None req-594ce8c1-1ab1-41b6-bcf3-6a5287350f9c tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Acquiring lock "15afaea0-65de-4f32-851c-365003e5498f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 935.125202] env[62066]: DEBUG oslo_concurrency.lockutils [None req-594ce8c1-1ab1-41b6-bcf3-6a5287350f9c tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Lock "15afaea0-65de-4f32-851c-365003e5498f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 935.125382] env[62066]: DEBUG oslo_concurrency.lockutils [None req-594ce8c1-1ab1-41b6-bcf3-6a5287350f9c tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Lock "15afaea0-65de-4f32-851c-365003e5498f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 935.127563] env[62066]: INFO nova.compute.manager [None req-594ce8c1-1ab1-41b6-bcf3-6a5287350f9c tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] [instance: 15afaea0-65de-4f32-851c-365003e5498f] Terminating instance [ 935.129562] env[62066]: DEBUG oslo_concurrency.lockutils [None req-594ce8c1-1ab1-41b6-bcf3-6a5287350f9c tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Acquiring lock "refresh_cache-15afaea0-65de-4f32-851c-365003e5498f" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 935.129727] env[62066]: DEBUG oslo_concurrency.lockutils [None req-594ce8c1-1ab1-41b6-bcf3-6a5287350f9c tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Acquired lock "refresh_cache-15afaea0-65de-4f32-851c-365003e5498f" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 935.130015] env[62066]: DEBUG nova.network.neutron [None req-594ce8c1-1ab1-41b6-bcf3-6a5287350f9c tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] [instance: 15afaea0-65de-4f32-851c-365003e5498f] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 935.135898] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97a16fab-813f-44ab-b5d1-f911b6e2c5e6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.145842] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91580d6d-2269-4677-9417-6e038dfac632 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.161498] env[62066]: DEBUG nova.compute.provider_tree [None req-bb369cbc-f881-4f95-b841-517845c5cb59 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 935.194817] env[62066]: DEBUG nova.network.neutron [None req-d267d329-25e8-414c-9540-a1c71897bff8 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 20987cb6-e12c-48c5-8fae-d990b1bb6b8a] Updating instance_info_cache with network_info: [{"id": "f24a3aad-eb9f-4bd9-aa9d-e7cc51421b37", "address": "fa:16:3e:62:1f:dd", "network": {"id": "9f3396ce-ee48-4eaf-8cc7-3e1711f37537", "bridge": "br-int", "label": "tempest-ServersTestJSON-2129775093-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6497ab02f327476d8ff81c2ecc0371e5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7ab8d568-adb0-4f3b-b6cc-68413e6546ae", "external-id": "nsx-vlan-transportzone-86", "segmentation_id": 86, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf24a3aad-eb", "ovs_interfaceid": "f24a3aad-eb9f-4bd9-aa9d-e7cc51421b37", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 935.610667] env[62066]: DEBUG nova.compute.manager [req-731f5a42-b150-47e9-ad1c-ed5fdd2b1b97 req-3ed0ad5a-e4e4-4391-a0f1-2f69b24006e9 service nova] [instance: 20987cb6-e12c-48c5-8fae-d990b1bb6b8a] Received event network-vif-plugged-f24a3aad-eb9f-4bd9-aa9d-e7cc51421b37 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 935.611062] env[62066]: DEBUG oslo_concurrency.lockutils [req-731f5a42-b150-47e9-ad1c-ed5fdd2b1b97 req-3ed0ad5a-e4e4-4391-a0f1-2f69b24006e9 service nova] Acquiring lock "20987cb6-e12c-48c5-8fae-d990b1bb6b8a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 935.611062] env[62066]: DEBUG oslo_concurrency.lockutils [req-731f5a42-b150-47e9-ad1c-ed5fdd2b1b97 req-3ed0ad5a-e4e4-4391-a0f1-2f69b24006e9 service nova] Lock "20987cb6-e12c-48c5-8fae-d990b1bb6b8a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 935.611398] env[62066]: DEBUG oslo_concurrency.lockutils [req-731f5a42-b150-47e9-ad1c-ed5fdd2b1b97 req-3ed0ad5a-e4e4-4391-a0f1-2f69b24006e9 service nova] Lock "20987cb6-e12c-48c5-8fae-d990b1bb6b8a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 935.611398] env[62066]: DEBUG nova.compute.manager [req-731f5a42-b150-47e9-ad1c-ed5fdd2b1b97 req-3ed0ad5a-e4e4-4391-a0f1-2f69b24006e9 service nova] [instance: 20987cb6-e12c-48c5-8fae-d990b1bb6b8a] No waiting events found dispatching network-vif-plugged-f24a3aad-eb9f-4bd9-aa9d-e7cc51421b37 {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 935.611557] env[62066]: WARNING nova.compute.manager [req-731f5a42-b150-47e9-ad1c-ed5fdd2b1b97 req-3ed0ad5a-e4e4-4391-a0f1-2f69b24006e9 service nova] [instance: 20987cb6-e12c-48c5-8fae-d990b1bb6b8a] Received unexpected event network-vif-plugged-f24a3aad-eb9f-4bd9-aa9d-e7cc51421b37 for instance with vm_state building and task_state spawning. [ 935.612059] env[62066]: DEBUG nova.compute.manager [req-731f5a42-b150-47e9-ad1c-ed5fdd2b1b97 req-3ed0ad5a-e4e4-4391-a0f1-2f69b24006e9 service nova] [instance: 20987cb6-e12c-48c5-8fae-d990b1bb6b8a] Received event network-changed-f24a3aad-eb9f-4bd9-aa9d-e7cc51421b37 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 935.612059] env[62066]: DEBUG nova.compute.manager [req-731f5a42-b150-47e9-ad1c-ed5fdd2b1b97 req-3ed0ad5a-e4e4-4391-a0f1-2f69b24006e9 service nova] [instance: 20987cb6-e12c-48c5-8fae-d990b1bb6b8a] Refreshing instance network info cache due to event network-changed-f24a3aad-eb9f-4bd9-aa9d-e7cc51421b37. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 935.612059] env[62066]: DEBUG oslo_concurrency.lockutils [req-731f5a42-b150-47e9-ad1c-ed5fdd2b1b97 req-3ed0ad5a-e4e4-4391-a0f1-2f69b24006e9 service nova] Acquiring lock "refresh_cache-20987cb6-e12c-48c5-8fae-d990b1bb6b8a" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 935.656382] env[62066]: DEBUG nova.network.neutron [None req-594ce8c1-1ab1-41b6-bcf3-6a5287350f9c tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] [instance: 15afaea0-65de-4f32-851c-365003e5498f] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 935.665294] env[62066]: DEBUG nova.scheduler.client.report [None req-bb369cbc-f881-4f95-b841-517845c5cb59 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 935.697644] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d267d329-25e8-414c-9540-a1c71897bff8 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Releasing lock "refresh_cache-20987cb6-e12c-48c5-8fae-d990b1bb6b8a" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 935.697949] env[62066]: DEBUG nova.compute.manager [None req-d267d329-25e8-414c-9540-a1c71897bff8 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 20987cb6-e12c-48c5-8fae-d990b1bb6b8a] Instance network_info: |[{"id": "f24a3aad-eb9f-4bd9-aa9d-e7cc51421b37", "address": "fa:16:3e:62:1f:dd", "network": {"id": "9f3396ce-ee48-4eaf-8cc7-3e1711f37537", "bridge": "br-int", "label": "tempest-ServersTestJSON-2129775093-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6497ab02f327476d8ff81c2ecc0371e5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7ab8d568-adb0-4f3b-b6cc-68413e6546ae", "external-id": "nsx-vlan-transportzone-86", "segmentation_id": 86, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf24a3aad-eb", "ovs_interfaceid": "f24a3aad-eb9f-4bd9-aa9d-e7cc51421b37", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 935.698369] env[62066]: DEBUG oslo_concurrency.lockutils [req-731f5a42-b150-47e9-ad1c-ed5fdd2b1b97 req-3ed0ad5a-e4e4-4391-a0f1-2f69b24006e9 service nova] Acquired lock "refresh_cache-20987cb6-e12c-48c5-8fae-d990b1bb6b8a" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 935.698600] env[62066]: DEBUG nova.network.neutron [req-731f5a42-b150-47e9-ad1c-ed5fdd2b1b97 req-3ed0ad5a-e4e4-4391-a0f1-2f69b24006e9 service nova] [instance: 20987cb6-e12c-48c5-8fae-d990b1bb6b8a] Refreshing network info cache for port f24a3aad-eb9f-4bd9-aa9d-e7cc51421b37 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 935.699953] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-d267d329-25e8-414c-9540-a1c71897bff8 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 20987cb6-e12c-48c5-8fae-d990b1bb6b8a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:62:1f:dd', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7ab8d568-adb0-4f3b-b6cc-68413e6546ae', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f24a3aad-eb9f-4bd9-aa9d-e7cc51421b37', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 935.707805] env[62066]: DEBUG oslo.service.loopingcall [None req-d267d329-25e8-414c-9540-a1c71897bff8 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 935.708610] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 20987cb6-e12c-48c5-8fae-d990b1bb6b8a] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 935.708903] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-779f6b34-3869-4102-9fda-53ab51321db7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.730249] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 935.730249] env[62066]: value = "task-1341177" [ 935.730249] env[62066]: _type = "Task" [ 935.730249] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.739494] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1341177, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.788751] env[62066]: DEBUG nova.network.neutron [None req-594ce8c1-1ab1-41b6-bcf3-6a5287350f9c tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] [instance: 15afaea0-65de-4f32-851c-365003e5498f] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 935.885586] env[62066]: DEBUG nova.compute.manager [None req-54702e84-4f3e-4046-b83c-40cd5c6f41f7 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 651e5142-7ab0-4040-97b4-4599bd4e419a] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 935.930860] env[62066]: DEBUG nova.virt.hardware [None req-54702e84-4f3e-4046-b83c-40cd5c6f41f7 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-23T13:40:41Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-23T13:40:23Z,direct_url=,disk_format='vmdk',id=50ff584c-3b50-4395-af07-3e66769bc9f7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='eb52f7069a374c61ae946f052007c6d9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-23T13:40:24Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 935.930860] env[62066]: DEBUG nova.virt.hardware [None req-54702e84-4f3e-4046-b83c-40cd5c6f41f7 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 935.930860] env[62066]: DEBUG nova.virt.hardware [None req-54702e84-4f3e-4046-b83c-40cd5c6f41f7 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 935.930860] env[62066]: DEBUG nova.virt.hardware [None req-54702e84-4f3e-4046-b83c-40cd5c6f41f7 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 935.930860] env[62066]: DEBUG nova.virt.hardware [None req-54702e84-4f3e-4046-b83c-40cd5c6f41f7 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 935.930860] env[62066]: DEBUG nova.virt.hardware [None req-54702e84-4f3e-4046-b83c-40cd5c6f41f7 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 935.930860] env[62066]: DEBUG nova.virt.hardware [None req-54702e84-4f3e-4046-b83c-40cd5c6f41f7 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 935.930860] env[62066]: DEBUG nova.virt.hardware [None req-54702e84-4f3e-4046-b83c-40cd5c6f41f7 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 935.930860] env[62066]: DEBUG nova.virt.hardware [None req-54702e84-4f3e-4046-b83c-40cd5c6f41f7 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 935.930860] env[62066]: DEBUG nova.virt.hardware [None req-54702e84-4f3e-4046-b83c-40cd5c6f41f7 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 935.930860] env[62066]: DEBUG nova.virt.hardware [None req-54702e84-4f3e-4046-b83c-40cd5c6f41f7 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 935.931856] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bb109c7-a80b-45e4-a4eb-6f3b1f153a11 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.939867] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5560c9f-3afc-4b94-bd94-f74d384b3beb {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.173198] env[62066]: DEBUG oslo_concurrency.lockutils [None req-bb369cbc-f881-4f95-b841-517845c5cb59 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.315s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 936.173770] env[62066]: DEBUG nova.compute.manager [None req-bb369cbc-f881-4f95-b841-517845c5cb59 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: 5693eea2-1cac-45d5-aa27-58526b65136e] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 936.183975] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9f08ab51-33e2-4c97-838b-004b2127a87f tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 1.456s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 936.183975] env[62066]: DEBUG nova.objects.instance [None req-9f08ab51-33e2-4c97-838b-004b2127a87f tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] [instance: 15afaea0-65de-4f32-851c-365003e5498f] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62066) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 936.240080] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1341177, 'name': CreateVM_Task, 'duration_secs': 0.499901} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.240265] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 20987cb6-e12c-48c5-8fae-d990b1bb6b8a] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 936.240926] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d267d329-25e8-414c-9540-a1c71897bff8 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 936.241204] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d267d329-25e8-414c-9540-a1c71897bff8 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Acquired lock "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 936.241455] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d267d329-25e8-414c-9540-a1c71897bff8 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 936.241712] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e8c914a1-261c-4d9e-8356-f0602beeec2f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.246402] env[62066]: DEBUG oslo_vmware.api [None req-d267d329-25e8-414c-9540-a1c71897bff8 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Waiting for the task: (returnval){ [ 936.246402] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]52e1f0eb-4dbd-cb6a-e855-ea9942f155e5" [ 936.246402] env[62066]: _type = "Task" [ 936.246402] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.254554] env[62066]: DEBUG oslo_vmware.api [None req-d267d329-25e8-414c-9540-a1c71897bff8 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52e1f0eb-4dbd-cb6a-e855-ea9942f155e5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.293197] env[62066]: DEBUG oslo_concurrency.lockutils [None req-594ce8c1-1ab1-41b6-bcf3-6a5287350f9c tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Releasing lock "refresh_cache-15afaea0-65de-4f32-851c-365003e5498f" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 936.293197] env[62066]: DEBUG nova.compute.manager [None req-594ce8c1-1ab1-41b6-bcf3-6a5287350f9c tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] [instance: 15afaea0-65de-4f32-851c-365003e5498f] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 936.293197] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-594ce8c1-1ab1-41b6-bcf3-6a5287350f9c tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] [instance: 15afaea0-65de-4f32-851c-365003e5498f] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 936.294133] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf5e45a4-36f9-48da-ad97-855c337bfc2d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.303444] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-594ce8c1-1ab1-41b6-bcf3-6a5287350f9c tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] [instance: 15afaea0-65de-4f32-851c-365003e5498f] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 936.303528] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6eecc4c4-b704-415c-b2b7-0edaf2233d51 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.309246] env[62066]: DEBUG oslo_vmware.api [None req-594ce8c1-1ab1-41b6-bcf3-6a5287350f9c tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Waiting for the task: (returnval){ [ 936.309246] env[62066]: value = "task-1341178" [ 936.309246] env[62066]: _type = "Task" [ 936.309246] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.317471] env[62066]: DEBUG oslo_vmware.api [None req-594ce8c1-1ab1-41b6-bcf3-6a5287350f9c tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Task: {'id': task-1341178, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.540120] env[62066]: DEBUG nova.network.neutron [req-731f5a42-b150-47e9-ad1c-ed5fdd2b1b97 req-3ed0ad5a-e4e4-4391-a0f1-2f69b24006e9 service nova] [instance: 20987cb6-e12c-48c5-8fae-d990b1bb6b8a] Updated VIF entry in instance network info cache for port f24a3aad-eb9f-4bd9-aa9d-e7cc51421b37. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 936.540563] env[62066]: DEBUG nova.network.neutron [req-731f5a42-b150-47e9-ad1c-ed5fdd2b1b97 req-3ed0ad5a-e4e4-4391-a0f1-2f69b24006e9 service nova] [instance: 20987cb6-e12c-48c5-8fae-d990b1bb6b8a] Updating instance_info_cache with network_info: [{"id": "f24a3aad-eb9f-4bd9-aa9d-e7cc51421b37", "address": "fa:16:3e:62:1f:dd", "network": {"id": "9f3396ce-ee48-4eaf-8cc7-3e1711f37537", "bridge": "br-int", "label": "tempest-ServersTestJSON-2129775093-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6497ab02f327476d8ff81c2ecc0371e5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7ab8d568-adb0-4f3b-b6cc-68413e6546ae", "external-id": "nsx-vlan-transportzone-86", "segmentation_id": 86, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf24a3aad-eb", "ovs_interfaceid": "f24a3aad-eb9f-4bd9-aa9d-e7cc51421b37", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 936.650837] env[62066]: DEBUG nova.network.neutron [None req-54702e84-4f3e-4046-b83c-40cd5c6f41f7 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 651e5142-7ab0-4040-97b4-4599bd4e419a] Successfully updated port: 66054023-ce27-4e2d-92f5-18f1dde734d0 {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 936.689253] env[62066]: DEBUG nova.compute.utils [None req-bb369cbc-f881-4f95-b841-517845c5cb59 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 936.693468] env[62066]: DEBUG nova.compute.manager [None req-bb369cbc-f881-4f95-b841-517845c5cb59 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: 5693eea2-1cac-45d5-aa27-58526b65136e] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 936.694090] env[62066]: DEBUG nova.network.neutron [None req-bb369cbc-f881-4f95-b841-517845c5cb59 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: 5693eea2-1cac-45d5-aa27-58526b65136e] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 936.758094] env[62066]: DEBUG oslo_vmware.api [None req-d267d329-25e8-414c-9540-a1c71897bff8 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52e1f0eb-4dbd-cb6a-e855-ea9942f155e5, 'name': SearchDatastore_Task, 'duration_secs': 0.017487} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.758402] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d267d329-25e8-414c-9540-a1c71897bff8 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Releasing lock "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 936.758695] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-d267d329-25e8-414c-9540-a1c71897bff8 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 20987cb6-e12c-48c5-8fae-d990b1bb6b8a] Processing image 50ff584c-3b50-4395-af07-3e66769bc9f7 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 936.758944] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d267d329-25e8-414c-9540-a1c71897bff8 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 936.759109] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d267d329-25e8-414c-9540-a1c71897bff8 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Acquired lock "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 936.759292] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-d267d329-25e8-414c-9540-a1c71897bff8 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 936.759821] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-36fbbdfa-e98c-4970-93a2-1a77cde3b5d1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.773132] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-d267d329-25e8-414c-9540-a1c71897bff8 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 936.773340] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-d267d329-25e8-414c-9540-a1c71897bff8 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 936.775971] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0fb77d89-dae1-410c-93bc-4aef3055b3e6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.780603] env[62066]: DEBUG oslo_vmware.api [None req-d267d329-25e8-414c-9540-a1c71897bff8 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Waiting for the task: (returnval){ [ 936.780603] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]526b0d94-4506-7ebf-d008-61c6d248f241" [ 936.780603] env[62066]: _type = "Task" [ 936.780603] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.786976] env[62066]: DEBUG nova.policy [None req-bb369cbc-f881-4f95-b841-517845c5cb59 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '588e53cee85f4ab484b76e7a59fcbe78', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8475ad5a900548cba568360999c846ea', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 936.789946] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b37dd465-c64a-41d9-9bff-6087484093ac tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Acquiring lock "9c989152-ec69-478e-a0b0-62983852576c" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 936.790290] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b37dd465-c64a-41d9-9bff-6087484093ac tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Lock "9c989152-ec69-478e-a0b0-62983852576c" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 936.790490] env[62066]: DEBUG nova.compute.manager [None req-b37dd465-c64a-41d9-9bff-6087484093ac tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 9c989152-ec69-478e-a0b0-62983852576c] Going to confirm migration 1 {{(pid=62066) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 936.796031] env[62066]: DEBUG oslo_vmware.api [None req-d267d329-25e8-414c-9540-a1c71897bff8 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]526b0d94-4506-7ebf-d008-61c6d248f241, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.822706] env[62066]: DEBUG oslo_vmware.api [None req-594ce8c1-1ab1-41b6-bcf3-6a5287350f9c tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Task: {'id': task-1341178, 'name': PowerOffVM_Task, 'duration_secs': 0.137335} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.822967] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-594ce8c1-1ab1-41b6-bcf3-6a5287350f9c tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] [instance: 15afaea0-65de-4f32-851c-365003e5498f] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 936.823145] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-594ce8c1-1ab1-41b6-bcf3-6a5287350f9c tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] [instance: 15afaea0-65de-4f32-851c-365003e5498f] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 936.823397] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-315dcce6-c671-4977-89ea-2b7d9662ad99 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.849211] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-594ce8c1-1ab1-41b6-bcf3-6a5287350f9c tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] [instance: 15afaea0-65de-4f32-851c-365003e5498f] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 936.849440] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-594ce8c1-1ab1-41b6-bcf3-6a5287350f9c tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] [instance: 15afaea0-65de-4f32-851c-365003e5498f] Deleting contents of the VM from datastore datastore1 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 936.849629] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-594ce8c1-1ab1-41b6-bcf3-6a5287350f9c tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Deleting the datastore file [datastore1] 15afaea0-65de-4f32-851c-365003e5498f {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 936.849957] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-fd7d995b-d9b3-4cdf-80d1-cca982c9f6cf {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.857058] env[62066]: DEBUG oslo_vmware.api [None req-594ce8c1-1ab1-41b6-bcf3-6a5287350f9c tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Waiting for the task: (returnval){ [ 936.857058] env[62066]: value = "task-1341180" [ 936.857058] env[62066]: _type = "Task" [ 936.857058] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.866594] env[62066]: DEBUG oslo_vmware.api [None req-594ce8c1-1ab1-41b6-bcf3-6a5287350f9c tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Task: {'id': task-1341180, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.045630] env[62066]: DEBUG oslo_concurrency.lockutils [req-731f5a42-b150-47e9-ad1c-ed5fdd2b1b97 req-3ed0ad5a-e4e4-4391-a0f1-2f69b24006e9 service nova] Releasing lock "refresh_cache-20987cb6-e12c-48c5-8fae-d990b1bb6b8a" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 937.156294] env[62066]: DEBUG oslo_concurrency.lockutils [None req-54702e84-4f3e-4046-b83c-40cd5c6f41f7 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Acquiring lock "refresh_cache-651e5142-7ab0-4040-97b4-4599bd4e419a" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 937.156453] env[62066]: DEBUG oslo_concurrency.lockutils [None req-54702e84-4f3e-4046-b83c-40cd5c6f41f7 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Acquired lock "refresh_cache-651e5142-7ab0-4040-97b4-4599bd4e419a" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 937.156662] env[62066]: DEBUG nova.network.neutron [None req-54702e84-4f3e-4046-b83c-40cd5c6f41f7 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 651e5142-7ab0-4040-97b4-4599bd4e419a] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 937.194665] env[62066]: DEBUG nova.compute.manager [None req-bb369cbc-f881-4f95-b841-517845c5cb59 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: 5693eea2-1cac-45d5-aa27-58526b65136e] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 937.199802] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9f08ab51-33e2-4c97-838b-004b2127a87f tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.022s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 937.233768] env[62066]: DEBUG oslo_concurrency.lockutils [None req-074ad1d4-5eea-488a-a57c-2b94f8a8b541 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Acquiring lock "interface-b718b09b-a7fa-445b-8be3-5b3ebca210a2-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 937.234153] env[62066]: DEBUG oslo_concurrency.lockutils [None req-074ad1d4-5eea-488a-a57c-2b94f8a8b541 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Lock "interface-b718b09b-a7fa-445b-8be3-5b3ebca210a2-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 937.234503] env[62066]: DEBUG nova.objects.instance [None req-074ad1d4-5eea-488a-a57c-2b94f8a8b541 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Lazy-loading 'flavor' on Instance uuid b718b09b-a7fa-445b-8be3-5b3ebca210a2 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 937.253357] env[62066]: DEBUG nova.network.neutron [None req-bb369cbc-f881-4f95-b841-517845c5cb59 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: 5693eea2-1cac-45d5-aa27-58526b65136e] Successfully created port: 7e38fd56-d921-4784-9206-fdff79574fa4 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 937.294426] env[62066]: DEBUG oslo_vmware.api [None req-d267d329-25e8-414c-9540-a1c71897bff8 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]526b0d94-4506-7ebf-d008-61c6d248f241, 'name': SearchDatastore_Task, 'duration_secs': 0.044021} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.295324] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ef431817-4656-44e3-86af-5689e8896786 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.307218] env[62066]: DEBUG oslo_vmware.api [None req-d267d329-25e8-414c-9540-a1c71897bff8 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Waiting for the task: (returnval){ [ 937.307218] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]524370e5-a8bd-41a9-edc2-cedbffaeffd5" [ 937.307218] env[62066]: _type = "Task" [ 937.307218] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.316190] env[62066]: DEBUG oslo_vmware.api [None req-d267d329-25e8-414c-9540-a1c71897bff8 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]524370e5-a8bd-41a9-edc2-cedbffaeffd5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.366490] env[62066]: DEBUG oslo_vmware.api [None req-594ce8c1-1ab1-41b6-bcf3-6a5287350f9c tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Task: {'id': task-1341180, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.179556} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.366735] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-594ce8c1-1ab1-41b6-bcf3-6a5287350f9c tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 937.366905] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-594ce8c1-1ab1-41b6-bcf3-6a5287350f9c tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] [instance: 15afaea0-65de-4f32-851c-365003e5498f] Deleted contents of the VM from datastore datastore1 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 937.367107] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-594ce8c1-1ab1-41b6-bcf3-6a5287350f9c tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] [instance: 15afaea0-65de-4f32-851c-365003e5498f] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 937.367294] env[62066]: INFO nova.compute.manager [None req-594ce8c1-1ab1-41b6-bcf3-6a5287350f9c tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] [instance: 15afaea0-65de-4f32-851c-365003e5498f] Took 1.07 seconds to destroy the instance on the hypervisor. [ 937.367552] env[62066]: DEBUG oslo.service.loopingcall [None req-594ce8c1-1ab1-41b6-bcf3-6a5287350f9c tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 937.367761] env[62066]: DEBUG nova.compute.manager [-] [instance: 15afaea0-65de-4f32-851c-365003e5498f] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 937.367856] env[62066]: DEBUG nova.network.neutron [-] [instance: 15afaea0-65de-4f32-851c-365003e5498f] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 937.387717] env[62066]: DEBUG nova.network.neutron [-] [instance: 15afaea0-65de-4f32-851c-365003e5498f] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 937.407241] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b37dd465-c64a-41d9-9bff-6087484093ac tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Acquiring lock "refresh_cache-9c989152-ec69-478e-a0b0-62983852576c" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 937.407446] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b37dd465-c64a-41d9-9bff-6087484093ac tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Acquired lock "refresh_cache-9c989152-ec69-478e-a0b0-62983852576c" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 937.407624] env[62066]: DEBUG nova.network.neutron [None req-b37dd465-c64a-41d9-9bff-6087484093ac tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 9c989152-ec69-478e-a0b0-62983852576c] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 937.407814] env[62066]: DEBUG nova.objects.instance [None req-b37dd465-c64a-41d9-9bff-6087484093ac tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Lazy-loading 'info_cache' on Instance uuid 9c989152-ec69-478e-a0b0-62983852576c {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 937.636509] env[62066]: DEBUG nova.compute.manager [req-94ca8320-f6b2-45f5-a10b-2f0463b2481f req-893d4c43-0986-4387-b3ed-a64a33b22d2c service nova] [instance: 651e5142-7ab0-4040-97b4-4599bd4e419a] Received event network-vif-plugged-66054023-ce27-4e2d-92f5-18f1dde734d0 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 937.636754] env[62066]: DEBUG oslo_concurrency.lockutils [req-94ca8320-f6b2-45f5-a10b-2f0463b2481f req-893d4c43-0986-4387-b3ed-a64a33b22d2c service nova] Acquiring lock "651e5142-7ab0-4040-97b4-4599bd4e419a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 937.637042] env[62066]: DEBUG oslo_concurrency.lockutils [req-94ca8320-f6b2-45f5-a10b-2f0463b2481f req-893d4c43-0986-4387-b3ed-a64a33b22d2c service nova] Lock "651e5142-7ab0-4040-97b4-4599bd4e419a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 937.637270] env[62066]: DEBUG oslo_concurrency.lockutils [req-94ca8320-f6b2-45f5-a10b-2f0463b2481f req-893d4c43-0986-4387-b3ed-a64a33b22d2c service nova] Lock "651e5142-7ab0-4040-97b4-4599bd4e419a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 937.637478] env[62066]: DEBUG nova.compute.manager [req-94ca8320-f6b2-45f5-a10b-2f0463b2481f req-893d4c43-0986-4387-b3ed-a64a33b22d2c service nova] [instance: 651e5142-7ab0-4040-97b4-4599bd4e419a] No waiting events found dispatching network-vif-plugged-66054023-ce27-4e2d-92f5-18f1dde734d0 {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 937.637683] env[62066]: WARNING nova.compute.manager [req-94ca8320-f6b2-45f5-a10b-2f0463b2481f req-893d4c43-0986-4387-b3ed-a64a33b22d2c service nova] [instance: 651e5142-7ab0-4040-97b4-4599bd4e419a] Received unexpected event network-vif-plugged-66054023-ce27-4e2d-92f5-18f1dde734d0 for instance with vm_state building and task_state spawning. [ 937.637902] env[62066]: DEBUG nova.compute.manager [req-94ca8320-f6b2-45f5-a10b-2f0463b2481f req-893d4c43-0986-4387-b3ed-a64a33b22d2c service nova] [instance: 651e5142-7ab0-4040-97b4-4599bd4e419a] Received event network-changed-66054023-ce27-4e2d-92f5-18f1dde734d0 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 937.638140] env[62066]: DEBUG nova.compute.manager [req-94ca8320-f6b2-45f5-a10b-2f0463b2481f req-893d4c43-0986-4387-b3ed-a64a33b22d2c service nova] [instance: 651e5142-7ab0-4040-97b4-4599bd4e419a] Refreshing instance network info cache due to event network-changed-66054023-ce27-4e2d-92f5-18f1dde734d0. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 937.638398] env[62066]: DEBUG oslo_concurrency.lockutils [req-94ca8320-f6b2-45f5-a10b-2f0463b2481f req-893d4c43-0986-4387-b3ed-a64a33b22d2c service nova] Acquiring lock "refresh_cache-651e5142-7ab0-4040-97b4-4599bd4e419a" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 937.690576] env[62066]: DEBUG nova.network.neutron [None req-54702e84-4f3e-4046-b83c-40cd5c6f41f7 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 651e5142-7ab0-4040-97b4-4599bd4e419a] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 937.740661] env[62066]: DEBUG nova.objects.instance [None req-074ad1d4-5eea-488a-a57c-2b94f8a8b541 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Lazy-loading 'pci_requests' on Instance uuid b718b09b-a7fa-445b-8be3-5b3ebca210a2 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 937.819032] env[62066]: DEBUG oslo_vmware.api [None req-d267d329-25e8-414c-9540-a1c71897bff8 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]524370e5-a8bd-41a9-edc2-cedbffaeffd5, 'name': SearchDatastore_Task, 'duration_secs': 0.042094} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.819032] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d267d329-25e8-414c-9540-a1c71897bff8 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Releasing lock "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 937.819032] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-d267d329-25e8-414c-9540-a1c71897bff8 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk to [datastore1] 20987cb6-e12c-48c5-8fae-d990b1bb6b8a/20987cb6-e12c-48c5-8fae-d990b1bb6b8a.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 937.819032] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-de6eac75-5ca7-4ebd-9267-b8a37f86f93d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.826040] env[62066]: DEBUG oslo_vmware.api [None req-d267d329-25e8-414c-9540-a1c71897bff8 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Waiting for the task: (returnval){ [ 937.826040] env[62066]: value = "task-1341181" [ 937.826040] env[62066]: _type = "Task" [ 937.826040] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.834365] env[62066]: DEBUG oslo_vmware.api [None req-d267d329-25e8-414c-9540-a1c71897bff8 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': task-1341181, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.835506] env[62066]: DEBUG nova.network.neutron [None req-54702e84-4f3e-4046-b83c-40cd5c6f41f7 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 651e5142-7ab0-4040-97b4-4599bd4e419a] Updating instance_info_cache with network_info: [{"id": "66054023-ce27-4e2d-92f5-18f1dde734d0", "address": "fa:16:3e:ea:f1:cb", "network": {"id": "a3852490-7a1f-499c-813a-46cbf52318da", "bridge": "br-int", "label": "tempest-ImagesTestJSON-735840736-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "75f0352c852947369474c9e05766c584", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "50171613-b419-45e3-9ada-fcb6cd921428", "external-id": "nsx-vlan-transportzone-914", "segmentation_id": 914, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap66054023-ce", "ovs_interfaceid": "66054023-ce27-4e2d-92f5-18f1dde734d0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 937.889928] env[62066]: DEBUG nova.network.neutron [-] [instance: 15afaea0-65de-4f32-851c-365003e5498f] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 938.207527] env[62066]: DEBUG nova.compute.manager [None req-bb369cbc-f881-4f95-b841-517845c5cb59 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: 5693eea2-1cac-45d5-aa27-58526b65136e] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 938.238015] env[62066]: DEBUG nova.virt.hardware [None req-bb369cbc-f881-4f95-b841-517845c5cb59 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-23T13:40:41Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-23T13:40:23Z,direct_url=,disk_format='vmdk',id=50ff584c-3b50-4395-af07-3e66769bc9f7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='eb52f7069a374c61ae946f052007c6d9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-23T13:40:24Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 938.238283] env[62066]: DEBUG nova.virt.hardware [None req-bb369cbc-f881-4f95-b841-517845c5cb59 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 938.238451] env[62066]: DEBUG nova.virt.hardware [None req-bb369cbc-f881-4f95-b841-517845c5cb59 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 938.238641] env[62066]: DEBUG nova.virt.hardware [None req-bb369cbc-f881-4f95-b841-517845c5cb59 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 938.238793] env[62066]: DEBUG nova.virt.hardware [None req-bb369cbc-f881-4f95-b841-517845c5cb59 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 938.238949] env[62066]: DEBUG nova.virt.hardware [None req-bb369cbc-f881-4f95-b841-517845c5cb59 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 938.239186] env[62066]: DEBUG nova.virt.hardware [None req-bb369cbc-f881-4f95-b841-517845c5cb59 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 938.239359] env[62066]: DEBUG nova.virt.hardware [None req-bb369cbc-f881-4f95-b841-517845c5cb59 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 938.239534] env[62066]: DEBUG nova.virt.hardware [None req-bb369cbc-f881-4f95-b841-517845c5cb59 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 938.239702] env[62066]: DEBUG nova.virt.hardware [None req-bb369cbc-f881-4f95-b841-517845c5cb59 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 938.239879] env[62066]: DEBUG nova.virt.hardware [None req-bb369cbc-f881-4f95-b841-517845c5cb59 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 938.241094] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-828acd05-02aa-4400-8bc3-cd9b0b3d71d9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.244077] env[62066]: DEBUG nova.objects.base [None req-074ad1d4-5eea-488a-a57c-2b94f8a8b541 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=62066) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 938.244286] env[62066]: DEBUG nova.network.neutron [None req-074ad1d4-5eea-488a-a57c-2b94f8a8b541 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: b718b09b-a7fa-445b-8be3-5b3ebca210a2] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 938.251961] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c5feb98-d6e0-45b2-89a2-b433d1461cdb {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.328638] env[62066]: DEBUG oslo_concurrency.lockutils [None req-074ad1d4-5eea-488a-a57c-2b94f8a8b541 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Lock "interface-b718b09b-a7fa-445b-8be3-5b3ebca210a2-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 1.094s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 938.340806] env[62066]: DEBUG oslo_concurrency.lockutils [None req-54702e84-4f3e-4046-b83c-40cd5c6f41f7 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Releasing lock "refresh_cache-651e5142-7ab0-4040-97b4-4599bd4e419a" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 938.340806] env[62066]: DEBUG nova.compute.manager [None req-54702e84-4f3e-4046-b83c-40cd5c6f41f7 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 651e5142-7ab0-4040-97b4-4599bd4e419a] Instance network_info: |[{"id": "66054023-ce27-4e2d-92f5-18f1dde734d0", "address": "fa:16:3e:ea:f1:cb", "network": {"id": "a3852490-7a1f-499c-813a-46cbf52318da", "bridge": "br-int", "label": "tempest-ImagesTestJSON-735840736-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "75f0352c852947369474c9e05766c584", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "50171613-b419-45e3-9ada-fcb6cd921428", "external-id": "nsx-vlan-transportzone-914", "segmentation_id": 914, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap66054023-ce", "ovs_interfaceid": "66054023-ce27-4e2d-92f5-18f1dde734d0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 938.341131] env[62066]: DEBUG oslo_vmware.api [None req-d267d329-25e8-414c-9540-a1c71897bff8 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': task-1341181, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.341450] env[62066]: DEBUG oslo_concurrency.lockutils [req-94ca8320-f6b2-45f5-a10b-2f0463b2481f req-893d4c43-0986-4387-b3ed-a64a33b22d2c service nova] Acquired lock "refresh_cache-651e5142-7ab0-4040-97b4-4599bd4e419a" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 938.341637] env[62066]: DEBUG nova.network.neutron [req-94ca8320-f6b2-45f5-a10b-2f0463b2481f req-893d4c43-0986-4387-b3ed-a64a33b22d2c service nova] [instance: 651e5142-7ab0-4040-97b4-4599bd4e419a] Refreshing network info cache for port 66054023-ce27-4e2d-92f5-18f1dde734d0 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 938.343350] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-54702e84-4f3e-4046-b83c-40cd5c6f41f7 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 651e5142-7ab0-4040-97b4-4599bd4e419a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ea:f1:cb', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '50171613-b419-45e3-9ada-fcb6cd921428', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '66054023-ce27-4e2d-92f5-18f1dde734d0', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 938.351992] env[62066]: DEBUG oslo.service.loopingcall [None req-54702e84-4f3e-4046-b83c-40cd5c6f41f7 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 938.353515] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 651e5142-7ab0-4040-97b4-4599bd4e419a] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 938.353813] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-327da71d-0a67-4e61-ac80-987d59d0a4d5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.375489] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 938.375489] env[62066]: value = "task-1341182" [ 938.375489] env[62066]: _type = "Task" [ 938.375489] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.385212] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1341182, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.392448] env[62066]: INFO nova.compute.manager [-] [instance: 15afaea0-65de-4f32-851c-365003e5498f] Took 1.02 seconds to deallocate network for instance. [ 938.838635] env[62066]: DEBUG nova.network.neutron [None req-bb369cbc-f881-4f95-b841-517845c5cb59 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: 5693eea2-1cac-45d5-aa27-58526b65136e] Successfully updated port: 7e38fd56-d921-4784-9206-fdff79574fa4 {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 938.846782] env[62066]: DEBUG oslo_vmware.api [None req-d267d329-25e8-414c-9540-a1c71897bff8 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': task-1341181, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.871453} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.846782] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-d267d329-25e8-414c-9540-a1c71897bff8 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk to [datastore1] 20987cb6-e12c-48c5-8fae-d990b1bb6b8a/20987cb6-e12c-48c5-8fae-d990b1bb6b8a.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 938.846782] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-d267d329-25e8-414c-9540-a1c71897bff8 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 20987cb6-e12c-48c5-8fae-d990b1bb6b8a] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 938.846782] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-483bc892-31a4-46e4-947e-16d7bc9e84ac {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.857087] env[62066]: DEBUG oslo_vmware.api [None req-d267d329-25e8-414c-9540-a1c71897bff8 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Waiting for the task: (returnval){ [ 938.857087] env[62066]: value = "task-1341183" [ 938.857087] env[62066]: _type = "Task" [ 938.857087] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.866774] env[62066]: DEBUG oslo_vmware.api [None req-d267d329-25e8-414c-9540-a1c71897bff8 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': task-1341183, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.867835] env[62066]: DEBUG nova.network.neutron [None req-b37dd465-c64a-41d9-9bff-6087484093ac tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 9c989152-ec69-478e-a0b0-62983852576c] Updating instance_info_cache with network_info: [{"id": "1c0d0075-7f22-441e-a283-21e9b5c5aecc", "address": "fa:16:3e:c5:c3:07", "network": {"id": "1c6c51e4-96f8-4175-b7f5-96b8a42fcb13", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-2006089809-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d609babc987a42e2a8ddb4bfb9c3b103", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a965790c-2d2f-4c2a-9ee7-745f4d53039b", "external-id": "nsx-vlan-transportzone-708", "segmentation_id": 708, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1c0d0075-7f", "ovs_interfaceid": "1c0d0075-7f22-441e-a283-21e9b5c5aecc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 938.885995] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1341182, 'name': CreateVM_Task} progress is 25%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.899736] env[62066]: DEBUG oslo_concurrency.lockutils [None req-594ce8c1-1ab1-41b6-bcf3-6a5287350f9c tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 938.900054] env[62066]: DEBUG oslo_concurrency.lockutils [None req-594ce8c1-1ab1-41b6-bcf3-6a5287350f9c tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 938.900347] env[62066]: DEBUG nova.objects.instance [None req-594ce8c1-1ab1-41b6-bcf3-6a5287350f9c tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Lazy-loading 'resources' on Instance uuid 15afaea0-65de-4f32-851c-365003e5498f {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 939.156172] env[62066]: DEBUG nova.network.neutron [req-94ca8320-f6b2-45f5-a10b-2f0463b2481f req-893d4c43-0986-4387-b3ed-a64a33b22d2c service nova] [instance: 651e5142-7ab0-4040-97b4-4599bd4e419a] Updated VIF entry in instance network info cache for port 66054023-ce27-4e2d-92f5-18f1dde734d0. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 939.156805] env[62066]: DEBUG nova.network.neutron [req-94ca8320-f6b2-45f5-a10b-2f0463b2481f req-893d4c43-0986-4387-b3ed-a64a33b22d2c service nova] [instance: 651e5142-7ab0-4040-97b4-4599bd4e419a] Updating instance_info_cache with network_info: [{"id": "66054023-ce27-4e2d-92f5-18f1dde734d0", "address": "fa:16:3e:ea:f1:cb", "network": {"id": "a3852490-7a1f-499c-813a-46cbf52318da", "bridge": "br-int", "label": "tempest-ImagesTestJSON-735840736-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "75f0352c852947369474c9e05766c584", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "50171613-b419-45e3-9ada-fcb6cd921428", "external-id": "nsx-vlan-transportzone-914", "segmentation_id": 914, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap66054023-ce", "ovs_interfaceid": "66054023-ce27-4e2d-92f5-18f1dde734d0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 939.346935] env[62066]: DEBUG oslo_concurrency.lockutils [None req-bb369cbc-f881-4f95-b841-517845c5cb59 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Acquiring lock "refresh_cache-5693eea2-1cac-45d5-aa27-58526b65136e" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 939.347139] env[62066]: DEBUG oslo_concurrency.lockutils [None req-bb369cbc-f881-4f95-b841-517845c5cb59 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Acquired lock "refresh_cache-5693eea2-1cac-45d5-aa27-58526b65136e" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 939.347301] env[62066]: DEBUG nova.network.neutron [None req-bb369cbc-f881-4f95-b841-517845c5cb59 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: 5693eea2-1cac-45d5-aa27-58526b65136e] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 939.367053] env[62066]: DEBUG oslo_vmware.api [None req-d267d329-25e8-414c-9540-a1c71897bff8 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': task-1341183, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067352} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.367385] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-d267d329-25e8-414c-9540-a1c71897bff8 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 20987cb6-e12c-48c5-8fae-d990b1bb6b8a] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 939.368200] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db2eec9f-1e27-4319-a228-10ef50c6f5c4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.370888] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b37dd465-c64a-41d9-9bff-6087484093ac tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Releasing lock "refresh_cache-9c989152-ec69-478e-a0b0-62983852576c" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 939.371134] env[62066]: DEBUG nova.objects.instance [None req-b37dd465-c64a-41d9-9bff-6087484093ac tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Lazy-loading 'migration_context' on Instance uuid 9c989152-ec69-478e-a0b0-62983852576c {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 939.392316] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-d267d329-25e8-414c-9540-a1c71897bff8 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 20987cb6-e12c-48c5-8fae-d990b1bb6b8a] Reconfiguring VM instance instance-0000005c to attach disk [datastore1] 20987cb6-e12c-48c5-8fae-d990b1bb6b8a/20987cb6-e12c-48c5-8fae-d990b1bb6b8a.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 939.396368] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-eff456a0-5813-440f-a005-d59404a08249 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.418348] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1341182, 'name': CreateVM_Task, 'duration_secs': 1.022857} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.419582] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 651e5142-7ab0-4040-97b4-4599bd4e419a] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 939.419965] env[62066]: DEBUG oslo_vmware.api [None req-d267d329-25e8-414c-9540-a1c71897bff8 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Waiting for the task: (returnval){ [ 939.419965] env[62066]: value = "task-1341184" [ 939.419965] env[62066]: _type = "Task" [ 939.419965] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.420624] env[62066]: DEBUG oslo_concurrency.lockutils [None req-54702e84-4f3e-4046-b83c-40cd5c6f41f7 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 939.420786] env[62066]: DEBUG oslo_concurrency.lockutils [None req-54702e84-4f3e-4046-b83c-40cd5c6f41f7 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Acquired lock "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 939.421115] env[62066]: DEBUG oslo_concurrency.lockutils [None req-54702e84-4f3e-4046-b83c-40cd5c6f41f7 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 939.421424] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f706ca00-4e7a-4bf1-86c0-dab2de968b12 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.428884] env[62066]: DEBUG oslo_vmware.api [None req-54702e84-4f3e-4046-b83c-40cd5c6f41f7 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Waiting for the task: (returnval){ [ 939.428884] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]52b31b6b-18e6-dcac-c39c-3c17da09a9c5" [ 939.428884] env[62066]: _type = "Task" [ 939.428884] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.431774] env[62066]: DEBUG oslo_vmware.api [None req-d267d329-25e8-414c-9540-a1c71897bff8 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': task-1341184, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.441797] env[62066]: DEBUG oslo_vmware.api [None req-54702e84-4f3e-4046-b83c-40cd5c6f41f7 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52b31b6b-18e6-dcac-c39c-3c17da09a9c5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.558169] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-730be412-54ee-44b8-913f-50bb02115317 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.565674] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab0ab5ab-dcdd-49be-b6d1-04cadd29534e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.596822] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aaded69e-a229-43b6-b7be-94a0c671fdfd {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.604063] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fc52b35-0da9-4e13-8aee-df3760552b35 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.617359] env[62066]: DEBUG nova.compute.provider_tree [None req-594ce8c1-1ab1-41b6-bcf3-6a5287350f9c tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 939.660520] env[62066]: DEBUG oslo_concurrency.lockutils [req-94ca8320-f6b2-45f5-a10b-2f0463b2481f req-893d4c43-0986-4387-b3ed-a64a33b22d2c service nova] Releasing lock "refresh_cache-651e5142-7ab0-4040-97b4-4599bd4e419a" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 939.661717] env[62066]: DEBUG nova.compute.manager [req-ab6edaf1-c61d-402d-b4e0-f356d446c437 req-8c34b220-c481-4325-9788-3ec2196d9226 service nova] [instance: 5693eea2-1cac-45d5-aa27-58526b65136e] Received event network-vif-plugged-7e38fd56-d921-4784-9206-fdff79574fa4 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 939.661927] env[62066]: DEBUG oslo_concurrency.lockutils [req-ab6edaf1-c61d-402d-b4e0-f356d446c437 req-8c34b220-c481-4325-9788-3ec2196d9226 service nova] Acquiring lock "5693eea2-1cac-45d5-aa27-58526b65136e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 939.662141] env[62066]: DEBUG oslo_concurrency.lockutils [req-ab6edaf1-c61d-402d-b4e0-f356d446c437 req-8c34b220-c481-4325-9788-3ec2196d9226 service nova] Lock "5693eea2-1cac-45d5-aa27-58526b65136e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 939.662337] env[62066]: DEBUG oslo_concurrency.lockutils [req-ab6edaf1-c61d-402d-b4e0-f356d446c437 req-8c34b220-c481-4325-9788-3ec2196d9226 service nova] Lock "5693eea2-1cac-45d5-aa27-58526b65136e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 939.662525] env[62066]: DEBUG nova.compute.manager [req-ab6edaf1-c61d-402d-b4e0-f356d446c437 req-8c34b220-c481-4325-9788-3ec2196d9226 service nova] [instance: 5693eea2-1cac-45d5-aa27-58526b65136e] No waiting events found dispatching network-vif-plugged-7e38fd56-d921-4784-9206-fdff79574fa4 {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 939.662697] env[62066]: WARNING nova.compute.manager [req-ab6edaf1-c61d-402d-b4e0-f356d446c437 req-8c34b220-c481-4325-9788-3ec2196d9226 service nova] [instance: 5693eea2-1cac-45d5-aa27-58526b65136e] Received unexpected event network-vif-plugged-7e38fd56-d921-4784-9206-fdff79574fa4 for instance with vm_state building and task_state spawning. [ 939.662863] env[62066]: DEBUG nova.compute.manager [req-ab6edaf1-c61d-402d-b4e0-f356d446c437 req-8c34b220-c481-4325-9788-3ec2196d9226 service nova] [instance: 5693eea2-1cac-45d5-aa27-58526b65136e] Received event network-changed-7e38fd56-d921-4784-9206-fdff79574fa4 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 939.663030] env[62066]: DEBUG nova.compute.manager [req-ab6edaf1-c61d-402d-b4e0-f356d446c437 req-8c34b220-c481-4325-9788-3ec2196d9226 service nova] [instance: 5693eea2-1cac-45d5-aa27-58526b65136e] Refreshing instance network info cache due to event network-changed-7e38fd56-d921-4784-9206-fdff79574fa4. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 939.663206] env[62066]: DEBUG oslo_concurrency.lockutils [req-ab6edaf1-c61d-402d-b4e0-f356d446c437 req-8c34b220-c481-4325-9788-3ec2196d9226 service nova] Acquiring lock "refresh_cache-5693eea2-1cac-45d5-aa27-58526b65136e" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 939.874153] env[62066]: DEBUG nova.objects.base [None req-b37dd465-c64a-41d9-9bff-6087484093ac tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Object Instance<9c989152-ec69-478e-a0b0-62983852576c> lazy-loaded attributes: info_cache,migration_context {{(pid=62066) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 939.875413] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c06f2ad-e979-4e03-afcd-5bb3b007fa6f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.895572] env[62066]: DEBUG nova.network.neutron [None req-bb369cbc-f881-4f95-b841-517845c5cb59 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: 5693eea2-1cac-45d5-aa27-58526b65136e] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 939.897429] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-43a7b8dd-48a2-48d3-98ad-b1676070290c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.903386] env[62066]: DEBUG oslo_vmware.api [None req-b37dd465-c64a-41d9-9bff-6087484093ac tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Waiting for the task: (returnval){ [ 939.903386] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]526dfa99-e0de-4aa8-4eac-2a7ecf49d7ce" [ 939.903386] env[62066]: _type = "Task" [ 939.903386] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.917571] env[62066]: DEBUG oslo_vmware.api [None req-b37dd465-c64a-41d9-9bff-6087484093ac tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]526dfa99-e0de-4aa8-4eac-2a7ecf49d7ce, 'name': SearchDatastore_Task, 'duration_secs': 0.006703} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.918219] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b37dd465-c64a-41d9-9bff-6087484093ac tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 939.929738] env[62066]: DEBUG oslo_vmware.api [None req-d267d329-25e8-414c-9540-a1c71897bff8 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': task-1341184, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.939780] env[62066]: DEBUG oslo_vmware.api [None req-54702e84-4f3e-4046-b83c-40cd5c6f41f7 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52b31b6b-18e6-dcac-c39c-3c17da09a9c5, 'name': SearchDatastore_Task, 'duration_secs': 0.02145} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.942028] env[62066]: DEBUG oslo_concurrency.lockutils [None req-54702e84-4f3e-4046-b83c-40cd5c6f41f7 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Releasing lock "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 939.942278] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-54702e84-4f3e-4046-b83c-40cd5c6f41f7 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 651e5142-7ab0-4040-97b4-4599bd4e419a] Processing image 50ff584c-3b50-4395-af07-3e66769bc9f7 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 939.942534] env[62066]: DEBUG oslo_concurrency.lockutils [None req-54702e84-4f3e-4046-b83c-40cd5c6f41f7 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 939.942683] env[62066]: DEBUG oslo_concurrency.lockutils [None req-54702e84-4f3e-4046-b83c-40cd5c6f41f7 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Acquired lock "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 939.942905] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-54702e84-4f3e-4046-b83c-40cd5c6f41f7 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 939.943181] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b4b79772-02d4-40e9-a443-ff478c2465d9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.951033] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-54702e84-4f3e-4046-b83c-40cd5c6f41f7 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 939.951212] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-54702e84-4f3e-4046-b83c-40cd5c6f41f7 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 939.951884] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-715851c6-6c1b-4218-a16f-50c985d8d10e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.956441] env[62066]: DEBUG oslo_vmware.api [None req-54702e84-4f3e-4046-b83c-40cd5c6f41f7 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Waiting for the task: (returnval){ [ 939.956441] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]52e63117-ba9d-743e-b51b-0380ebdceb14" [ 939.956441] env[62066]: _type = "Task" [ 939.956441] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.964130] env[62066]: DEBUG oslo_vmware.api [None req-54702e84-4f3e-4046-b83c-40cd5c6f41f7 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52e63117-ba9d-743e-b51b-0380ebdceb14, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.074237] env[62066]: DEBUG nova.network.neutron [None req-bb369cbc-f881-4f95-b841-517845c5cb59 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: 5693eea2-1cac-45d5-aa27-58526b65136e] Updating instance_info_cache with network_info: [{"id": "7e38fd56-d921-4784-9206-fdff79574fa4", "address": "fa:16:3e:84:3d:8e", "network": {"id": "7617f887-bdd0-400a-8b5f-606d7785df78", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-558766873-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8475ad5a900548cba568360999c846ea", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2b6a4065-12af-4fb9-ac47-ec9143f7297e", "external-id": "nsx-vlan-transportzone-95", "segmentation_id": 95, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7e38fd56-d9", "ovs_interfaceid": "7e38fd56-d921-4784-9206-fdff79574fa4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 940.120858] env[62066]: DEBUG nova.scheduler.client.report [None req-594ce8c1-1ab1-41b6-bcf3-6a5287350f9c tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 940.276128] env[62066]: DEBUG oslo_concurrency.lockutils [None req-77f7408f-8ed6-44d2-926f-666964de9ad4 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Acquiring lock "interface-b718b09b-a7fa-445b-8be3-5b3ebca210a2-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 940.276368] env[62066]: DEBUG oslo_concurrency.lockutils [None req-77f7408f-8ed6-44d2-926f-666964de9ad4 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Lock "interface-b718b09b-a7fa-445b-8be3-5b3ebca210a2-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 940.276699] env[62066]: DEBUG nova.objects.instance [None req-77f7408f-8ed6-44d2-926f-666964de9ad4 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Lazy-loading 'flavor' on Instance uuid b718b09b-a7fa-445b-8be3-5b3ebca210a2 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 940.431749] env[62066]: DEBUG oslo_vmware.api [None req-d267d329-25e8-414c-9540-a1c71897bff8 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': task-1341184, 'name': ReconfigVM_Task, 'duration_secs': 0.606325} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.432036] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-d267d329-25e8-414c-9540-a1c71897bff8 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 20987cb6-e12c-48c5-8fae-d990b1bb6b8a] Reconfigured VM instance instance-0000005c to attach disk [datastore1] 20987cb6-e12c-48c5-8fae-d990b1bb6b8a/20987cb6-e12c-48c5-8fae-d990b1bb6b8a.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 940.432678] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1b3dbbd9-d843-4049-a92e-daec07032531 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.438214] env[62066]: DEBUG oslo_vmware.api [None req-d267d329-25e8-414c-9540-a1c71897bff8 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Waiting for the task: (returnval){ [ 940.438214] env[62066]: value = "task-1341185" [ 940.438214] env[62066]: _type = "Task" [ 940.438214] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.445191] env[62066]: DEBUG oslo_vmware.api [None req-d267d329-25e8-414c-9540-a1c71897bff8 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': task-1341185, 'name': Rename_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.468050] env[62066]: DEBUG oslo_vmware.api [None req-54702e84-4f3e-4046-b83c-40cd5c6f41f7 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52e63117-ba9d-743e-b51b-0380ebdceb14, 'name': SearchDatastore_Task, 'duration_secs': 0.010581} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.468050] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d6734805-3f1a-4d97-92d0-659c57d4a2f3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.472375] env[62066]: DEBUG oslo_vmware.api [None req-54702e84-4f3e-4046-b83c-40cd5c6f41f7 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Waiting for the task: (returnval){ [ 940.472375] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]52e5f0f9-8cb5-cdc7-5463-6507b6da9bd9" [ 940.472375] env[62066]: _type = "Task" [ 940.472375] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.480448] env[62066]: DEBUG oslo_vmware.api [None req-54702e84-4f3e-4046-b83c-40cd5c6f41f7 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52e5f0f9-8cb5-cdc7-5463-6507b6da9bd9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.577020] env[62066]: DEBUG oslo_concurrency.lockutils [None req-bb369cbc-f881-4f95-b841-517845c5cb59 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Releasing lock "refresh_cache-5693eea2-1cac-45d5-aa27-58526b65136e" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 940.577359] env[62066]: DEBUG nova.compute.manager [None req-bb369cbc-f881-4f95-b841-517845c5cb59 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: 5693eea2-1cac-45d5-aa27-58526b65136e] Instance network_info: |[{"id": "7e38fd56-d921-4784-9206-fdff79574fa4", "address": "fa:16:3e:84:3d:8e", "network": {"id": "7617f887-bdd0-400a-8b5f-606d7785df78", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-558766873-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8475ad5a900548cba568360999c846ea", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2b6a4065-12af-4fb9-ac47-ec9143f7297e", "external-id": "nsx-vlan-transportzone-95", "segmentation_id": 95, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7e38fd56-d9", "ovs_interfaceid": "7e38fd56-d921-4784-9206-fdff79574fa4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 940.577679] env[62066]: DEBUG oslo_concurrency.lockutils [req-ab6edaf1-c61d-402d-b4e0-f356d446c437 req-8c34b220-c481-4325-9788-3ec2196d9226 service nova] Acquired lock "refresh_cache-5693eea2-1cac-45d5-aa27-58526b65136e" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 940.577862] env[62066]: DEBUG nova.network.neutron [req-ab6edaf1-c61d-402d-b4e0-f356d446c437 req-8c34b220-c481-4325-9788-3ec2196d9226 service nova] [instance: 5693eea2-1cac-45d5-aa27-58526b65136e] Refreshing network info cache for port 7e38fd56-d921-4784-9206-fdff79574fa4 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 940.579145] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-bb369cbc-f881-4f95-b841-517845c5cb59 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: 5693eea2-1cac-45d5-aa27-58526b65136e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:84:3d:8e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '2b6a4065-12af-4fb9-ac47-ec9143f7297e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7e38fd56-d921-4784-9206-fdff79574fa4', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 940.586691] env[62066]: DEBUG oslo.service.loopingcall [None req-bb369cbc-f881-4f95-b841-517845c5cb59 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 940.587712] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5693eea2-1cac-45d5-aa27-58526b65136e] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 940.587959] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-740b108f-1e0b-4767-b23d-0ac5f8d229c9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.608114] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 940.608114] env[62066]: value = "task-1341186" [ 940.608114] env[62066]: _type = "Task" [ 940.608114] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.617637] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1341186, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.625741] env[62066]: DEBUG oslo_concurrency.lockutils [None req-594ce8c1-1ab1-41b6-bcf3-6a5287350f9c tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.726s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 940.627835] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b37dd465-c64a-41d9-9bff-6087484093ac tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.710s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 940.644403] env[62066]: INFO nova.scheduler.client.report [None req-594ce8c1-1ab1-41b6-bcf3-6a5287350f9c tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Deleted allocations for instance 15afaea0-65de-4f32-851c-365003e5498f [ 940.854895] env[62066]: DEBUG nova.objects.instance [None req-77f7408f-8ed6-44d2-926f-666964de9ad4 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Lazy-loading 'pci_requests' on Instance uuid b718b09b-a7fa-445b-8be3-5b3ebca210a2 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 940.949476] env[62066]: DEBUG oslo_vmware.api [None req-d267d329-25e8-414c-9540-a1c71897bff8 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': task-1341185, 'name': Rename_Task} progress is 99%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.981832] env[62066]: DEBUG oslo_vmware.api [None req-54702e84-4f3e-4046-b83c-40cd5c6f41f7 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52e5f0f9-8cb5-cdc7-5463-6507b6da9bd9, 'name': SearchDatastore_Task, 'duration_secs': 0.008776} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.982079] env[62066]: DEBUG oslo_concurrency.lockutils [None req-54702e84-4f3e-4046-b83c-40cd5c6f41f7 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Releasing lock "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 940.982346] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-54702e84-4f3e-4046-b83c-40cd5c6f41f7 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk to [datastore1] 651e5142-7ab0-4040-97b4-4599bd4e419a/651e5142-7ab0-4040-97b4-4599bd4e419a.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 940.982600] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8b8ff5fb-b43f-4763-8fb4-b75104c10e37 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.989054] env[62066]: DEBUG oslo_vmware.api [None req-54702e84-4f3e-4046-b83c-40cd5c6f41f7 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Waiting for the task: (returnval){ [ 940.989054] env[62066]: value = "task-1341187" [ 940.989054] env[62066]: _type = "Task" [ 940.989054] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.995880] env[62066]: DEBUG oslo_vmware.api [None req-54702e84-4f3e-4046-b83c-40cd5c6f41f7 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': task-1341187, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.118007] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1341186, 'name': CreateVM_Task, 'duration_secs': 0.336126} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.120103] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5693eea2-1cac-45d5-aa27-58526b65136e] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 941.120763] env[62066]: DEBUG oslo_concurrency.lockutils [None req-bb369cbc-f881-4f95-b841-517845c5cb59 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 941.120927] env[62066]: DEBUG oslo_concurrency.lockutils [None req-bb369cbc-f881-4f95-b841-517845c5cb59 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Acquired lock "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 941.121256] env[62066]: DEBUG oslo_concurrency.lockutils [None req-bb369cbc-f881-4f95-b841-517845c5cb59 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 941.122069] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ae6d5a16-fb3b-450d-9e39-06dc7f2f5f9e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.126433] env[62066]: DEBUG oslo_vmware.api [None req-bb369cbc-f881-4f95-b841-517845c5cb59 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Waiting for the task: (returnval){ [ 941.126433] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]52ca4591-596c-8079-c68d-70b8bf690cce" [ 941.126433] env[62066]: _type = "Task" [ 941.126433] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.134994] env[62066]: DEBUG oslo_vmware.api [None req-bb369cbc-f881-4f95-b841-517845c5cb59 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52ca4591-596c-8079-c68d-70b8bf690cce, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.155715] env[62066]: DEBUG oslo_concurrency.lockutils [None req-594ce8c1-1ab1-41b6-bcf3-6a5287350f9c tempest-ServerShowV257Test-303939077 tempest-ServerShowV257Test-303939077-project-member] Lock "15afaea0-65de-4f32-851c-365003e5498f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.031s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 941.473191] env[62066]: DEBUG nova.objects.base [None req-77f7408f-8ed6-44d2-926f-666964de9ad4 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=62066) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 941.473191] env[62066]: DEBUG nova.network.neutron [None req-77f7408f-8ed6-44d2-926f-666964de9ad4 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: b718b09b-a7fa-445b-8be3-5b3ebca210a2] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 941.473191] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6655b798-907f-4fef-bf71-28a3c92c1c40 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.473191] env[62066]: DEBUG nova.network.neutron [req-ab6edaf1-c61d-402d-b4e0-f356d446c437 req-8c34b220-c481-4325-9788-3ec2196d9226 service nova] [instance: 5693eea2-1cac-45d5-aa27-58526b65136e] Updated VIF entry in instance network info cache for port 7e38fd56-d921-4784-9206-fdff79574fa4. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 941.473191] env[62066]: DEBUG nova.network.neutron [req-ab6edaf1-c61d-402d-b4e0-f356d446c437 req-8c34b220-c481-4325-9788-3ec2196d9226 service nova] [instance: 5693eea2-1cac-45d5-aa27-58526b65136e] Updating instance_info_cache with network_info: [{"id": "7e38fd56-d921-4784-9206-fdff79574fa4", "address": "fa:16:3e:84:3d:8e", "network": {"id": "7617f887-bdd0-400a-8b5f-606d7785df78", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-558766873-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8475ad5a900548cba568360999c846ea", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2b6a4065-12af-4fb9-ac47-ec9143f7297e", "external-id": "nsx-vlan-transportzone-95", "segmentation_id": 95, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7e38fd56-d9", "ovs_interfaceid": "7e38fd56-d921-4784-9206-fdff79574fa4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 941.473191] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1adb835a-fdb2-4ffa-880e-494bd54c6f6d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.473191] env[62066]: DEBUG nova.policy [None req-77f7408f-8ed6-44d2-926f-666964de9ad4 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '747026d272d4467bbf14a0c30589c3b7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0d4e33a99c7741fb8cdd97f4ec5dbbd8', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 941.473191] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea926b83-b6ea-4052-8fe7-e19831354d60 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.473191] env[62066]: DEBUG oslo_vmware.api [None req-d267d329-25e8-414c-9540-a1c71897bff8 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': task-1341185, 'name': Rename_Task} progress is 99%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.473191] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17d6ec4c-2dbc-406a-9eb1-df0a0c260885 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.479714] env[62066]: DEBUG nova.compute.provider_tree [None req-b37dd465-c64a-41d9-9bff-6087484093ac tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 941.498581] env[62066]: DEBUG oslo_vmware.api [None req-54702e84-4f3e-4046-b83c-40cd5c6f41f7 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': task-1341187, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.636812] env[62066]: DEBUG oslo_vmware.api [None req-bb369cbc-f881-4f95-b841-517845c5cb59 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52ca4591-596c-8079-c68d-70b8bf690cce, 'name': SearchDatastore_Task, 'duration_secs': 0.008444} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.637161] env[62066]: DEBUG oslo_concurrency.lockutils [None req-bb369cbc-f881-4f95-b841-517845c5cb59 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Releasing lock "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 941.637457] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-bb369cbc-f881-4f95-b841-517845c5cb59 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: 5693eea2-1cac-45d5-aa27-58526b65136e] Processing image 50ff584c-3b50-4395-af07-3e66769bc9f7 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 941.637642] env[62066]: DEBUG oslo_concurrency.lockutils [None req-bb369cbc-f881-4f95-b841-517845c5cb59 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 941.637792] env[62066]: DEBUG oslo_concurrency.lockutils [None req-bb369cbc-f881-4f95-b841-517845c5cb59 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Acquired lock "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 941.637974] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-bb369cbc-f881-4f95-b841-517845c5cb59 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 941.638271] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b5da5c1b-11fd-42f6-a32c-7f584c85daa2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.646226] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-bb369cbc-f881-4f95-b841-517845c5cb59 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 941.646410] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-bb369cbc-f881-4f95-b841-517845c5cb59 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 941.647116] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6cf111e0-b5ae-41a6-9ef5-b76841e65d09 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.652583] env[62066]: DEBUG oslo_vmware.api [None req-bb369cbc-f881-4f95-b841-517845c5cb59 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Waiting for the task: (returnval){ [ 941.652583] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]52e9f1f6-3c38-790a-d2a5-14e8e56d6e0c" [ 941.652583] env[62066]: _type = "Task" [ 941.652583] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.660463] env[62066]: DEBUG oslo_vmware.api [None req-bb369cbc-f881-4f95-b841-517845c5cb59 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52e9f1f6-3c38-790a-d2a5-14e8e56d6e0c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.736070] env[62066]: DEBUG nova.network.neutron [None req-77f7408f-8ed6-44d2-926f-666964de9ad4 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: b718b09b-a7fa-445b-8be3-5b3ebca210a2] Successfully created port: 26c5cd4b-0e0d-4e99-a7fe-7a46764dc642 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 941.898020] env[62066]: DEBUG oslo_concurrency.lockutils [req-ab6edaf1-c61d-402d-b4e0-f356d446c437 req-8c34b220-c481-4325-9788-3ec2196d9226 service nova] Releasing lock "refresh_cache-5693eea2-1cac-45d5-aa27-58526b65136e" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 941.955773] env[62066]: DEBUG oslo_vmware.api [None req-d267d329-25e8-414c-9540-a1c71897bff8 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': task-1341185, 'name': Rename_Task, 'duration_secs': 1.170437} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.956143] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-d267d329-25e8-414c-9540-a1c71897bff8 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 20987cb6-e12c-48c5-8fae-d990b1bb6b8a] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 941.956371] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-cc0203cc-f0ad-406a-acd0-981c6eb0b869 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.966297] env[62066]: DEBUG oslo_vmware.api [None req-d267d329-25e8-414c-9540-a1c71897bff8 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Waiting for the task: (returnval){ [ 941.966297] env[62066]: value = "task-1341188" [ 941.966297] env[62066]: _type = "Task" [ 941.966297] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.976179] env[62066]: DEBUG oslo_vmware.api [None req-d267d329-25e8-414c-9540-a1c71897bff8 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': task-1341188, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.983370] env[62066]: DEBUG nova.scheduler.client.report [None req-b37dd465-c64a-41d9-9bff-6087484093ac tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 942.001767] env[62066]: DEBUG oslo_vmware.api [None req-54702e84-4f3e-4046-b83c-40cd5c6f41f7 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': task-1341187, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.163652] env[62066]: DEBUG oslo_vmware.api [None req-bb369cbc-f881-4f95-b841-517845c5cb59 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52e9f1f6-3c38-790a-d2a5-14e8e56d6e0c, 'name': SearchDatastore_Task, 'duration_secs': 0.040319} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.164498] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c1ba5671-9bf8-4f56-805a-8312d4c84e6f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.169688] env[62066]: DEBUG oslo_vmware.api [None req-bb369cbc-f881-4f95-b841-517845c5cb59 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Waiting for the task: (returnval){ [ 942.169688] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]52cdd73b-c494-1a65-0546-00bd9208361b" [ 942.169688] env[62066]: _type = "Task" [ 942.169688] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.177681] env[62066]: DEBUG oslo_vmware.api [None req-bb369cbc-f881-4f95-b841-517845c5cb59 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52cdd73b-c494-1a65-0546-00bd9208361b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.476133] env[62066]: DEBUG oslo_vmware.api [None req-d267d329-25e8-414c-9540-a1c71897bff8 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': task-1341188, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.500439] env[62066]: DEBUG oslo_vmware.api [None req-54702e84-4f3e-4046-b83c-40cd5c6f41f7 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': task-1341187, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.194175} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.500439] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-54702e84-4f3e-4046-b83c-40cd5c6f41f7 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk to [datastore1] 651e5142-7ab0-4040-97b4-4599bd4e419a/651e5142-7ab0-4040-97b4-4599bd4e419a.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 942.500714] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-54702e84-4f3e-4046-b83c-40cd5c6f41f7 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 651e5142-7ab0-4040-97b4-4599bd4e419a] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 942.500846] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-07140791-5cb3-41ad-a4f6-10998e138d2b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.507929] env[62066]: DEBUG oslo_vmware.api [None req-54702e84-4f3e-4046-b83c-40cd5c6f41f7 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Waiting for the task: (returnval){ [ 942.507929] env[62066]: value = "task-1341189" [ 942.507929] env[62066]: _type = "Task" [ 942.507929] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.515333] env[62066]: DEBUG oslo_vmware.api [None req-54702e84-4f3e-4046-b83c-40cd5c6f41f7 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': task-1341189, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.680068] env[62066]: DEBUG oslo_vmware.api [None req-bb369cbc-f881-4f95-b841-517845c5cb59 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52cdd73b-c494-1a65-0546-00bd9208361b, 'name': SearchDatastore_Task, 'duration_secs': 0.013472} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.680374] env[62066]: DEBUG oslo_concurrency.lockutils [None req-bb369cbc-f881-4f95-b841-517845c5cb59 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Releasing lock "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 942.680641] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-bb369cbc-f881-4f95-b841-517845c5cb59 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk to [datastore1] 5693eea2-1cac-45d5-aa27-58526b65136e/5693eea2-1cac-45d5-aa27-58526b65136e.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 942.680897] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-eb3d93ab-6629-431b-af33-666dc361720b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.687542] env[62066]: DEBUG oslo_vmware.api [None req-bb369cbc-f881-4f95-b841-517845c5cb59 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Waiting for the task: (returnval){ [ 942.687542] env[62066]: value = "task-1341190" [ 942.687542] env[62066]: _type = "Task" [ 942.687542] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.697587] env[62066]: DEBUG oslo_vmware.api [None req-bb369cbc-f881-4f95-b841-517845c5cb59 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341190, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.976983] env[62066]: DEBUG oslo_vmware.api [None req-d267d329-25e8-414c-9540-a1c71897bff8 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': task-1341188, 'name': PowerOnVM_Task, 'duration_secs': 0.733109} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.977357] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-d267d329-25e8-414c-9540-a1c71897bff8 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 20987cb6-e12c-48c5-8fae-d990b1bb6b8a] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 942.977480] env[62066]: INFO nova.compute.manager [None req-d267d329-25e8-414c-9540-a1c71897bff8 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 20987cb6-e12c-48c5-8fae-d990b1bb6b8a] Took 9.49 seconds to spawn the instance on the hypervisor. [ 942.977665] env[62066]: DEBUG nova.compute.manager [None req-d267d329-25e8-414c-9540-a1c71897bff8 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 20987cb6-e12c-48c5-8fae-d990b1bb6b8a] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 942.978430] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e5d1cca-68fa-4c98-9f8f-ba26d81d4c8c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.996834] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b37dd465-c64a-41d9-9bff-6087484093ac tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.369s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 943.016988] env[62066]: DEBUG oslo_vmware.api [None req-54702e84-4f3e-4046-b83c-40cd5c6f41f7 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': task-1341189, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.059263} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.017268] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-54702e84-4f3e-4046-b83c-40cd5c6f41f7 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 651e5142-7ab0-4040-97b4-4599bd4e419a] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 943.018018] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-089cce4b-cc3e-4e26-9ab1-95de8c84930d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.041085] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-54702e84-4f3e-4046-b83c-40cd5c6f41f7 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 651e5142-7ab0-4040-97b4-4599bd4e419a] Reconfiguring VM instance instance-0000005d to attach disk [datastore1] 651e5142-7ab0-4040-97b4-4599bd4e419a/651e5142-7ab0-4040-97b4-4599bd4e419a.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 943.042076] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3f5a22d0-deaa-4a6f-81fd-e0e38ed12a3e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.061337] env[62066]: DEBUG oslo_vmware.api [None req-54702e84-4f3e-4046-b83c-40cd5c6f41f7 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Waiting for the task: (returnval){ [ 943.061337] env[62066]: value = "task-1341191" [ 943.061337] env[62066]: _type = "Task" [ 943.061337] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.069352] env[62066]: DEBUG oslo_vmware.api [None req-54702e84-4f3e-4046-b83c-40cd5c6f41f7 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': task-1341191, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.156878] env[62066]: DEBUG nova.compute.manager [req-b14aa611-ed9d-4218-a1a7-f8abbf2dbd6a req-348795b6-888b-4e6d-935f-174511aa4175 service nova] [instance: b718b09b-a7fa-445b-8be3-5b3ebca210a2] Received event network-vif-plugged-26c5cd4b-0e0d-4e99-a7fe-7a46764dc642 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 943.157088] env[62066]: DEBUG oslo_concurrency.lockutils [req-b14aa611-ed9d-4218-a1a7-f8abbf2dbd6a req-348795b6-888b-4e6d-935f-174511aa4175 service nova] Acquiring lock "b718b09b-a7fa-445b-8be3-5b3ebca210a2-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 943.157299] env[62066]: DEBUG oslo_concurrency.lockutils [req-b14aa611-ed9d-4218-a1a7-f8abbf2dbd6a req-348795b6-888b-4e6d-935f-174511aa4175 service nova] Lock "b718b09b-a7fa-445b-8be3-5b3ebca210a2-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 943.157475] env[62066]: DEBUG oslo_concurrency.lockutils [req-b14aa611-ed9d-4218-a1a7-f8abbf2dbd6a req-348795b6-888b-4e6d-935f-174511aa4175 service nova] Lock "b718b09b-a7fa-445b-8be3-5b3ebca210a2-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 943.157699] env[62066]: DEBUG nova.compute.manager [req-b14aa611-ed9d-4218-a1a7-f8abbf2dbd6a req-348795b6-888b-4e6d-935f-174511aa4175 service nova] [instance: b718b09b-a7fa-445b-8be3-5b3ebca210a2] No waiting events found dispatching network-vif-plugged-26c5cd4b-0e0d-4e99-a7fe-7a46764dc642 {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 943.157917] env[62066]: WARNING nova.compute.manager [req-b14aa611-ed9d-4218-a1a7-f8abbf2dbd6a req-348795b6-888b-4e6d-935f-174511aa4175 service nova] [instance: b718b09b-a7fa-445b-8be3-5b3ebca210a2] Received unexpected event network-vif-plugged-26c5cd4b-0e0d-4e99-a7fe-7a46764dc642 for instance with vm_state active and task_state None. [ 943.197598] env[62066]: DEBUG oslo_vmware.api [None req-bb369cbc-f881-4f95-b841-517845c5cb59 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341190, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.248131] env[62066]: DEBUG nova.network.neutron [None req-77f7408f-8ed6-44d2-926f-666964de9ad4 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: b718b09b-a7fa-445b-8be3-5b3ebca210a2] Successfully updated port: 26c5cd4b-0e0d-4e99-a7fe-7a46764dc642 {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 943.497867] env[62066]: INFO nova.compute.manager [None req-d267d329-25e8-414c-9540-a1c71897bff8 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 20987cb6-e12c-48c5-8fae-d990b1bb6b8a] Took 23.55 seconds to build instance. [ 943.570415] env[62066]: INFO nova.scheduler.client.report [None req-b37dd465-c64a-41d9-9bff-6087484093ac tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Deleted allocation for migration aba80e34-5809-409a-9dcc-c487ec0ebe3f [ 943.575720] env[62066]: DEBUG oslo_vmware.api [None req-54702e84-4f3e-4046-b83c-40cd5c6f41f7 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': task-1341191, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.699386] env[62066]: DEBUG oslo_vmware.api [None req-bb369cbc-f881-4f95-b841-517845c5cb59 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341190, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.750878] env[62066]: DEBUG oslo_concurrency.lockutils [None req-77f7408f-8ed6-44d2-926f-666964de9ad4 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Acquiring lock "refresh_cache-b718b09b-a7fa-445b-8be3-5b3ebca210a2" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 943.751098] env[62066]: DEBUG oslo_concurrency.lockutils [None req-77f7408f-8ed6-44d2-926f-666964de9ad4 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Acquired lock "refresh_cache-b718b09b-a7fa-445b-8be3-5b3ebca210a2" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 943.751232] env[62066]: DEBUG nova.network.neutron [None req-77f7408f-8ed6-44d2-926f-666964de9ad4 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: b718b09b-a7fa-445b-8be3-5b3ebca210a2] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 944.000338] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d267d329-25e8-414c-9540-a1c71897bff8 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Lock "20987cb6-e12c-48c5-8fae-d990b1bb6b8a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 25.066s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 944.072180] env[62066]: DEBUG oslo_vmware.api [None req-54702e84-4f3e-4046-b83c-40cd5c6f41f7 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': task-1341191, 'name': ReconfigVM_Task, 'duration_secs': 0.911873} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.072504] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-54702e84-4f3e-4046-b83c-40cd5c6f41f7 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 651e5142-7ab0-4040-97b4-4599bd4e419a] Reconfigured VM instance instance-0000005d to attach disk [datastore1] 651e5142-7ab0-4040-97b4-4599bd4e419a/651e5142-7ab0-4040-97b4-4599bd4e419a.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 944.073233] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d10f8851-6a49-4299-81ff-3fe8f8742506 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.079739] env[62066]: DEBUG oslo_vmware.api [None req-54702e84-4f3e-4046-b83c-40cd5c6f41f7 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Waiting for the task: (returnval){ [ 944.079739] env[62066]: value = "task-1341192" [ 944.079739] env[62066]: _type = "Task" [ 944.079739] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.080847] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b37dd465-c64a-41d9-9bff-6087484093ac tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Lock "9c989152-ec69-478e-a0b0-62983852576c" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 7.291s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 944.092153] env[62066]: DEBUG oslo_vmware.api [None req-54702e84-4f3e-4046-b83c-40cd5c6f41f7 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': task-1341192, 'name': Rename_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.199606] env[62066]: DEBUG oslo_vmware.api [None req-bb369cbc-f881-4f95-b841-517845c5cb59 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341190, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.14904} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.199911] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-bb369cbc-f881-4f95-b841-517845c5cb59 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk to [datastore1] 5693eea2-1cac-45d5-aa27-58526b65136e/5693eea2-1cac-45d5-aa27-58526b65136e.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 944.200247] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-bb369cbc-f881-4f95-b841-517845c5cb59 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: 5693eea2-1cac-45d5-aa27-58526b65136e] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 944.200804] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-40c5346b-8ed0-4cb8-9c54-a46e84514619 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.206666] env[62066]: DEBUG oslo_vmware.api [None req-bb369cbc-f881-4f95-b841-517845c5cb59 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Waiting for the task: (returnval){ [ 944.206666] env[62066]: value = "task-1341193" [ 944.206666] env[62066]: _type = "Task" [ 944.206666] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.214549] env[62066]: DEBUG oslo_vmware.api [None req-bb369cbc-f881-4f95-b841-517845c5cb59 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341193, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.289441] env[62066]: WARNING nova.network.neutron [None req-77f7408f-8ed6-44d2-926f-666964de9ad4 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: b718b09b-a7fa-445b-8be3-5b3ebca210a2] 449dfe4e-bb49-43b6-9ba7-b57af74ebfb3 already exists in list: networks containing: ['449dfe4e-bb49-43b6-9ba7-b57af74ebfb3']. ignoring it [ 944.566252] env[62066]: DEBUG nova.network.neutron [None req-77f7408f-8ed6-44d2-926f-666964de9ad4 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: b718b09b-a7fa-445b-8be3-5b3ebca210a2] Updating instance_info_cache with network_info: [{"id": "3d01745e-0d9e-4bae-ba4d-242380edbab6", "address": "fa:16:3e:ed:53:4c", "network": {"id": "449dfe4e-bb49-43b6-9ba7-b57af74ebfb3", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-282029676-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.224", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0d4e33a99c7741fb8cdd97f4ec5dbbd8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db68bd64-5b56-49af-a075-13dcf85cb2e0", "external-id": "nsx-vlan-transportzone-590", "segmentation_id": 590, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3d01745e-0d", "ovs_interfaceid": "3d01745e-0d9e-4bae-ba4d-242380edbab6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "26c5cd4b-0e0d-4e99-a7fe-7a46764dc642", "address": "fa:16:3e:ed:67:f7", "network": {"id": "449dfe4e-bb49-43b6-9ba7-b57af74ebfb3", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-282029676-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0d4e33a99c7741fb8cdd97f4ec5dbbd8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db68bd64-5b56-49af-a075-13dcf85cb2e0", "external-id": "nsx-vlan-transportzone-590", "segmentation_id": 590, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap26c5cd4b-0e", "ovs_interfaceid": "26c5cd4b-0e0d-4e99-a7fe-7a46764dc642", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 944.591609] env[62066]: DEBUG oslo_vmware.api [None req-54702e84-4f3e-4046-b83c-40cd5c6f41f7 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': task-1341192, 'name': Rename_Task, 'duration_secs': 0.146564} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.591887] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-54702e84-4f3e-4046-b83c-40cd5c6f41f7 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 651e5142-7ab0-4040-97b4-4599bd4e419a] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 944.592734] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-bd8c88ec-1d19-4c53-9f6a-434a4e34b129 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.598603] env[62066]: DEBUG oslo_vmware.api [None req-54702e84-4f3e-4046-b83c-40cd5c6f41f7 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Waiting for the task: (returnval){ [ 944.598603] env[62066]: value = "task-1341194" [ 944.598603] env[62066]: _type = "Task" [ 944.598603] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.606476] env[62066]: DEBUG oslo_vmware.api [None req-54702e84-4f3e-4046-b83c-40cd5c6f41f7 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': task-1341194, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.716886] env[62066]: DEBUG oslo_vmware.api [None req-bb369cbc-f881-4f95-b841-517845c5cb59 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341193, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06687} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.717409] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-bb369cbc-f881-4f95-b841-517845c5cb59 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: 5693eea2-1cac-45d5-aa27-58526b65136e] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 944.717986] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68384676-668a-471e-96ca-0632b63f7dca {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.740863] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-bb369cbc-f881-4f95-b841-517845c5cb59 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: 5693eea2-1cac-45d5-aa27-58526b65136e] Reconfiguring VM instance instance-0000005e to attach disk [datastore1] 5693eea2-1cac-45d5-aa27-58526b65136e/5693eea2-1cac-45d5-aa27-58526b65136e.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 944.741621] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-466316dc-154f-4c16-a33a-3bc7a5502d8a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.763025] env[62066]: DEBUG oslo_vmware.api [None req-bb369cbc-f881-4f95-b841-517845c5cb59 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Waiting for the task: (returnval){ [ 944.763025] env[62066]: value = "task-1341195" [ 944.763025] env[62066]: _type = "Task" [ 944.763025] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.770174] env[62066]: DEBUG oslo_vmware.api [None req-bb369cbc-f881-4f95-b841-517845c5cb59 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341195, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.868188] env[62066]: DEBUG oslo_concurrency.lockutils [None req-43f2b4bd-4e37-439b-bcb4-879202e419df tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Acquiring lock "20987cb6-e12c-48c5-8fae-d990b1bb6b8a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 944.868423] env[62066]: DEBUG oslo_concurrency.lockutils [None req-43f2b4bd-4e37-439b-bcb4-879202e419df tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Lock "20987cb6-e12c-48c5-8fae-d990b1bb6b8a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 944.868669] env[62066]: DEBUG oslo_concurrency.lockutils [None req-43f2b4bd-4e37-439b-bcb4-879202e419df tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Acquiring lock "20987cb6-e12c-48c5-8fae-d990b1bb6b8a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 944.868893] env[62066]: DEBUG oslo_concurrency.lockutils [None req-43f2b4bd-4e37-439b-bcb4-879202e419df tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Lock "20987cb6-e12c-48c5-8fae-d990b1bb6b8a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 944.869093] env[62066]: DEBUG oslo_concurrency.lockutils [None req-43f2b4bd-4e37-439b-bcb4-879202e419df tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Lock "20987cb6-e12c-48c5-8fae-d990b1bb6b8a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 944.871514] env[62066]: INFO nova.compute.manager [None req-43f2b4bd-4e37-439b-bcb4-879202e419df tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 20987cb6-e12c-48c5-8fae-d990b1bb6b8a] Terminating instance [ 944.874052] env[62066]: DEBUG nova.compute.manager [None req-43f2b4bd-4e37-439b-bcb4-879202e419df tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 20987cb6-e12c-48c5-8fae-d990b1bb6b8a] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 944.874052] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-43f2b4bd-4e37-439b-bcb4-879202e419df tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 20987cb6-e12c-48c5-8fae-d990b1bb6b8a] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 944.874814] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7368d8b-7f53-4419-84b7-192e5ba3daf6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.884861] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-43f2b4bd-4e37-439b-bcb4-879202e419df tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 20987cb6-e12c-48c5-8fae-d990b1bb6b8a] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 944.885161] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5c13f7aa-9e60-4f15-bd6a-f18a869945a4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.891409] env[62066]: DEBUG oslo_vmware.api [None req-43f2b4bd-4e37-439b-bcb4-879202e419df tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Waiting for the task: (returnval){ [ 944.891409] env[62066]: value = "task-1341196" [ 944.891409] env[62066]: _type = "Task" [ 944.891409] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.899208] env[62066]: DEBUG oslo_vmware.api [None req-43f2b4bd-4e37-439b-bcb4-879202e419df tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': task-1341196, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.069496] env[62066]: DEBUG oslo_concurrency.lockutils [None req-77f7408f-8ed6-44d2-926f-666964de9ad4 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Releasing lock "refresh_cache-b718b09b-a7fa-445b-8be3-5b3ebca210a2" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 945.070427] env[62066]: DEBUG oslo_concurrency.lockutils [None req-77f7408f-8ed6-44d2-926f-666964de9ad4 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Acquiring lock "b718b09b-a7fa-445b-8be3-5b3ebca210a2" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 945.070692] env[62066]: DEBUG oslo_concurrency.lockutils [None req-77f7408f-8ed6-44d2-926f-666964de9ad4 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Acquired lock "b718b09b-a7fa-445b-8be3-5b3ebca210a2" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 945.072084] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31d516a1-ca48-4dc7-b785-33ce3131dedf {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.089685] env[62066]: DEBUG nova.virt.hardware [None req-77f7408f-8ed6-44d2-926f-666964de9ad4 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-23T13:40:41Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 945.089947] env[62066]: DEBUG nova.virt.hardware [None req-77f7408f-8ed6-44d2-926f-666964de9ad4 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 945.090181] env[62066]: DEBUG nova.virt.hardware [None req-77f7408f-8ed6-44d2-926f-666964de9ad4 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 945.090435] env[62066]: DEBUG nova.virt.hardware [None req-77f7408f-8ed6-44d2-926f-666964de9ad4 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 945.090671] env[62066]: DEBUG nova.virt.hardware [None req-77f7408f-8ed6-44d2-926f-666964de9ad4 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 945.090752] env[62066]: DEBUG nova.virt.hardware [None req-77f7408f-8ed6-44d2-926f-666964de9ad4 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 945.090963] env[62066]: DEBUG nova.virt.hardware [None req-77f7408f-8ed6-44d2-926f-666964de9ad4 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 945.091146] env[62066]: DEBUG nova.virt.hardware [None req-77f7408f-8ed6-44d2-926f-666964de9ad4 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 945.091344] env[62066]: DEBUG nova.virt.hardware [None req-77f7408f-8ed6-44d2-926f-666964de9ad4 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 945.091533] env[62066]: DEBUG nova.virt.hardware [None req-77f7408f-8ed6-44d2-926f-666964de9ad4 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 945.091798] env[62066]: DEBUG nova.virt.hardware [None req-77f7408f-8ed6-44d2-926f-666964de9ad4 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 945.098273] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-77f7408f-8ed6-44d2-926f-666964de9ad4 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: b718b09b-a7fa-445b-8be3-5b3ebca210a2] Reconfiguring VM to attach interface {{(pid=62066) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 945.098682] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4ab1972f-f4e4-475b-ba5f-17f2ef04d06e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.126424] env[62066]: DEBUG oslo_vmware.api [None req-54702e84-4f3e-4046-b83c-40cd5c6f41f7 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': task-1341194, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.127955] env[62066]: DEBUG oslo_vmware.api [None req-77f7408f-8ed6-44d2-926f-666964de9ad4 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Waiting for the task: (returnval){ [ 945.127955] env[62066]: value = "task-1341197" [ 945.127955] env[62066]: _type = "Task" [ 945.127955] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.136506] env[62066]: DEBUG oslo_vmware.api [None req-77f7408f-8ed6-44d2-926f-666964de9ad4 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': task-1341197, 'name': ReconfigVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.205248] env[62066]: DEBUG oslo_concurrency.lockutils [None req-91ab760c-87a2-479d-a9f4-ed9e891980c3 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Acquiring lock "9c989152-ec69-478e-a0b0-62983852576c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 945.205658] env[62066]: DEBUG oslo_concurrency.lockutils [None req-91ab760c-87a2-479d-a9f4-ed9e891980c3 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Lock "9c989152-ec69-478e-a0b0-62983852576c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 945.206174] env[62066]: DEBUG oslo_concurrency.lockutils [None req-91ab760c-87a2-479d-a9f4-ed9e891980c3 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Acquiring lock "9c989152-ec69-478e-a0b0-62983852576c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 945.206391] env[62066]: DEBUG oslo_concurrency.lockutils [None req-91ab760c-87a2-479d-a9f4-ed9e891980c3 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Lock "9c989152-ec69-478e-a0b0-62983852576c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 945.206495] env[62066]: DEBUG oslo_concurrency.lockutils [None req-91ab760c-87a2-479d-a9f4-ed9e891980c3 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Lock "9c989152-ec69-478e-a0b0-62983852576c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 945.208896] env[62066]: INFO nova.compute.manager [None req-91ab760c-87a2-479d-a9f4-ed9e891980c3 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 9c989152-ec69-478e-a0b0-62983852576c] Terminating instance [ 945.210846] env[62066]: DEBUG nova.compute.manager [None req-91ab760c-87a2-479d-a9f4-ed9e891980c3 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 9c989152-ec69-478e-a0b0-62983852576c] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 945.211017] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-91ab760c-87a2-479d-a9f4-ed9e891980c3 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 9c989152-ec69-478e-a0b0-62983852576c] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 945.211900] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40395b73-148e-47bb-8466-14a49eb2e779 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.219318] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-91ab760c-87a2-479d-a9f4-ed9e891980c3 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 9c989152-ec69-478e-a0b0-62983852576c] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 945.219572] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8fa1d624-eaab-461c-ac51-b183094adede {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.225945] env[62066]: DEBUG oslo_vmware.api [None req-91ab760c-87a2-479d-a9f4-ed9e891980c3 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Waiting for the task: (returnval){ [ 945.225945] env[62066]: value = "task-1341198" [ 945.225945] env[62066]: _type = "Task" [ 945.225945] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.234482] env[62066]: DEBUG oslo_vmware.api [None req-91ab760c-87a2-479d-a9f4-ed9e891980c3 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1341198, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.245525] env[62066]: DEBUG nova.compute.manager [req-3686639f-0a7f-4545-a2e0-d5d4877abed2 req-d60c890f-c977-4ab0-8000-f229857c905d service nova] [instance: b718b09b-a7fa-445b-8be3-5b3ebca210a2] Received event network-changed-26c5cd4b-0e0d-4e99-a7fe-7a46764dc642 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 945.245707] env[62066]: DEBUG nova.compute.manager [req-3686639f-0a7f-4545-a2e0-d5d4877abed2 req-d60c890f-c977-4ab0-8000-f229857c905d service nova] [instance: b718b09b-a7fa-445b-8be3-5b3ebca210a2] Refreshing instance network info cache due to event network-changed-26c5cd4b-0e0d-4e99-a7fe-7a46764dc642. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 945.245889] env[62066]: DEBUG oslo_concurrency.lockutils [req-3686639f-0a7f-4545-a2e0-d5d4877abed2 req-d60c890f-c977-4ab0-8000-f229857c905d service nova] Acquiring lock "refresh_cache-b718b09b-a7fa-445b-8be3-5b3ebca210a2" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 945.246067] env[62066]: DEBUG oslo_concurrency.lockutils [req-3686639f-0a7f-4545-a2e0-d5d4877abed2 req-d60c890f-c977-4ab0-8000-f229857c905d service nova] Acquired lock "refresh_cache-b718b09b-a7fa-445b-8be3-5b3ebca210a2" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 945.246241] env[62066]: DEBUG nova.network.neutron [req-3686639f-0a7f-4545-a2e0-d5d4877abed2 req-d60c890f-c977-4ab0-8000-f229857c905d service nova] [instance: b718b09b-a7fa-445b-8be3-5b3ebca210a2] Refreshing network info cache for port 26c5cd4b-0e0d-4e99-a7fe-7a46764dc642 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 945.274324] env[62066]: DEBUG oslo_vmware.api [None req-bb369cbc-f881-4f95-b841-517845c5cb59 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341195, 'name': ReconfigVM_Task, 'duration_secs': 0.274863} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.274644] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-bb369cbc-f881-4f95-b841-517845c5cb59 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: 5693eea2-1cac-45d5-aa27-58526b65136e] Reconfigured VM instance instance-0000005e to attach disk [datastore1] 5693eea2-1cac-45d5-aa27-58526b65136e/5693eea2-1cac-45d5-aa27-58526b65136e.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 945.275340] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-41d78ab5-3c3d-430e-ba22-ee885d81deb6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.281950] env[62066]: DEBUG oslo_vmware.api [None req-bb369cbc-f881-4f95-b841-517845c5cb59 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Waiting for the task: (returnval){ [ 945.281950] env[62066]: value = "task-1341199" [ 945.281950] env[62066]: _type = "Task" [ 945.281950] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.290111] env[62066]: DEBUG oslo_vmware.api [None req-bb369cbc-f881-4f95-b841-517845c5cb59 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341199, 'name': Rename_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.401698] env[62066]: DEBUG oslo_vmware.api [None req-43f2b4bd-4e37-439b-bcb4-879202e419df tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': task-1341196, 'name': PowerOffVM_Task, 'duration_secs': 0.216834} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.401988] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-43f2b4bd-4e37-439b-bcb4-879202e419df tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 20987cb6-e12c-48c5-8fae-d990b1bb6b8a] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 945.402308] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-43f2b4bd-4e37-439b-bcb4-879202e419df tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 20987cb6-e12c-48c5-8fae-d990b1bb6b8a] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 945.402630] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-54bcc2fc-2891-4919-a943-f63167d535c5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.504297] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-43f2b4bd-4e37-439b-bcb4-879202e419df tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 20987cb6-e12c-48c5-8fae-d990b1bb6b8a] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 945.504536] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-43f2b4bd-4e37-439b-bcb4-879202e419df tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 20987cb6-e12c-48c5-8fae-d990b1bb6b8a] Deleting contents of the VM from datastore datastore1 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 945.504727] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-43f2b4bd-4e37-439b-bcb4-879202e419df tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Deleting the datastore file [datastore1] 20987cb6-e12c-48c5-8fae-d990b1bb6b8a {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 945.505035] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a6ddb7e8-4afe-4219-b428-584bdbd0de3d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.511977] env[62066]: DEBUG oslo_vmware.api [None req-43f2b4bd-4e37-439b-bcb4-879202e419df tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Waiting for the task: (returnval){ [ 945.511977] env[62066]: value = "task-1341201" [ 945.511977] env[62066]: _type = "Task" [ 945.511977] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.520262] env[62066]: DEBUG oslo_vmware.api [None req-43f2b4bd-4e37-439b-bcb4-879202e419df tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': task-1341201, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.621679] env[62066]: DEBUG oslo_vmware.api [None req-54702e84-4f3e-4046-b83c-40cd5c6f41f7 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': task-1341194, 'name': PowerOnVM_Task, 'duration_secs': 0.834263} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.622109] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-54702e84-4f3e-4046-b83c-40cd5c6f41f7 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 651e5142-7ab0-4040-97b4-4599bd4e419a] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 945.622487] env[62066]: INFO nova.compute.manager [None req-54702e84-4f3e-4046-b83c-40cd5c6f41f7 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 651e5142-7ab0-4040-97b4-4599bd4e419a] Took 9.74 seconds to spawn the instance on the hypervisor. [ 945.622822] env[62066]: DEBUG nova.compute.manager [None req-54702e84-4f3e-4046-b83c-40cd5c6f41f7 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 651e5142-7ab0-4040-97b4-4599bd4e419a] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 945.623828] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3dbd7834-8b84-422b-86ca-898737674825 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.640802] env[62066]: DEBUG oslo_vmware.api [None req-77f7408f-8ed6-44d2-926f-666964de9ad4 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': task-1341197, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.716457] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d41b2ff6-ff8d-407b-a347-8a3e15a584f9 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Acquiring lock "ca504be5-209c-456c-af00-d403bfa2d634" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 945.716705] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d41b2ff6-ff8d-407b-a347-8a3e15a584f9 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Lock "ca504be5-209c-456c-af00-d403bfa2d634" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 945.737361] env[62066]: DEBUG oslo_vmware.api [None req-91ab760c-87a2-479d-a9f4-ed9e891980c3 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1341198, 'name': PowerOffVM_Task, 'duration_secs': 0.275079} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.737361] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-91ab760c-87a2-479d-a9f4-ed9e891980c3 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 9c989152-ec69-478e-a0b0-62983852576c] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 945.737573] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-91ab760c-87a2-479d-a9f4-ed9e891980c3 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 9c989152-ec69-478e-a0b0-62983852576c] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 945.737706] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-80b13812-de1b-4c8d-a1f2-ff2ec6f8b96c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.793384] env[62066]: DEBUG oslo_vmware.api [None req-bb369cbc-f881-4f95-b841-517845c5cb59 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341199, 'name': Rename_Task, 'duration_secs': 0.155322} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.793710] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-bb369cbc-f881-4f95-b841-517845c5cb59 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: 5693eea2-1cac-45d5-aa27-58526b65136e] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 945.794071] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6fa952fb-a8ff-4a80-8d13-b6e1618aa6f3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.800768] env[62066]: DEBUG oslo_vmware.api [None req-bb369cbc-f881-4f95-b841-517845c5cb59 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Waiting for the task: (returnval){ [ 945.800768] env[62066]: value = "task-1341203" [ 945.800768] env[62066]: _type = "Task" [ 945.800768] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.808736] env[62066]: DEBUG oslo_vmware.api [None req-bb369cbc-f881-4f95-b841-517845c5cb59 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341203, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.813375] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-91ab760c-87a2-479d-a9f4-ed9e891980c3 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 9c989152-ec69-478e-a0b0-62983852576c] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 945.813692] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-91ab760c-87a2-479d-a9f4-ed9e891980c3 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 9c989152-ec69-478e-a0b0-62983852576c] Deleting contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 945.813925] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-91ab760c-87a2-479d-a9f4-ed9e891980c3 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Deleting the datastore file [datastore2] 9c989152-ec69-478e-a0b0-62983852576c {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 945.814290] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6bee066c-e497-45bb-8bbb-9ecb3e08df5b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.821370] env[62066]: DEBUG oslo_vmware.api [None req-91ab760c-87a2-479d-a9f4-ed9e891980c3 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Waiting for the task: (returnval){ [ 945.821370] env[62066]: value = "task-1341204" [ 945.821370] env[62066]: _type = "Task" [ 945.821370] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.829557] env[62066]: DEBUG oslo_vmware.api [None req-91ab760c-87a2-479d-a9f4-ed9e891980c3 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1341204, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.974182] env[62066]: DEBUG nova.network.neutron [req-3686639f-0a7f-4545-a2e0-d5d4877abed2 req-d60c890f-c977-4ab0-8000-f229857c905d service nova] [instance: b718b09b-a7fa-445b-8be3-5b3ebca210a2] Updated VIF entry in instance network info cache for port 26c5cd4b-0e0d-4e99-a7fe-7a46764dc642. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 945.974787] env[62066]: DEBUG nova.network.neutron [req-3686639f-0a7f-4545-a2e0-d5d4877abed2 req-d60c890f-c977-4ab0-8000-f229857c905d service nova] [instance: b718b09b-a7fa-445b-8be3-5b3ebca210a2] Updating instance_info_cache with network_info: [{"id": "3d01745e-0d9e-4bae-ba4d-242380edbab6", "address": "fa:16:3e:ed:53:4c", "network": {"id": "449dfe4e-bb49-43b6-9ba7-b57af74ebfb3", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-282029676-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.224", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0d4e33a99c7741fb8cdd97f4ec5dbbd8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db68bd64-5b56-49af-a075-13dcf85cb2e0", "external-id": "nsx-vlan-transportzone-590", "segmentation_id": 590, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3d01745e-0d", "ovs_interfaceid": "3d01745e-0d9e-4bae-ba4d-242380edbab6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "26c5cd4b-0e0d-4e99-a7fe-7a46764dc642", "address": "fa:16:3e:ed:67:f7", "network": {"id": "449dfe4e-bb49-43b6-9ba7-b57af74ebfb3", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-282029676-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0d4e33a99c7741fb8cdd97f4ec5dbbd8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db68bd64-5b56-49af-a075-13dcf85cb2e0", "external-id": "nsx-vlan-transportzone-590", "segmentation_id": 590, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap26c5cd4b-0e", "ovs_interfaceid": "26c5cd4b-0e0d-4e99-a7fe-7a46764dc642", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 946.022661] env[62066]: DEBUG oslo_vmware.api [None req-43f2b4bd-4e37-439b-bcb4-879202e419df tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': task-1341201, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.287966} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.023457] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-43f2b4bd-4e37-439b-bcb4-879202e419df tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 946.023848] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-43f2b4bd-4e37-439b-bcb4-879202e419df tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 20987cb6-e12c-48c5-8fae-d990b1bb6b8a] Deleted contents of the VM from datastore datastore1 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 946.024041] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-43f2b4bd-4e37-439b-bcb4-879202e419df tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 20987cb6-e12c-48c5-8fae-d990b1bb6b8a] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 946.024321] env[62066]: INFO nova.compute.manager [None req-43f2b4bd-4e37-439b-bcb4-879202e419df tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 20987cb6-e12c-48c5-8fae-d990b1bb6b8a] Took 1.15 seconds to destroy the instance on the hypervisor. [ 946.024575] env[62066]: DEBUG oslo.service.loopingcall [None req-43f2b4bd-4e37-439b-bcb4-879202e419df tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 946.024776] env[62066]: DEBUG nova.compute.manager [-] [instance: 20987cb6-e12c-48c5-8fae-d990b1bb6b8a] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 946.024871] env[62066]: DEBUG nova.network.neutron [-] [instance: 20987cb6-e12c-48c5-8fae-d990b1bb6b8a] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 946.138632] env[62066]: DEBUG oslo_vmware.api [None req-77f7408f-8ed6-44d2-926f-666964de9ad4 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': task-1341197, 'name': ReconfigVM_Task, 'duration_secs': 0.744066} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.139160] env[62066]: DEBUG oslo_concurrency.lockutils [None req-77f7408f-8ed6-44d2-926f-666964de9ad4 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Releasing lock "b718b09b-a7fa-445b-8be3-5b3ebca210a2" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 946.139377] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-77f7408f-8ed6-44d2-926f-666964de9ad4 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: b718b09b-a7fa-445b-8be3-5b3ebca210a2] Reconfigured VM to attach interface {{(pid=62066) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 946.153229] env[62066]: INFO nova.compute.manager [None req-54702e84-4f3e-4046-b83c-40cd5c6f41f7 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 651e5142-7ab0-4040-97b4-4599bd4e419a] Took 25.89 seconds to build instance. [ 946.219377] env[62066]: DEBUG nova.compute.manager [None req-d41b2ff6-ff8d-407b-a347-8a3e15a584f9 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: ca504be5-209c-456c-af00-d403bfa2d634] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 946.313258] env[62066]: DEBUG oslo_vmware.api [None req-bb369cbc-f881-4f95-b841-517845c5cb59 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341203, 'name': PowerOnVM_Task, 'duration_secs': 0.470746} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.313505] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-bb369cbc-f881-4f95-b841-517845c5cb59 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: 5693eea2-1cac-45d5-aa27-58526b65136e] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 946.313741] env[62066]: INFO nova.compute.manager [None req-bb369cbc-f881-4f95-b841-517845c5cb59 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: 5693eea2-1cac-45d5-aa27-58526b65136e] Took 8.11 seconds to spawn the instance on the hypervisor. [ 946.313930] env[62066]: DEBUG nova.compute.manager [None req-bb369cbc-f881-4f95-b841-517845c5cb59 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: 5693eea2-1cac-45d5-aa27-58526b65136e] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 946.314756] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff32afda-06fa-4c04-9076-cc538dfe9ec5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.332152] env[62066]: DEBUG oslo_vmware.api [None req-91ab760c-87a2-479d-a9f4-ed9e891980c3 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1341204, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.210763} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.332266] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-91ab760c-87a2-479d-a9f4-ed9e891980c3 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 946.332453] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-91ab760c-87a2-479d-a9f4-ed9e891980c3 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 9c989152-ec69-478e-a0b0-62983852576c] Deleted contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 946.333705] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-91ab760c-87a2-479d-a9f4-ed9e891980c3 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 9c989152-ec69-478e-a0b0-62983852576c] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 946.333705] env[62066]: INFO nova.compute.manager [None req-91ab760c-87a2-479d-a9f4-ed9e891980c3 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 9c989152-ec69-478e-a0b0-62983852576c] Took 1.12 seconds to destroy the instance on the hypervisor. [ 946.333705] env[62066]: DEBUG oslo.service.loopingcall [None req-91ab760c-87a2-479d-a9f4-ed9e891980c3 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 946.333705] env[62066]: DEBUG nova.compute.manager [-] [instance: 9c989152-ec69-478e-a0b0-62983852576c] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 946.333705] env[62066]: DEBUG nova.network.neutron [-] [instance: 9c989152-ec69-478e-a0b0-62983852576c] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 946.477827] env[62066]: DEBUG oslo_concurrency.lockutils [req-3686639f-0a7f-4545-a2e0-d5d4877abed2 req-d60c890f-c977-4ab0-8000-f229857c905d service nova] Releasing lock "refresh_cache-b718b09b-a7fa-445b-8be3-5b3ebca210a2" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 946.654672] env[62066]: DEBUG oslo_concurrency.lockutils [None req-77f7408f-8ed6-44d2-926f-666964de9ad4 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Lock "interface-b718b09b-a7fa-445b-8be3-5b3ebca210a2-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 6.378s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 946.657414] env[62066]: DEBUG oslo_concurrency.lockutils [None req-54702e84-4f3e-4046-b83c-40cd5c6f41f7 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Lock "651e5142-7ab0-4040-97b4-4599bd4e419a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 27.397s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 946.743644] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d41b2ff6-ff8d-407b-a347-8a3e15a584f9 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 946.743917] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d41b2ff6-ff8d-407b-a347-8a3e15a584f9 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 946.745514] env[62066]: INFO nova.compute.claims [None req-d41b2ff6-ff8d-407b-a347-8a3e15a584f9 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: ca504be5-209c-456c-af00-d403bfa2d634] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 946.778563] env[62066]: DEBUG nova.network.neutron [-] [instance: 20987cb6-e12c-48c5-8fae-d990b1bb6b8a] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 946.833976] env[62066]: INFO nova.compute.manager [None req-bb369cbc-f881-4f95-b841-517845c5cb59 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: 5693eea2-1cac-45d5-aa27-58526b65136e] Took 20.39 seconds to build instance. [ 947.054196] env[62066]: DEBUG nova.network.neutron [-] [instance: 9c989152-ec69-478e-a0b0-62983852576c] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 947.159545] env[62066]: DEBUG nova.compute.manager [None req-ca9c32b8-6257-4e95-8675-f49503dbe681 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 651e5142-7ab0-4040-97b4-4599bd4e419a] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 947.161436] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d2d2ba6-12f7-4c05-b090-d537e68a4a86 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.273115] env[62066]: DEBUG nova.compute.manager [req-286f4b6a-e648-4807-afed-074d258bae42 req-933c3352-5ab6-4005-9052-bc3d13fb2558 service nova] [instance: 20987cb6-e12c-48c5-8fae-d990b1bb6b8a] Received event network-vif-deleted-f24a3aad-eb9f-4bd9-aa9d-e7cc51421b37 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 947.273391] env[62066]: DEBUG nova.compute.manager [req-286f4b6a-e648-4807-afed-074d258bae42 req-933c3352-5ab6-4005-9052-bc3d13fb2558 service nova] [instance: 9c989152-ec69-478e-a0b0-62983852576c] Received event network-vif-deleted-1c0d0075-7f22-441e-a283-21e9b5c5aecc {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 947.281409] env[62066]: INFO nova.compute.manager [-] [instance: 20987cb6-e12c-48c5-8fae-d990b1bb6b8a] Took 1.26 seconds to deallocate network for instance. [ 947.336008] env[62066]: DEBUG oslo_concurrency.lockutils [None req-bb369cbc-f881-4f95-b841-517845c5cb59 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Lock "5693eea2-1cac-45d5-aa27-58526b65136e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.898s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 947.556534] env[62066]: INFO nova.compute.manager [-] [instance: 9c989152-ec69-478e-a0b0-62983852576c] Took 1.22 seconds to deallocate network for instance. [ 947.676908] env[62066]: INFO nova.compute.manager [None req-ca9c32b8-6257-4e95-8675-f49503dbe681 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 651e5142-7ab0-4040-97b4-4599bd4e419a] instance snapshotting [ 947.676908] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5c53486-2554-4c82-a234-fee1265585af {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.699543] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e95baec-0a4c-46f2-a551-b6f3e6a72956 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.788605] env[62066]: DEBUG oslo_concurrency.lockutils [None req-43f2b4bd-4e37-439b-bcb4-879202e419df tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 947.913677] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd7a00a3-d979-434b-8a0b-941b25942bea {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.921576] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acd2d24c-bc8a-4681-b306-8aef38fd6864 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.950834] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-331ba17e-b4a4-40ac-92e0-02cbea19b070 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.958144] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47151cb5-8a48-498e-9b12-7b70797754a6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.971049] env[62066]: DEBUG nova.compute.provider_tree [None req-d41b2ff6-ff8d-407b-a347-8a3e15a584f9 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 948.065489] env[62066]: DEBUG oslo_concurrency.lockutils [None req-91ab760c-87a2-479d-a9f4-ed9e891980c3 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 948.210026] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-ca9c32b8-6257-4e95-8675-f49503dbe681 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 651e5142-7ab0-4040-97b4-4599bd4e419a] Creating Snapshot of the VM instance {{(pid=62066) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 948.210457] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-edcc500f-6476-416d-8cae-fda06fa312cc {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.218636] env[62066]: DEBUG oslo_vmware.api [None req-ca9c32b8-6257-4e95-8675-f49503dbe681 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Waiting for the task: (returnval){ [ 948.218636] env[62066]: value = "task-1341205" [ 948.218636] env[62066]: _type = "Task" [ 948.218636] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.228497] env[62066]: DEBUG oslo_vmware.api [None req-ca9c32b8-6257-4e95-8675-f49503dbe681 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': task-1341205, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.297823] env[62066]: DEBUG oslo_concurrency.lockutils [None req-239d0b06-f99e-40e5-8e60-8ebe96ea1723 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Acquiring lock "interface-b718b09b-a7fa-445b-8be3-5b3ebca210a2-c9952f47-6c37-48e3-ad4e-0c2cc1c5bdb0" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 948.298188] env[62066]: DEBUG oslo_concurrency.lockutils [None req-239d0b06-f99e-40e5-8e60-8ebe96ea1723 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Lock "interface-b718b09b-a7fa-445b-8be3-5b3ebca210a2-c9952f47-6c37-48e3-ad4e-0c2cc1c5bdb0" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 948.298616] env[62066]: DEBUG nova.objects.instance [None req-239d0b06-f99e-40e5-8e60-8ebe96ea1723 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Lazy-loading 'flavor' on Instance uuid b718b09b-a7fa-445b-8be3-5b3ebca210a2 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 948.474853] env[62066]: DEBUG nova.scheduler.client.report [None req-d41b2ff6-ff8d-407b-a347-8a3e15a584f9 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 948.730320] env[62066]: DEBUG oslo_vmware.api [None req-ca9c32b8-6257-4e95-8675-f49503dbe681 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': task-1341205, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.768303] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c7b14697-dee9-4e9c-8ed4-e9c67e984eb5 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Acquiring lock "5693eea2-1cac-45d5-aa27-58526b65136e" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 948.768552] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c7b14697-dee9-4e9c-8ed4-e9c67e984eb5 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Lock "5693eea2-1cac-45d5-aa27-58526b65136e" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 948.914811] env[62066]: DEBUG nova.objects.instance [None req-239d0b06-f99e-40e5-8e60-8ebe96ea1723 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Lazy-loading 'pci_requests' on Instance uuid b718b09b-a7fa-445b-8be3-5b3ebca210a2 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 948.980040] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d41b2ff6-ff8d-407b-a347-8a3e15a584f9 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.236s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 948.980752] env[62066]: DEBUG nova.compute.manager [None req-d41b2ff6-ff8d-407b-a347-8a3e15a584f9 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: ca504be5-209c-456c-af00-d403bfa2d634] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 948.984206] env[62066]: DEBUG oslo_concurrency.lockutils [None req-43f2b4bd-4e37-439b-bcb4-879202e419df tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.196s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 948.984591] env[62066]: DEBUG nova.objects.instance [None req-43f2b4bd-4e37-439b-bcb4-879202e419df tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Lazy-loading 'resources' on Instance uuid 20987cb6-e12c-48c5-8fae-d990b1bb6b8a {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 949.230028] env[62066]: DEBUG oslo_vmware.api [None req-ca9c32b8-6257-4e95-8675-f49503dbe681 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': task-1341205, 'name': CreateSnapshot_Task, 'duration_secs': 0.664376} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.230446] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-ca9c32b8-6257-4e95-8675-f49503dbe681 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 651e5142-7ab0-4040-97b4-4599bd4e419a] Created Snapshot of the VM instance {{(pid=62066) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 949.231121] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89903251-cefc-4c23-94e1-a0e9fb855b42 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.272126] env[62066]: DEBUG nova.compute.utils [None req-c7b14697-dee9-4e9c-8ed4-e9c67e984eb5 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 949.417411] env[62066]: DEBUG nova.objects.base [None req-239d0b06-f99e-40e5-8e60-8ebe96ea1723 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=62066) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 949.417665] env[62066]: DEBUG nova.network.neutron [None req-239d0b06-f99e-40e5-8e60-8ebe96ea1723 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: b718b09b-a7fa-445b-8be3-5b3ebca210a2] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 949.481981] env[62066]: DEBUG nova.policy [None req-239d0b06-f99e-40e5-8e60-8ebe96ea1723 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '747026d272d4467bbf14a0c30589c3b7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0d4e33a99c7741fb8cdd97f4ec5dbbd8', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 949.485451] env[62066]: DEBUG nova.compute.utils [None req-d41b2ff6-ff8d-407b-a347-8a3e15a584f9 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 949.486701] env[62066]: DEBUG nova.compute.manager [None req-d41b2ff6-ff8d-407b-a347-8a3e15a584f9 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: ca504be5-209c-456c-af00-d403bfa2d634] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 949.486901] env[62066]: DEBUG nova.network.neutron [None req-d41b2ff6-ff8d-407b-a347-8a3e15a584f9 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: ca504be5-209c-456c-af00-d403bfa2d634] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 949.527471] env[62066]: DEBUG nova.policy [None req-d41b2ff6-ff8d-407b-a347-8a3e15a584f9 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2ab20cbd0bab4ae7ba46d9135605a509', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd609babc987a42e2a8ddb4bfb9c3b103', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 949.635833] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e7b4ef7-e8a6-48ba-972f-32c2e2a3e2d1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.644774] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de234819-7d3a-4851-b60d-9cf05c788c99 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.683996] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99d21995-9dd6-49cb-9ce1-6b704bdad9d3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.692256] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a26d0b97-88c6-4360-a66e-edd23a233379 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.706012] env[62066]: DEBUG nova.compute.provider_tree [None req-43f2b4bd-4e37-439b-bcb4-879202e419df tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 949.749817] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-ca9c32b8-6257-4e95-8675-f49503dbe681 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 651e5142-7ab0-4040-97b4-4599bd4e419a] Creating linked-clone VM from snapshot {{(pid=62066) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 949.750563] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-39832e58-708c-4618-9edf-3cafa5ccb890 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.758834] env[62066]: DEBUG oslo_vmware.api [None req-ca9c32b8-6257-4e95-8675-f49503dbe681 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Waiting for the task: (returnval){ [ 949.758834] env[62066]: value = "task-1341206" [ 949.758834] env[62066]: _type = "Task" [ 949.758834] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.766879] env[62066]: DEBUG oslo_vmware.api [None req-ca9c32b8-6257-4e95-8675-f49503dbe681 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': task-1341206, 'name': CloneVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.774658] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c7b14697-dee9-4e9c-8ed4-e9c67e984eb5 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Lock "5693eea2-1cac-45d5-aa27-58526b65136e" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 949.780431] env[62066]: DEBUG nova.network.neutron [None req-d41b2ff6-ff8d-407b-a347-8a3e15a584f9 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: ca504be5-209c-456c-af00-d403bfa2d634] Successfully created port: 11deacd3-d3a5-49a4-98f3-3226eb532e76 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 949.992427] env[62066]: DEBUG nova.compute.manager [None req-d41b2ff6-ff8d-407b-a347-8a3e15a584f9 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: ca504be5-209c-456c-af00-d403bfa2d634] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 950.209958] env[62066]: DEBUG nova.scheduler.client.report [None req-43f2b4bd-4e37-439b-bcb4-879202e419df tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 950.269296] env[62066]: DEBUG oslo_vmware.api [None req-ca9c32b8-6257-4e95-8675-f49503dbe681 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': task-1341206, 'name': CloneVM_Task} progress is 94%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.714801] env[62066]: DEBUG oslo_concurrency.lockutils [None req-43f2b4bd-4e37-439b-bcb4-879202e419df tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.730s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 950.716415] env[62066]: DEBUG oslo_concurrency.lockutils [None req-91ab760c-87a2-479d-a9f4-ed9e891980c3 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.651s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 950.716653] env[62066]: DEBUG oslo_concurrency.lockutils [None req-91ab760c-87a2-479d-a9f4-ed9e891980c3 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 950.736137] env[62066]: INFO nova.scheduler.client.report [None req-91ab760c-87a2-479d-a9f4-ed9e891980c3 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Deleted allocations for instance 9c989152-ec69-478e-a0b0-62983852576c [ 950.738236] env[62066]: INFO nova.scheduler.client.report [None req-43f2b4bd-4e37-439b-bcb4-879202e419df tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Deleted allocations for instance 20987cb6-e12c-48c5-8fae-d990b1bb6b8a [ 950.771579] env[62066]: DEBUG oslo_vmware.api [None req-ca9c32b8-6257-4e95-8675-f49503dbe681 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': task-1341206, 'name': CloneVM_Task} progress is 94%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.844216] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c7b14697-dee9-4e9c-8ed4-e9c67e984eb5 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Acquiring lock "5693eea2-1cac-45d5-aa27-58526b65136e" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 950.844489] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c7b14697-dee9-4e9c-8ed4-e9c67e984eb5 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Lock "5693eea2-1cac-45d5-aa27-58526b65136e" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 950.844957] env[62066]: INFO nova.compute.manager [None req-c7b14697-dee9-4e9c-8ed4-e9c67e984eb5 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: 5693eea2-1cac-45d5-aa27-58526b65136e] Attaching volume e50e4638-0cff-4a5c-b7f9-9f8f39406291 to /dev/sdb [ 950.882023] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5380bec2-fed2-4c88-a15b-c5c4ba510324 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.886582] env[62066]: DEBUG nova.compute.manager [req-0711e9e3-4709-40a6-8c37-e47bd0380fd0 req-a0abb100-5a21-495e-a513-f57c57d153ac service nova] [instance: b718b09b-a7fa-445b-8be3-5b3ebca210a2] Received event network-vif-plugged-c9952f47-6c37-48e3-ad4e-0c2cc1c5bdb0 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 950.887908] env[62066]: DEBUG oslo_concurrency.lockutils [req-0711e9e3-4709-40a6-8c37-e47bd0380fd0 req-a0abb100-5a21-495e-a513-f57c57d153ac service nova] Acquiring lock "b718b09b-a7fa-445b-8be3-5b3ebca210a2-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 950.887908] env[62066]: DEBUG oslo_concurrency.lockutils [req-0711e9e3-4709-40a6-8c37-e47bd0380fd0 req-a0abb100-5a21-495e-a513-f57c57d153ac service nova] Lock "b718b09b-a7fa-445b-8be3-5b3ebca210a2-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 950.887908] env[62066]: DEBUG oslo_concurrency.lockutils [req-0711e9e3-4709-40a6-8c37-e47bd0380fd0 req-a0abb100-5a21-495e-a513-f57c57d153ac service nova] Lock "b718b09b-a7fa-445b-8be3-5b3ebca210a2-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 950.887908] env[62066]: DEBUG nova.compute.manager [req-0711e9e3-4709-40a6-8c37-e47bd0380fd0 req-a0abb100-5a21-495e-a513-f57c57d153ac service nova] [instance: b718b09b-a7fa-445b-8be3-5b3ebca210a2] No waiting events found dispatching network-vif-plugged-c9952f47-6c37-48e3-ad4e-0c2cc1c5bdb0 {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 950.887908] env[62066]: WARNING nova.compute.manager [req-0711e9e3-4709-40a6-8c37-e47bd0380fd0 req-a0abb100-5a21-495e-a513-f57c57d153ac service nova] [instance: b718b09b-a7fa-445b-8be3-5b3ebca210a2] Received unexpected event network-vif-plugged-c9952f47-6c37-48e3-ad4e-0c2cc1c5bdb0 for instance with vm_state active and task_state None. [ 950.892846] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4213624a-4f50-405d-9484-b2822140f5a1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.906370] env[62066]: DEBUG nova.virt.block_device [None req-c7b14697-dee9-4e9c-8ed4-e9c67e984eb5 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: 5693eea2-1cac-45d5-aa27-58526b65136e] Updating existing volume attachment record: ec38892c-61de-4e8a-ba92-9adab43fc8b1 {{(pid=62066) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 950.982353] env[62066]: DEBUG nova.network.neutron [None req-239d0b06-f99e-40e5-8e60-8ebe96ea1723 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: b718b09b-a7fa-445b-8be3-5b3ebca210a2] Successfully updated port: c9952f47-6c37-48e3-ad4e-0c2cc1c5bdb0 {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 951.001274] env[62066]: DEBUG nova.compute.manager [None req-d41b2ff6-ff8d-407b-a347-8a3e15a584f9 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: ca504be5-209c-456c-af00-d403bfa2d634] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 951.029420] env[62066]: DEBUG nova.virt.hardware [None req-d41b2ff6-ff8d-407b-a347-8a3e15a584f9 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-23T13:40:41Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-23T13:40:23Z,direct_url=,disk_format='vmdk',id=50ff584c-3b50-4395-af07-3e66769bc9f7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='eb52f7069a374c61ae946f052007c6d9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-23T13:40:24Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 951.029775] env[62066]: DEBUG nova.virt.hardware [None req-d41b2ff6-ff8d-407b-a347-8a3e15a584f9 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 951.029988] env[62066]: DEBUG nova.virt.hardware [None req-d41b2ff6-ff8d-407b-a347-8a3e15a584f9 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 951.030270] env[62066]: DEBUG nova.virt.hardware [None req-d41b2ff6-ff8d-407b-a347-8a3e15a584f9 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 951.030451] env[62066]: DEBUG nova.virt.hardware [None req-d41b2ff6-ff8d-407b-a347-8a3e15a584f9 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 951.030667] env[62066]: DEBUG nova.virt.hardware [None req-d41b2ff6-ff8d-407b-a347-8a3e15a584f9 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 951.030950] env[62066]: DEBUG nova.virt.hardware [None req-d41b2ff6-ff8d-407b-a347-8a3e15a584f9 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 951.031160] env[62066]: DEBUG nova.virt.hardware [None req-d41b2ff6-ff8d-407b-a347-8a3e15a584f9 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 951.031537] env[62066]: DEBUG nova.virt.hardware [None req-d41b2ff6-ff8d-407b-a347-8a3e15a584f9 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 951.031780] env[62066]: DEBUG nova.virt.hardware [None req-d41b2ff6-ff8d-407b-a347-8a3e15a584f9 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 951.031971] env[62066]: DEBUG nova.virt.hardware [None req-d41b2ff6-ff8d-407b-a347-8a3e15a584f9 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 951.033420] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b044cf3-6bb4-446f-80e7-08842cb8a9a2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.041277] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c32eb673-c287-48ba-94b5-6c736048b54e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.199157] env[62066]: DEBUG nova.compute.manager [req-0c088a15-0d34-4107-956f-6107d7603605 req-f1b1cb01-4de9-4219-ac1a-11fb6967843c service nova] [instance: ca504be5-209c-456c-af00-d403bfa2d634] Received event network-vif-plugged-11deacd3-d3a5-49a4-98f3-3226eb532e76 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 951.199382] env[62066]: DEBUG oslo_concurrency.lockutils [req-0c088a15-0d34-4107-956f-6107d7603605 req-f1b1cb01-4de9-4219-ac1a-11fb6967843c service nova] Acquiring lock "ca504be5-209c-456c-af00-d403bfa2d634-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 951.199746] env[62066]: DEBUG oslo_concurrency.lockutils [req-0c088a15-0d34-4107-956f-6107d7603605 req-f1b1cb01-4de9-4219-ac1a-11fb6967843c service nova] Lock "ca504be5-209c-456c-af00-d403bfa2d634-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 951.199794] env[62066]: DEBUG oslo_concurrency.lockutils [req-0c088a15-0d34-4107-956f-6107d7603605 req-f1b1cb01-4de9-4219-ac1a-11fb6967843c service nova] Lock "ca504be5-209c-456c-af00-d403bfa2d634-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 951.199927] env[62066]: DEBUG nova.compute.manager [req-0c088a15-0d34-4107-956f-6107d7603605 req-f1b1cb01-4de9-4219-ac1a-11fb6967843c service nova] [instance: ca504be5-209c-456c-af00-d403bfa2d634] No waiting events found dispatching network-vif-plugged-11deacd3-d3a5-49a4-98f3-3226eb532e76 {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 951.200098] env[62066]: WARNING nova.compute.manager [req-0c088a15-0d34-4107-956f-6107d7603605 req-f1b1cb01-4de9-4219-ac1a-11fb6967843c service nova] [instance: ca504be5-209c-456c-af00-d403bfa2d634] Received unexpected event network-vif-plugged-11deacd3-d3a5-49a4-98f3-3226eb532e76 for instance with vm_state building and task_state spawning. [ 951.249141] env[62066]: DEBUG oslo_concurrency.lockutils [None req-91ab760c-87a2-479d-a9f4-ed9e891980c3 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Lock "9c989152-ec69-478e-a0b0-62983852576c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.043s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 951.250234] env[62066]: DEBUG oslo_concurrency.lockutils [None req-43f2b4bd-4e37-439b-bcb4-879202e419df tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Lock "20987cb6-e12c-48c5-8fae-d990b1bb6b8a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.382s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 951.274467] env[62066]: DEBUG oslo_vmware.api [None req-ca9c32b8-6257-4e95-8675-f49503dbe681 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': task-1341206, 'name': CloneVM_Task} progress is 94%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.284377] env[62066]: DEBUG nova.network.neutron [None req-d41b2ff6-ff8d-407b-a347-8a3e15a584f9 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: ca504be5-209c-456c-af00-d403bfa2d634] Successfully updated port: 11deacd3-d3a5-49a4-98f3-3226eb532e76 {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 951.485150] env[62066]: DEBUG oslo_concurrency.lockutils [None req-239d0b06-f99e-40e5-8e60-8ebe96ea1723 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Acquiring lock "refresh_cache-b718b09b-a7fa-445b-8be3-5b3ebca210a2" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 951.485450] env[62066]: DEBUG oslo_concurrency.lockutils [None req-239d0b06-f99e-40e5-8e60-8ebe96ea1723 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Acquired lock "refresh_cache-b718b09b-a7fa-445b-8be3-5b3ebca210a2" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 951.485699] env[62066]: DEBUG nova.network.neutron [None req-239d0b06-f99e-40e5-8e60-8ebe96ea1723 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: b718b09b-a7fa-445b-8be3-5b3ebca210a2] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 951.774725] env[62066]: DEBUG oslo_vmware.api [None req-ca9c32b8-6257-4e95-8675-f49503dbe681 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': task-1341206, 'name': CloneVM_Task} progress is 100%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.786419] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d41b2ff6-ff8d-407b-a347-8a3e15a584f9 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Acquiring lock "refresh_cache-ca504be5-209c-456c-af00-d403bfa2d634" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 951.786596] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d41b2ff6-ff8d-407b-a347-8a3e15a584f9 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Acquired lock "refresh_cache-ca504be5-209c-456c-af00-d403bfa2d634" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 951.786729] env[62066]: DEBUG nova.network.neutron [None req-d41b2ff6-ff8d-407b-a347-8a3e15a584f9 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: ca504be5-209c-456c-af00-d403bfa2d634] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 952.023631] env[62066]: WARNING nova.network.neutron [None req-239d0b06-f99e-40e5-8e60-8ebe96ea1723 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: b718b09b-a7fa-445b-8be3-5b3ebca210a2] 449dfe4e-bb49-43b6-9ba7-b57af74ebfb3 already exists in list: networks containing: ['449dfe4e-bb49-43b6-9ba7-b57af74ebfb3']. ignoring it [ 952.023861] env[62066]: WARNING nova.network.neutron [None req-239d0b06-f99e-40e5-8e60-8ebe96ea1723 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: b718b09b-a7fa-445b-8be3-5b3ebca210a2] 449dfe4e-bb49-43b6-9ba7-b57af74ebfb3 already exists in list: networks containing: ['449dfe4e-bb49-43b6-9ba7-b57af74ebfb3']. ignoring it [ 952.132615] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f28185e3-e8c2-4352-b716-1069aa95f5fc tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Acquiring lock "4e39b9b0-4ab0-4f48-8a3f-6bc1895597eb" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 952.132884] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f28185e3-e8c2-4352-b716-1069aa95f5fc tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Lock "4e39b9b0-4ab0-4f48-8a3f-6bc1895597eb" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 952.133108] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f28185e3-e8c2-4352-b716-1069aa95f5fc tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Acquiring lock "4e39b9b0-4ab0-4f48-8a3f-6bc1895597eb-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 952.133305] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f28185e3-e8c2-4352-b716-1069aa95f5fc tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Lock "4e39b9b0-4ab0-4f48-8a3f-6bc1895597eb-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 952.133481] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f28185e3-e8c2-4352-b716-1069aa95f5fc tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Lock "4e39b9b0-4ab0-4f48-8a3f-6bc1895597eb-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 952.137524] env[62066]: INFO nova.compute.manager [None req-f28185e3-e8c2-4352-b716-1069aa95f5fc tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 4e39b9b0-4ab0-4f48-8a3f-6bc1895597eb] Terminating instance [ 952.139694] env[62066]: DEBUG nova.compute.manager [None req-f28185e3-e8c2-4352-b716-1069aa95f5fc tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 4e39b9b0-4ab0-4f48-8a3f-6bc1895597eb] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 952.139896] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-f28185e3-e8c2-4352-b716-1069aa95f5fc tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 4e39b9b0-4ab0-4f48-8a3f-6bc1895597eb] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 952.140757] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e201f20b-b751-4332-8717-391aad176f0a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.148840] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-f28185e3-e8c2-4352-b716-1069aa95f5fc tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 4e39b9b0-4ab0-4f48-8a3f-6bc1895597eb] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 952.148840] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c300a6f7-9af8-4022-9c08-c5dcd79a861a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.154871] env[62066]: DEBUG oslo_vmware.api [None req-f28185e3-e8c2-4352-b716-1069aa95f5fc tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Waiting for the task: (returnval){ [ 952.154871] env[62066]: value = "task-1341210" [ 952.154871] env[62066]: _type = "Task" [ 952.154871] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.162193] env[62066]: DEBUG oslo_vmware.api [None req-f28185e3-e8c2-4352-b716-1069aa95f5fc tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': task-1341210, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.272022] env[62066]: DEBUG oslo_vmware.api [None req-ca9c32b8-6257-4e95-8675-f49503dbe681 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': task-1341206, 'name': CloneVM_Task, 'duration_secs': 2.020051} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.272318] env[62066]: INFO nova.virt.vmwareapi.vmops [None req-ca9c32b8-6257-4e95-8675-f49503dbe681 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 651e5142-7ab0-4040-97b4-4599bd4e419a] Created linked-clone VM from snapshot [ 952.273052] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f74007fb-f223-4487-bc46-5ad14bea1a38 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.279689] env[62066]: DEBUG nova.virt.vmwareapi.images [None req-ca9c32b8-6257-4e95-8675-f49503dbe681 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 651e5142-7ab0-4040-97b4-4599bd4e419a] Uploading image 8544c972-199b-4087-8d05-4a253d6e0256 {{(pid=62066) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 952.290351] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca9c32b8-6257-4e95-8675-f49503dbe681 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 651e5142-7ab0-4040-97b4-4599bd4e419a] Destroying the VM {{(pid=62066) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 952.290831] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-5e1673ec-9757-46b9-91a7-3ecadd4032ad {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.296868] env[62066]: DEBUG oslo_vmware.api [None req-ca9c32b8-6257-4e95-8675-f49503dbe681 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Waiting for the task: (returnval){ [ 952.296868] env[62066]: value = "task-1341211" [ 952.296868] env[62066]: _type = "Task" [ 952.296868] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.304443] env[62066]: DEBUG oslo_vmware.api [None req-ca9c32b8-6257-4e95-8675-f49503dbe681 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': task-1341211, 'name': Destroy_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.319301] env[62066]: DEBUG nova.network.neutron [None req-d41b2ff6-ff8d-407b-a347-8a3e15a584f9 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: ca504be5-209c-456c-af00-d403bfa2d634] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 952.423026] env[62066]: DEBUG nova.network.neutron [None req-239d0b06-f99e-40e5-8e60-8ebe96ea1723 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: b718b09b-a7fa-445b-8be3-5b3ebca210a2] Updating instance_info_cache with network_info: [{"id": "3d01745e-0d9e-4bae-ba4d-242380edbab6", "address": "fa:16:3e:ed:53:4c", "network": {"id": "449dfe4e-bb49-43b6-9ba7-b57af74ebfb3", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-282029676-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.224", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0d4e33a99c7741fb8cdd97f4ec5dbbd8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db68bd64-5b56-49af-a075-13dcf85cb2e0", "external-id": "nsx-vlan-transportzone-590", "segmentation_id": 590, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3d01745e-0d", "ovs_interfaceid": "3d01745e-0d9e-4bae-ba4d-242380edbab6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "26c5cd4b-0e0d-4e99-a7fe-7a46764dc642", "address": "fa:16:3e:ed:67:f7", "network": {"id": "449dfe4e-bb49-43b6-9ba7-b57af74ebfb3", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-282029676-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0d4e33a99c7741fb8cdd97f4ec5dbbd8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db68bd64-5b56-49af-a075-13dcf85cb2e0", "external-id": "nsx-vlan-transportzone-590", "segmentation_id": 590, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap26c5cd4b-0e", "ovs_interfaceid": "26c5cd4b-0e0d-4e99-a7fe-7a46764dc642", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "c9952f47-6c37-48e3-ad4e-0c2cc1c5bdb0", "address": "fa:16:3e:d3:cb:83", "network": {"id": "449dfe4e-bb49-43b6-9ba7-b57af74ebfb3", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-282029676-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0d4e33a99c7741fb8cdd97f4ec5dbbd8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db68bd64-5b56-49af-a075-13dcf85cb2e0", "external-id": "nsx-vlan-transportzone-590", "segmentation_id": 590, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc9952f47-6c", "ovs_interfaceid": "c9952f47-6c37-48e3-ad4e-0c2cc1c5bdb0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 952.451217] env[62066]: DEBUG nova.network.neutron [None req-d41b2ff6-ff8d-407b-a347-8a3e15a584f9 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: ca504be5-209c-456c-af00-d403bfa2d634] Updating instance_info_cache with network_info: [{"id": "11deacd3-d3a5-49a4-98f3-3226eb532e76", "address": "fa:16:3e:3a:b5:74", "network": {"id": "1c6c51e4-96f8-4175-b7f5-96b8a42fcb13", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-2006089809-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d609babc987a42e2a8ddb4bfb9c3b103", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a965790c-2d2f-4c2a-9ee7-745f4d53039b", "external-id": "nsx-vlan-transportzone-708", "segmentation_id": 708, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap11deacd3-d3", "ovs_interfaceid": "11deacd3-d3a5-49a4-98f3-3226eb532e76", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 952.665664] env[62066]: DEBUG oslo_vmware.api [None req-f28185e3-e8c2-4352-b716-1069aa95f5fc tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': task-1341210, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.808386] env[62066]: DEBUG oslo_vmware.api [None req-ca9c32b8-6257-4e95-8675-f49503dbe681 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': task-1341211, 'name': Destroy_Task} progress is 100%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.926353] env[62066]: DEBUG oslo_concurrency.lockutils [None req-239d0b06-f99e-40e5-8e60-8ebe96ea1723 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Releasing lock "refresh_cache-b718b09b-a7fa-445b-8be3-5b3ebca210a2" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 952.927063] env[62066]: DEBUG oslo_concurrency.lockutils [None req-239d0b06-f99e-40e5-8e60-8ebe96ea1723 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Acquiring lock "b718b09b-a7fa-445b-8be3-5b3ebca210a2" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 952.927241] env[62066]: DEBUG oslo_concurrency.lockutils [None req-239d0b06-f99e-40e5-8e60-8ebe96ea1723 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Acquired lock "b718b09b-a7fa-445b-8be3-5b3ebca210a2" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 952.928136] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93ecc538-f6d8-4518-bbeb-9ecde8866000 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.945695] env[62066]: DEBUG nova.virt.hardware [None req-239d0b06-f99e-40e5-8e60-8ebe96ea1723 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-23T13:40:41Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 952.945941] env[62066]: DEBUG nova.virt.hardware [None req-239d0b06-f99e-40e5-8e60-8ebe96ea1723 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 952.946115] env[62066]: DEBUG nova.virt.hardware [None req-239d0b06-f99e-40e5-8e60-8ebe96ea1723 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 952.946311] env[62066]: DEBUG nova.virt.hardware [None req-239d0b06-f99e-40e5-8e60-8ebe96ea1723 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 952.946464] env[62066]: DEBUG nova.virt.hardware [None req-239d0b06-f99e-40e5-8e60-8ebe96ea1723 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 952.946615] env[62066]: DEBUG nova.virt.hardware [None req-239d0b06-f99e-40e5-8e60-8ebe96ea1723 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 952.946823] env[62066]: DEBUG nova.virt.hardware [None req-239d0b06-f99e-40e5-8e60-8ebe96ea1723 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 952.946986] env[62066]: DEBUG nova.virt.hardware [None req-239d0b06-f99e-40e5-8e60-8ebe96ea1723 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 952.947172] env[62066]: DEBUG nova.virt.hardware [None req-239d0b06-f99e-40e5-8e60-8ebe96ea1723 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 952.947341] env[62066]: DEBUG nova.virt.hardware [None req-239d0b06-f99e-40e5-8e60-8ebe96ea1723 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 952.947518] env[62066]: DEBUG nova.virt.hardware [None req-239d0b06-f99e-40e5-8e60-8ebe96ea1723 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 952.954194] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-239d0b06-f99e-40e5-8e60-8ebe96ea1723 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: b718b09b-a7fa-445b-8be3-5b3ebca210a2] Reconfiguring VM to attach interface {{(pid=62066) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 952.955225] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d41b2ff6-ff8d-407b-a347-8a3e15a584f9 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Releasing lock "refresh_cache-ca504be5-209c-456c-af00-d403bfa2d634" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 952.955495] env[62066]: DEBUG nova.compute.manager [None req-d41b2ff6-ff8d-407b-a347-8a3e15a584f9 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: ca504be5-209c-456c-af00-d403bfa2d634] Instance network_info: |[{"id": "11deacd3-d3a5-49a4-98f3-3226eb532e76", "address": "fa:16:3e:3a:b5:74", "network": {"id": "1c6c51e4-96f8-4175-b7f5-96b8a42fcb13", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-2006089809-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d609babc987a42e2a8ddb4bfb9c3b103", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a965790c-2d2f-4c2a-9ee7-745f4d53039b", "external-id": "nsx-vlan-transportzone-708", "segmentation_id": 708, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap11deacd3-d3", "ovs_interfaceid": "11deacd3-d3a5-49a4-98f3-3226eb532e76", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 952.955734] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a4b112ce-e22b-4ccf-b09c-9715855826b7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.968468] env[62066]: DEBUG nova.compute.manager [req-cdf8f96c-2620-4898-9d84-b77a2dbaba7b req-a3b6416f-6fb6-4454-9c94-9f87eb472640 service nova] [instance: b718b09b-a7fa-445b-8be3-5b3ebca210a2] Received event network-changed-c9952f47-6c37-48e3-ad4e-0c2cc1c5bdb0 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 952.968574] env[62066]: DEBUG nova.compute.manager [req-cdf8f96c-2620-4898-9d84-b77a2dbaba7b req-a3b6416f-6fb6-4454-9c94-9f87eb472640 service nova] [instance: b718b09b-a7fa-445b-8be3-5b3ebca210a2] Refreshing instance network info cache due to event network-changed-c9952f47-6c37-48e3-ad4e-0c2cc1c5bdb0. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 952.968783] env[62066]: DEBUG oslo_concurrency.lockutils [req-cdf8f96c-2620-4898-9d84-b77a2dbaba7b req-a3b6416f-6fb6-4454-9c94-9f87eb472640 service nova] Acquiring lock "refresh_cache-b718b09b-a7fa-445b-8be3-5b3ebca210a2" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 952.968985] env[62066]: DEBUG oslo_concurrency.lockutils [req-cdf8f96c-2620-4898-9d84-b77a2dbaba7b req-a3b6416f-6fb6-4454-9c94-9f87eb472640 service nova] Acquired lock "refresh_cache-b718b09b-a7fa-445b-8be3-5b3ebca210a2" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 952.969194] env[62066]: DEBUG nova.network.neutron [req-cdf8f96c-2620-4898-9d84-b77a2dbaba7b req-a3b6416f-6fb6-4454-9c94-9f87eb472640 service nova] [instance: b718b09b-a7fa-445b-8be3-5b3ebca210a2] Refreshing network info cache for port c9952f47-6c37-48e3-ad4e-0c2cc1c5bdb0 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 952.970465] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-d41b2ff6-ff8d-407b-a347-8a3e15a584f9 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: ca504be5-209c-456c-af00-d403bfa2d634] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:3a:b5:74', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a965790c-2d2f-4c2a-9ee7-745f4d53039b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '11deacd3-d3a5-49a4-98f3-3226eb532e76', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 952.977756] env[62066]: DEBUG oslo.service.loopingcall [None req-d41b2ff6-ff8d-407b-a347-8a3e15a584f9 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 952.980547] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ca504be5-209c-456c-af00-d403bfa2d634] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 952.981688] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-53da963f-5164-40dd-821b-a083db08a2cb {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.998389] env[62066]: DEBUG oslo_vmware.api [None req-239d0b06-f99e-40e5-8e60-8ebe96ea1723 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Waiting for the task: (returnval){ [ 952.998389] env[62066]: value = "task-1341212" [ 952.998389] env[62066]: _type = "Task" [ 952.998389] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.003640] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 953.003640] env[62066]: value = "task-1341213" [ 953.003640] env[62066]: _type = "Task" [ 953.003640] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.007041] env[62066]: DEBUG oslo_vmware.api [None req-239d0b06-f99e-40e5-8e60-8ebe96ea1723 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': task-1341212, 'name': ReconfigVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.013930] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1341213, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.166024] env[62066]: DEBUG oslo_vmware.api [None req-f28185e3-e8c2-4352-b716-1069aa95f5fc tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': task-1341210, 'name': PowerOffVM_Task, 'duration_secs': 0.656934} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.166303] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-f28185e3-e8c2-4352-b716-1069aa95f5fc tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 4e39b9b0-4ab0-4f48-8a3f-6bc1895597eb] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 953.166480] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-f28185e3-e8c2-4352-b716-1069aa95f5fc tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 4e39b9b0-4ab0-4f48-8a3f-6bc1895597eb] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 953.166738] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4911d779-33b5-4f3e-8af3-a101d14c416e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.206451] env[62066]: DEBUG nova.network.neutron [req-cdf8f96c-2620-4898-9d84-b77a2dbaba7b req-a3b6416f-6fb6-4454-9c94-9f87eb472640 service nova] [instance: b718b09b-a7fa-445b-8be3-5b3ebca210a2] Updated VIF entry in instance network info cache for port c9952f47-6c37-48e3-ad4e-0c2cc1c5bdb0. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 953.206942] env[62066]: DEBUG nova.network.neutron [req-cdf8f96c-2620-4898-9d84-b77a2dbaba7b req-a3b6416f-6fb6-4454-9c94-9f87eb472640 service nova] [instance: b718b09b-a7fa-445b-8be3-5b3ebca210a2] Updating instance_info_cache with network_info: [{"id": "3d01745e-0d9e-4bae-ba4d-242380edbab6", "address": "fa:16:3e:ed:53:4c", "network": {"id": "449dfe4e-bb49-43b6-9ba7-b57af74ebfb3", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-282029676-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.224", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0d4e33a99c7741fb8cdd97f4ec5dbbd8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db68bd64-5b56-49af-a075-13dcf85cb2e0", "external-id": "nsx-vlan-transportzone-590", "segmentation_id": 590, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3d01745e-0d", "ovs_interfaceid": "3d01745e-0d9e-4bae-ba4d-242380edbab6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "26c5cd4b-0e0d-4e99-a7fe-7a46764dc642", "address": "fa:16:3e:ed:67:f7", "network": {"id": "449dfe4e-bb49-43b6-9ba7-b57af74ebfb3", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-282029676-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0d4e33a99c7741fb8cdd97f4ec5dbbd8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db68bd64-5b56-49af-a075-13dcf85cb2e0", "external-id": "nsx-vlan-transportzone-590", "segmentation_id": 590, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap26c5cd4b-0e", "ovs_interfaceid": "26c5cd4b-0e0d-4e99-a7fe-7a46764dc642", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "c9952f47-6c37-48e3-ad4e-0c2cc1c5bdb0", "address": "fa:16:3e:d3:cb:83", "network": {"id": "449dfe4e-bb49-43b6-9ba7-b57af74ebfb3", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-282029676-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0d4e33a99c7741fb8cdd97f4ec5dbbd8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db68bd64-5b56-49af-a075-13dcf85cb2e0", "external-id": "nsx-vlan-transportzone-590", "segmentation_id": 590, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc9952f47-6c", "ovs_interfaceid": "c9952f47-6c37-48e3-ad4e-0c2cc1c5bdb0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 953.230506] env[62066]: DEBUG nova.compute.manager [req-052def76-7fd7-4551-aea7-02d8c1a48c18 req-dbf19da7-b185-4db3-909a-0267015d80e5 service nova] [instance: ca504be5-209c-456c-af00-d403bfa2d634] Received event network-changed-11deacd3-d3a5-49a4-98f3-3226eb532e76 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 953.230925] env[62066]: DEBUG nova.compute.manager [req-052def76-7fd7-4551-aea7-02d8c1a48c18 req-dbf19da7-b185-4db3-909a-0267015d80e5 service nova] [instance: ca504be5-209c-456c-af00-d403bfa2d634] Refreshing instance network info cache due to event network-changed-11deacd3-d3a5-49a4-98f3-3226eb532e76. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 953.230925] env[62066]: DEBUG oslo_concurrency.lockutils [req-052def76-7fd7-4551-aea7-02d8c1a48c18 req-dbf19da7-b185-4db3-909a-0267015d80e5 service nova] Acquiring lock "refresh_cache-ca504be5-209c-456c-af00-d403bfa2d634" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 953.231106] env[62066]: DEBUG oslo_concurrency.lockutils [req-052def76-7fd7-4551-aea7-02d8c1a48c18 req-dbf19da7-b185-4db3-909a-0267015d80e5 service nova] Acquired lock "refresh_cache-ca504be5-209c-456c-af00-d403bfa2d634" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 953.231244] env[62066]: DEBUG nova.network.neutron [req-052def76-7fd7-4551-aea7-02d8c1a48c18 req-dbf19da7-b185-4db3-909a-0267015d80e5 service nova] [instance: ca504be5-209c-456c-af00-d403bfa2d634] Refreshing network info cache for port 11deacd3-d3a5-49a4-98f3-3226eb532e76 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 953.237080] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-f28185e3-e8c2-4352-b716-1069aa95f5fc tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 4e39b9b0-4ab0-4f48-8a3f-6bc1895597eb] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 953.237238] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-f28185e3-e8c2-4352-b716-1069aa95f5fc tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 4e39b9b0-4ab0-4f48-8a3f-6bc1895597eb] Deleting contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 953.237536] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-f28185e3-e8c2-4352-b716-1069aa95f5fc tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Deleting the datastore file [datastore2] 4e39b9b0-4ab0-4f48-8a3f-6bc1895597eb {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 953.237866] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3a7bc4d5-2430-421e-94f7-df620bd3f06d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.244784] env[62066]: DEBUG oslo_vmware.api [None req-f28185e3-e8c2-4352-b716-1069aa95f5fc tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Waiting for the task: (returnval){ [ 953.244784] env[62066]: value = "task-1341216" [ 953.244784] env[62066]: _type = "Task" [ 953.244784] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.253433] env[62066]: DEBUG oslo_vmware.api [None req-f28185e3-e8c2-4352-b716-1069aa95f5fc tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': task-1341216, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.309170] env[62066]: DEBUG oslo_vmware.api [None req-ca9c32b8-6257-4e95-8675-f49503dbe681 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': task-1341211, 'name': Destroy_Task, 'duration_secs': 0.513059} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.309567] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-ca9c32b8-6257-4e95-8675-f49503dbe681 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 651e5142-7ab0-4040-97b4-4599bd4e419a] Destroyed the VM [ 953.309938] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-ca9c32b8-6257-4e95-8675-f49503dbe681 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 651e5142-7ab0-4040-97b4-4599bd4e419a] Deleting Snapshot of the VM instance {{(pid=62066) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 953.310263] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-a8ae5b17-bdcd-4826-9655-a32043c89a46 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.317345] env[62066]: DEBUG oslo_vmware.api [None req-ca9c32b8-6257-4e95-8675-f49503dbe681 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Waiting for the task: (returnval){ [ 953.317345] env[62066]: value = "task-1341217" [ 953.317345] env[62066]: _type = "Task" [ 953.317345] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.325723] env[62066]: DEBUG oslo_vmware.api [None req-ca9c32b8-6257-4e95-8675-f49503dbe681 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': task-1341217, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.509237] env[62066]: DEBUG oslo_vmware.api [None req-239d0b06-f99e-40e5-8e60-8ebe96ea1723 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': task-1341212, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.517300] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1341213, 'name': CreateVM_Task, 'duration_secs': 0.437013} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.517509] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ca504be5-209c-456c-af00-d403bfa2d634] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 953.518236] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d41b2ff6-ff8d-407b-a347-8a3e15a584f9 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 953.518484] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d41b2ff6-ff8d-407b-a347-8a3e15a584f9 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Acquired lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 953.518852] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d41b2ff6-ff8d-407b-a347-8a3e15a584f9 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 953.519165] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9446eb20-df3f-4e24-b08d-e8ed7d5e17b4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.523561] env[62066]: DEBUG oslo_vmware.api [None req-d41b2ff6-ff8d-407b-a347-8a3e15a584f9 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Waiting for the task: (returnval){ [ 953.523561] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]526b60a3-ec0a-8d54-b213-cea491db9d33" [ 953.523561] env[62066]: _type = "Task" [ 953.523561] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.530644] env[62066]: DEBUG oslo_vmware.api [None req-d41b2ff6-ff8d-407b-a347-8a3e15a584f9 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]526b60a3-ec0a-8d54-b213-cea491db9d33, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.710653] env[62066]: DEBUG oslo_concurrency.lockutils [req-cdf8f96c-2620-4898-9d84-b77a2dbaba7b req-a3b6416f-6fb6-4454-9c94-9f87eb472640 service nova] Releasing lock "refresh_cache-b718b09b-a7fa-445b-8be3-5b3ebca210a2" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 953.755755] env[62066]: DEBUG oslo_vmware.api [None req-f28185e3-e8c2-4352-b716-1069aa95f5fc tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': task-1341216, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.218587} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.756033] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-f28185e3-e8c2-4352-b716-1069aa95f5fc tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 953.756228] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-f28185e3-e8c2-4352-b716-1069aa95f5fc tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 4e39b9b0-4ab0-4f48-8a3f-6bc1895597eb] Deleted contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 953.756409] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-f28185e3-e8c2-4352-b716-1069aa95f5fc tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 4e39b9b0-4ab0-4f48-8a3f-6bc1895597eb] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 953.756587] env[62066]: INFO nova.compute.manager [None req-f28185e3-e8c2-4352-b716-1069aa95f5fc tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 4e39b9b0-4ab0-4f48-8a3f-6bc1895597eb] Took 1.62 seconds to destroy the instance on the hypervisor. [ 953.756835] env[62066]: DEBUG oslo.service.loopingcall [None req-f28185e3-e8c2-4352-b716-1069aa95f5fc tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 953.757045] env[62066]: DEBUG nova.compute.manager [-] [instance: 4e39b9b0-4ab0-4f48-8a3f-6bc1895597eb] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 953.757138] env[62066]: DEBUG nova.network.neutron [-] [instance: 4e39b9b0-4ab0-4f48-8a3f-6bc1895597eb] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 953.828749] env[62066]: DEBUG oslo_vmware.api [None req-ca9c32b8-6257-4e95-8675-f49503dbe681 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': task-1341217, 'name': RemoveSnapshot_Task} progress is 97%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.951573] env[62066]: DEBUG nova.network.neutron [req-052def76-7fd7-4551-aea7-02d8c1a48c18 req-dbf19da7-b185-4db3-909a-0267015d80e5 service nova] [instance: ca504be5-209c-456c-af00-d403bfa2d634] Updated VIF entry in instance network info cache for port 11deacd3-d3a5-49a4-98f3-3226eb532e76. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 953.951972] env[62066]: DEBUG nova.network.neutron [req-052def76-7fd7-4551-aea7-02d8c1a48c18 req-dbf19da7-b185-4db3-909a-0267015d80e5 service nova] [instance: ca504be5-209c-456c-af00-d403bfa2d634] Updating instance_info_cache with network_info: [{"id": "11deacd3-d3a5-49a4-98f3-3226eb532e76", "address": "fa:16:3e:3a:b5:74", "network": {"id": "1c6c51e4-96f8-4175-b7f5-96b8a42fcb13", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-2006089809-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d609babc987a42e2a8ddb4bfb9c3b103", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a965790c-2d2f-4c2a-9ee7-745f4d53039b", "external-id": "nsx-vlan-transportzone-708", "segmentation_id": 708, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap11deacd3-d3", "ovs_interfaceid": "11deacd3-d3a5-49a4-98f3-3226eb532e76", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 954.014379] env[62066]: DEBUG oslo_vmware.api [None req-239d0b06-f99e-40e5-8e60-8ebe96ea1723 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': task-1341212, 'name': ReconfigVM_Task, 'duration_secs': 0.943756} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.016025] env[62066]: DEBUG oslo_concurrency.lockutils [None req-239d0b06-f99e-40e5-8e60-8ebe96ea1723 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Releasing lock "b718b09b-a7fa-445b-8be3-5b3ebca210a2" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 954.016025] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-239d0b06-f99e-40e5-8e60-8ebe96ea1723 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: b718b09b-a7fa-445b-8be3-5b3ebca210a2] Reconfigured VM to attach interface {{(pid=62066) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 954.036545] env[62066]: DEBUG oslo_vmware.api [None req-d41b2ff6-ff8d-407b-a347-8a3e15a584f9 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]526b60a3-ec0a-8d54-b213-cea491db9d33, 'name': SearchDatastore_Task, 'duration_secs': 0.01214} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.037187] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d41b2ff6-ff8d-407b-a347-8a3e15a584f9 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Releasing lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 954.039018] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-d41b2ff6-ff8d-407b-a347-8a3e15a584f9 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: ca504be5-209c-456c-af00-d403bfa2d634] Processing image 50ff584c-3b50-4395-af07-3e66769bc9f7 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 954.039018] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d41b2ff6-ff8d-407b-a347-8a3e15a584f9 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 954.039018] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d41b2ff6-ff8d-407b-a347-8a3e15a584f9 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Acquired lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 954.039018] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-d41b2ff6-ff8d-407b-a347-8a3e15a584f9 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 954.039018] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-91fe3f4c-6d18-49cf-ab44-0faa8a103163 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.050015] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-d41b2ff6-ff8d-407b-a347-8a3e15a584f9 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 954.050015] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-d41b2ff6-ff8d-407b-a347-8a3e15a584f9 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 954.050015] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0d397e67-ef43-43c5-bc14-a93c7ce3edb8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.057230] env[62066]: DEBUG oslo_vmware.api [None req-d41b2ff6-ff8d-407b-a347-8a3e15a584f9 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Waiting for the task: (returnval){ [ 954.057230] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]52440208-6839-2d08-2feb-3ad4e850eb5b" [ 954.057230] env[62066]: _type = "Task" [ 954.057230] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.065765] env[62066]: DEBUG oslo_vmware.api [None req-d41b2ff6-ff8d-407b-a347-8a3e15a584f9 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52440208-6839-2d08-2feb-3ad4e850eb5b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.327759] env[62066]: DEBUG oslo_vmware.api [None req-ca9c32b8-6257-4e95-8675-f49503dbe681 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': task-1341217, 'name': RemoveSnapshot_Task, 'duration_secs': 0.802881} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.329028] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-ca9c32b8-6257-4e95-8675-f49503dbe681 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 651e5142-7ab0-4040-97b4-4599bd4e419a] Deleted Snapshot of the VM instance {{(pid=62066) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 954.366636] env[62066]: DEBUG oslo_service.periodic_task [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 954.366994] env[62066]: DEBUG oslo_service.periodic_task [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 954.455059] env[62066]: DEBUG oslo_concurrency.lockutils [req-052def76-7fd7-4551-aea7-02d8c1a48c18 req-dbf19da7-b185-4db3-909a-0267015d80e5 service nova] Releasing lock "refresh_cache-ca504be5-209c-456c-af00-d403bfa2d634" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 954.522176] env[62066]: DEBUG oslo_concurrency.lockutils [None req-239d0b06-f99e-40e5-8e60-8ebe96ea1723 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Lock "interface-b718b09b-a7fa-445b-8be3-5b3ebca210a2-c9952f47-6c37-48e3-ad4e-0c2cc1c5bdb0" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 6.224s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 954.565813] env[62066]: DEBUG oslo_vmware.api [None req-d41b2ff6-ff8d-407b-a347-8a3e15a584f9 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52440208-6839-2d08-2feb-3ad4e850eb5b, 'name': SearchDatastore_Task, 'duration_secs': 0.01472} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.566597] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-920f4479-3bf8-4a44-935f-d6f5a1fcd9d0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.571947] env[62066]: DEBUG oslo_vmware.api [None req-d41b2ff6-ff8d-407b-a347-8a3e15a584f9 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Waiting for the task: (returnval){ [ 954.571947] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]52534b4b-88ea-7b26-1bf2-9437080bec8b" [ 954.571947] env[62066]: _type = "Task" [ 954.571947] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.578939] env[62066]: DEBUG oslo_vmware.api [None req-d41b2ff6-ff8d-407b-a347-8a3e15a584f9 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52534b4b-88ea-7b26-1bf2-9437080bec8b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.602485] env[62066]: DEBUG nova.network.neutron [-] [instance: 4e39b9b0-4ab0-4f48-8a3f-6bc1895597eb] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 954.834225] env[62066]: WARNING nova.compute.manager [None req-ca9c32b8-6257-4e95-8675-f49503dbe681 tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 651e5142-7ab0-4040-97b4-4599bd4e419a] Image not found during snapshot: nova.exception.ImageNotFound: Image 8544c972-199b-4087-8d05-4a253d6e0256 could not be found. [ 954.875448] env[62066]: DEBUG oslo_service.periodic_task [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 954.875835] env[62066]: DEBUG nova.compute.manager [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Starting heal instance info cache {{(pid=62066) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 955.082494] env[62066]: DEBUG oslo_vmware.api [None req-d41b2ff6-ff8d-407b-a347-8a3e15a584f9 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52534b4b-88ea-7b26-1bf2-9437080bec8b, 'name': SearchDatastore_Task, 'duration_secs': 0.009624} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.082709] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d41b2ff6-ff8d-407b-a347-8a3e15a584f9 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Releasing lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 955.082972] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-d41b2ff6-ff8d-407b-a347-8a3e15a584f9 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk to [datastore2] ca504be5-209c-456c-af00-d403bfa2d634/ca504be5-209c-456c-af00-d403bfa2d634.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 955.083244] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6ecf12af-d7a9-465b-a99f-6201d09bf084 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.089498] env[62066]: DEBUG oslo_vmware.api [None req-d41b2ff6-ff8d-407b-a347-8a3e15a584f9 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Waiting for the task: (returnval){ [ 955.089498] env[62066]: value = "task-1341218" [ 955.089498] env[62066]: _type = "Task" [ 955.089498] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.096807] env[62066]: DEBUG oslo_vmware.api [None req-d41b2ff6-ff8d-407b-a347-8a3e15a584f9 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1341218, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.104361] env[62066]: INFO nova.compute.manager [-] [instance: 4e39b9b0-4ab0-4f48-8a3f-6bc1895597eb] Took 1.35 seconds to deallocate network for instance. [ 955.320011] env[62066]: DEBUG nova.compute.manager [req-fe1ad684-c79a-426b-bd95-ef5a6b4d33e8 req-40cad751-08fe-45a2-b912-ee293e3f63d4 service nova] [instance: 4e39b9b0-4ab0-4f48-8a3f-6bc1895597eb] Received event network-vif-deleted-843a9051-1a6f-42b2-b482-2baf7ffb6ca4 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 955.453020] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-c7b14697-dee9-4e9c-8ed4-e9c67e984eb5 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: 5693eea2-1cac-45d5-aa27-58526b65136e] Volume attach. Driver type: vmdk {{(pid=62066) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 955.453288] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-c7b14697-dee9-4e9c-8ed4-e9c67e984eb5 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: 5693eea2-1cac-45d5-aa27-58526b65136e] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-286114', 'volume_id': 'e50e4638-0cff-4a5c-b7f9-9f8f39406291', 'name': 'volume-e50e4638-0cff-4a5c-b7f9-9f8f39406291', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '5693eea2-1cac-45d5-aa27-58526b65136e', 'attached_at': '', 'detached_at': '', 'volume_id': 'e50e4638-0cff-4a5c-b7f9-9f8f39406291', 'serial': 'e50e4638-0cff-4a5c-b7f9-9f8f39406291'} {{(pid=62066) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 955.454205] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80186073-21f3-4b7f-b6ee-7164f4b06651 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.473484] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-462285de-55c5-40a4-9154-f08e58ccb668 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.501842] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-c7b14697-dee9-4e9c-8ed4-e9c67e984eb5 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: 5693eea2-1cac-45d5-aa27-58526b65136e] Reconfiguring VM instance instance-0000005e to attach disk [datastore2] volume-e50e4638-0cff-4a5c-b7f9-9f8f39406291/volume-e50e4638-0cff-4a5c-b7f9-9f8f39406291.vmdk or device None with type thin {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 955.502372] env[62066]: DEBUG oslo_concurrency.lockutils [None req-980eec8a-3d17-48a4-863b-a262428cc66c tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Acquiring lock "651e5142-7ab0-4040-97b4-4599bd4e419a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 955.502625] env[62066]: DEBUG oslo_concurrency.lockutils [None req-980eec8a-3d17-48a4-863b-a262428cc66c tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Lock "651e5142-7ab0-4040-97b4-4599bd4e419a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 955.502972] env[62066]: DEBUG oslo_concurrency.lockutils [None req-980eec8a-3d17-48a4-863b-a262428cc66c tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Acquiring lock "651e5142-7ab0-4040-97b4-4599bd4e419a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 955.503089] env[62066]: DEBUG oslo_concurrency.lockutils [None req-980eec8a-3d17-48a4-863b-a262428cc66c tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Lock "651e5142-7ab0-4040-97b4-4599bd4e419a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 955.503227] env[62066]: DEBUG oslo_concurrency.lockutils [None req-980eec8a-3d17-48a4-863b-a262428cc66c tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Lock "651e5142-7ab0-4040-97b4-4599bd4e419a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 955.504967] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ab2ef155-590b-4202-b5fd-e06f611b3b38 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.518526] env[62066]: INFO nova.compute.manager [None req-980eec8a-3d17-48a4-863b-a262428cc66c tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 651e5142-7ab0-4040-97b4-4599bd4e419a] Terminating instance [ 955.522120] env[62066]: DEBUG nova.compute.manager [None req-980eec8a-3d17-48a4-863b-a262428cc66c tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 651e5142-7ab0-4040-97b4-4599bd4e419a] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 955.522537] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-980eec8a-3d17-48a4-863b-a262428cc66c tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 651e5142-7ab0-4040-97b4-4599bd4e419a] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 955.523960] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47bac902-e813-46a2-ba83-98158620a25f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.531505] env[62066]: DEBUG oslo_vmware.api [None req-c7b14697-dee9-4e9c-8ed4-e9c67e984eb5 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Waiting for the task: (returnval){ [ 955.531505] env[62066]: value = "task-1341219" [ 955.531505] env[62066]: _type = "Task" [ 955.531505] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.537712] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-980eec8a-3d17-48a4-863b-a262428cc66c tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 651e5142-7ab0-4040-97b4-4599bd4e419a] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 955.538534] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a5f10847-4d91-42b2-8225-74b73706a71a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.546192] env[62066]: DEBUG oslo_vmware.api [None req-c7b14697-dee9-4e9c-8ed4-e9c67e984eb5 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341219, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.553606] env[62066]: DEBUG oslo_vmware.api [None req-980eec8a-3d17-48a4-863b-a262428cc66c tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Waiting for the task: (returnval){ [ 955.553606] env[62066]: value = "task-1341220" [ 955.553606] env[62066]: _type = "Task" [ 955.553606] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.562914] env[62066]: DEBUG oslo_vmware.api [None req-980eec8a-3d17-48a4-863b-a262428cc66c tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': task-1341220, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.601789] env[62066]: DEBUG oslo_vmware.api [None req-d41b2ff6-ff8d-407b-a347-8a3e15a584f9 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1341218, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.611035] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f28185e3-e8c2-4352-b716-1069aa95f5fc tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 955.611381] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f28185e3-e8c2-4352-b716-1069aa95f5fc tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 955.611713] env[62066]: DEBUG nova.objects.instance [None req-f28185e3-e8c2-4352-b716-1069aa95f5fc tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Lazy-loading 'resources' on Instance uuid 4e39b9b0-4ab0-4f48-8a3f-6bc1895597eb {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 956.041795] env[62066]: DEBUG oslo_vmware.api [None req-c7b14697-dee9-4e9c-8ed4-e9c67e984eb5 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341219, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.061600] env[62066]: DEBUG oslo_vmware.api [None req-980eec8a-3d17-48a4-863b-a262428cc66c tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': task-1341220, 'name': PowerOffVM_Task, 'duration_secs': 0.27557} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.061960] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-980eec8a-3d17-48a4-863b-a262428cc66c tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 651e5142-7ab0-4040-97b4-4599bd4e419a] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 956.062156] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-980eec8a-3d17-48a4-863b-a262428cc66c tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 651e5142-7ab0-4040-97b4-4599bd4e419a] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 956.062409] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-da05d6fa-b7cd-4250-beba-96371ff681d2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.102030] env[62066]: DEBUG oslo_vmware.api [None req-d41b2ff6-ff8d-407b-a347-8a3e15a584f9 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1341218, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.69152} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.102030] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-d41b2ff6-ff8d-407b-a347-8a3e15a584f9 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk to [datastore2] ca504be5-209c-456c-af00-d403bfa2d634/ca504be5-209c-456c-af00-d403bfa2d634.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 956.102030] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-d41b2ff6-ff8d-407b-a347-8a3e15a584f9 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: ca504be5-209c-456c-af00-d403bfa2d634] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 956.102030] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6c3c65fb-edb0-4988-b7d0-68c97da90648 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.105950] env[62066]: DEBUG oslo_vmware.api [None req-d41b2ff6-ff8d-407b-a347-8a3e15a584f9 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Waiting for the task: (returnval){ [ 956.105950] env[62066]: value = "task-1341222" [ 956.105950] env[62066]: _type = "Task" [ 956.105950] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.113825] env[62066]: DEBUG oslo_vmware.api [None req-d41b2ff6-ff8d-407b-a347-8a3e15a584f9 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1341222, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.125612] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-980eec8a-3d17-48a4-863b-a262428cc66c tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 651e5142-7ab0-4040-97b4-4599bd4e419a] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 956.125967] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-980eec8a-3d17-48a4-863b-a262428cc66c tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 651e5142-7ab0-4040-97b4-4599bd4e419a] Deleting contents of the VM from datastore datastore1 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 956.126275] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-980eec8a-3d17-48a4-863b-a262428cc66c tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Deleting the datastore file [datastore1] 651e5142-7ab0-4040-97b4-4599bd4e419a {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 956.126643] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c8038220-c9fb-4089-92eb-fd1b9e99590e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.133912] env[62066]: DEBUG oslo_vmware.api [None req-980eec8a-3d17-48a4-863b-a262428cc66c tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Waiting for the task: (returnval){ [ 956.133912] env[62066]: value = "task-1341223" [ 956.133912] env[62066]: _type = "Task" [ 956.133912] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.146109] env[62066]: DEBUG oslo_vmware.api [None req-980eec8a-3d17-48a4-863b-a262428cc66c tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': task-1341223, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.154720] env[62066]: DEBUG oslo_concurrency.lockutils [None req-96a3683c-7d82-48dc-bf96-f0467366c2eb tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Acquiring lock "interface-b718b09b-a7fa-445b-8be3-5b3ebca210a2-26c5cd4b-0e0d-4e99-a7fe-7a46764dc642" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 956.155098] env[62066]: DEBUG oslo_concurrency.lockutils [None req-96a3683c-7d82-48dc-bf96-f0467366c2eb tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Lock "interface-b718b09b-a7fa-445b-8be3-5b3ebca210a2-26c5cd4b-0e0d-4e99-a7fe-7a46764dc642" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 956.269834] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a665977d-86b0-4f16-9cae-369586917991 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.277089] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a3fb0bd-c79a-4a62-b6d9-b1dbf6d7ca4b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.307190] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a40383a9-2f50-47f4-ae30-53f78715133c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.314280] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc00c15c-0667-4417-a0cb-fe916e1c5f5b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.328380] env[62066]: DEBUG nova.compute.provider_tree [None req-f28185e3-e8c2-4352-b716-1069aa95f5fc tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 956.542943] env[62066]: DEBUG oslo_vmware.api [None req-c7b14697-dee9-4e9c-8ed4-e9c67e984eb5 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341219, 'name': ReconfigVM_Task, 'duration_secs': 0.671059} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.543336] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-c7b14697-dee9-4e9c-8ed4-e9c67e984eb5 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: 5693eea2-1cac-45d5-aa27-58526b65136e] Reconfigured VM instance instance-0000005e to attach disk [datastore2] volume-e50e4638-0cff-4a5c-b7f9-9f8f39406291/volume-e50e4638-0cff-4a5c-b7f9-9f8f39406291.vmdk or device None with type thin {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 956.547877] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5ef140c6-00d9-4c59-84f1-57fa8ce42d9f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.562376] env[62066]: DEBUG oslo_vmware.api [None req-c7b14697-dee9-4e9c-8ed4-e9c67e984eb5 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Waiting for the task: (returnval){ [ 956.562376] env[62066]: value = "task-1341224" [ 956.562376] env[62066]: _type = "Task" [ 956.562376] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.570047] env[62066]: DEBUG oslo_vmware.api [None req-c7b14697-dee9-4e9c-8ed4-e9c67e984eb5 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341224, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.614773] env[62066]: DEBUG oslo_vmware.api [None req-d41b2ff6-ff8d-407b-a347-8a3e15a584f9 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1341222, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072791} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.615061] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-d41b2ff6-ff8d-407b-a347-8a3e15a584f9 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: ca504be5-209c-456c-af00-d403bfa2d634] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 956.615834] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bbe7d01-3080-4cc3-92df-dc4381a26bd6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.637542] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-d41b2ff6-ff8d-407b-a347-8a3e15a584f9 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: ca504be5-209c-456c-af00-d403bfa2d634] Reconfiguring VM instance instance-0000005f to attach disk [datastore2] ca504be5-209c-456c-af00-d403bfa2d634/ca504be5-209c-456c-af00-d403bfa2d634.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 956.637861] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b2e65e9f-649f-4745-bfaa-84da5024f851 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.657932] env[62066]: DEBUG oslo_concurrency.lockutils [None req-96a3683c-7d82-48dc-bf96-f0467366c2eb tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Acquiring lock "b718b09b-a7fa-445b-8be3-5b3ebca210a2" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 956.658144] env[62066]: DEBUG oslo_concurrency.lockutils [None req-96a3683c-7d82-48dc-bf96-f0467366c2eb tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Acquired lock "b718b09b-a7fa-445b-8be3-5b3ebca210a2" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 956.658883] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b592b800-20ca-46fd-9262-18b44570f393 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.664974] env[62066]: DEBUG oslo_vmware.api [None req-980eec8a-3d17-48a4-863b-a262428cc66c tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Task: {'id': task-1341223, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.139446} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.665256] env[62066]: DEBUG oslo_vmware.api [None req-d41b2ff6-ff8d-407b-a347-8a3e15a584f9 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Waiting for the task: (returnval){ [ 956.665256] env[62066]: value = "task-1341225" [ 956.665256] env[62066]: _type = "Task" [ 956.665256] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.665859] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-980eec8a-3d17-48a4-863b-a262428cc66c tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 956.666111] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-980eec8a-3d17-48a4-863b-a262428cc66c tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 651e5142-7ab0-4040-97b4-4599bd4e419a] Deleted contents of the VM from datastore datastore1 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 956.666324] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-980eec8a-3d17-48a4-863b-a262428cc66c tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 651e5142-7ab0-4040-97b4-4599bd4e419a] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 956.666512] env[62066]: INFO nova.compute.manager [None req-980eec8a-3d17-48a4-863b-a262428cc66c tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] [instance: 651e5142-7ab0-4040-97b4-4599bd4e419a] Took 1.14 seconds to destroy the instance on the hypervisor. [ 956.666762] env[62066]: DEBUG oslo.service.loopingcall [None req-980eec8a-3d17-48a4-863b-a262428cc66c tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 956.667098] env[62066]: DEBUG nova.compute.manager [-] [instance: 651e5142-7ab0-4040-97b4-4599bd4e419a] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 956.667266] env[62066]: DEBUG nova.network.neutron [-] [instance: 651e5142-7ab0-4040-97b4-4599bd4e419a] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 956.685082] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a042c261-8bfa-4eb0-9160-c6e36e06a341 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.690712] env[62066]: DEBUG oslo_vmware.api [None req-d41b2ff6-ff8d-407b-a347-8a3e15a584f9 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1341225, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.718100] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-96a3683c-7d82-48dc-bf96-f0467366c2eb tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: b718b09b-a7fa-445b-8be3-5b3ebca210a2] Reconfiguring VM to detach interface {{(pid=62066) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 956.720479] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d003daf7-1e6e-42b7-9673-cc5f65e3a577 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.739482] env[62066]: DEBUG oslo_vmware.api [None req-96a3683c-7d82-48dc-bf96-f0467366c2eb tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Waiting for the task: (returnval){ [ 956.739482] env[62066]: value = "task-1341226" [ 956.739482] env[62066]: _type = "Task" [ 956.739482] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.747679] env[62066]: DEBUG oslo_vmware.api [None req-96a3683c-7d82-48dc-bf96-f0467366c2eb tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': task-1341226, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.832133] env[62066]: DEBUG nova.scheduler.client.report [None req-f28185e3-e8c2-4352-b716-1069aa95f5fc tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 957.072532] env[62066]: DEBUG oslo_vmware.api [None req-c7b14697-dee9-4e9c-8ed4-e9c67e984eb5 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341224, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.176062] env[62066]: DEBUG oslo_vmware.api [None req-d41b2ff6-ff8d-407b-a347-8a3e15a584f9 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1341225, 'name': ReconfigVM_Task, 'duration_secs': 0.275136} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.176334] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-d41b2ff6-ff8d-407b-a347-8a3e15a584f9 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: ca504be5-209c-456c-af00-d403bfa2d634] Reconfigured VM instance instance-0000005f to attach disk [datastore2] ca504be5-209c-456c-af00-d403bfa2d634/ca504be5-209c-456c-af00-d403bfa2d634.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 957.176977] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-26c24ccc-4794-40a4-9d4e-a91af720eb8c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.183778] env[62066]: DEBUG oslo_vmware.api [None req-d41b2ff6-ff8d-407b-a347-8a3e15a584f9 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Waiting for the task: (returnval){ [ 957.183778] env[62066]: value = "task-1341227" [ 957.183778] env[62066]: _type = "Task" [ 957.183778] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.191169] env[62066]: DEBUG oslo_vmware.api [None req-d41b2ff6-ff8d-407b-a347-8a3e15a584f9 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1341227, 'name': Rename_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.249376] env[62066]: DEBUG oslo_vmware.api [None req-96a3683c-7d82-48dc-bf96-f0467366c2eb tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': task-1341226, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.338532] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f28185e3-e8c2-4352-b716-1069aa95f5fc tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.727s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 957.357425] env[62066]: INFO nova.scheduler.client.report [None req-f28185e3-e8c2-4352-b716-1069aa95f5fc tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Deleted allocations for instance 4e39b9b0-4ab0-4f48-8a3f-6bc1895597eb [ 957.367667] env[62066]: DEBUG nova.compute.manager [req-6edfa8f6-8e3a-4096-a63d-ce199481ef59 req-fa86a79a-ac92-4a90-916f-dbb0b8b49ee0 service nova] [instance: 651e5142-7ab0-4040-97b4-4599bd4e419a] Received event network-vif-deleted-66054023-ce27-4e2d-92f5-18f1dde734d0 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 957.367925] env[62066]: INFO nova.compute.manager [req-6edfa8f6-8e3a-4096-a63d-ce199481ef59 req-fa86a79a-ac92-4a90-916f-dbb0b8b49ee0 service nova] [instance: 651e5142-7ab0-4040-97b4-4599bd4e419a] Neutron deleted interface 66054023-ce27-4e2d-92f5-18f1dde734d0; detaching it from the instance and deleting it from the info cache [ 957.368453] env[62066]: DEBUG nova.network.neutron [req-6edfa8f6-8e3a-4096-a63d-ce199481ef59 req-fa86a79a-ac92-4a90-916f-dbb0b8b49ee0 service nova] [instance: 651e5142-7ab0-4040-97b4-4599bd4e419a] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 957.463215] env[62066]: DEBUG nova.network.neutron [-] [instance: 651e5142-7ab0-4040-97b4-4599bd4e419a] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 957.573017] env[62066]: DEBUG oslo_vmware.api [None req-c7b14697-dee9-4e9c-8ed4-e9c67e984eb5 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341224, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.694738] env[62066]: DEBUG oslo_vmware.api [None req-d41b2ff6-ff8d-407b-a347-8a3e15a584f9 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1341227, 'name': Rename_Task, 'duration_secs': 0.136695} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.694964] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-d41b2ff6-ff8d-407b-a347-8a3e15a584f9 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: ca504be5-209c-456c-af00-d403bfa2d634] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 957.695224] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4ea6cda5-9d23-444e-975f-8563eeefbf06 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.701699] env[62066]: DEBUG oslo_vmware.api [None req-d41b2ff6-ff8d-407b-a347-8a3e15a584f9 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Waiting for the task: (returnval){ [ 957.701699] env[62066]: value = "task-1341228" [ 957.701699] env[62066]: _type = "Task" [ 957.701699] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.709418] env[62066]: DEBUG oslo_vmware.api [None req-d41b2ff6-ff8d-407b-a347-8a3e15a584f9 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1341228, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.750195] env[62066]: DEBUG oslo_vmware.api [None req-96a3683c-7d82-48dc-bf96-f0467366c2eb tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': task-1341226, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.867018] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f28185e3-e8c2-4352-b716-1069aa95f5fc tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Lock "4e39b9b0-4ab0-4f48-8a3f-6bc1895597eb" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.734s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 957.870974] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-dfa85f83-e3a8-4439-bc23-98a2e7c2ecfb {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.880746] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e531019-9c0a-4eed-b4d9-121e7171d49d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.893431] env[62066]: DEBUG nova.compute.manager [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Didn't find any instances for network info cache update. {{(pid=62066) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 957.893635] env[62066]: DEBUG oslo_service.periodic_task [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 957.893878] env[62066]: DEBUG oslo_service.periodic_task [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 957.894088] env[62066]: DEBUG oslo_service.periodic_task [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 957.894287] env[62066]: DEBUG oslo_service.periodic_task [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 957.894478] env[62066]: DEBUG oslo_service.periodic_task [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 957.894671] env[62066]: DEBUG oslo_service.periodic_task [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 957.894830] env[62066]: DEBUG nova.compute.manager [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62066) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 957.894987] env[62066]: DEBUG oslo_service.periodic_task [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Running periodic task ComputeManager.update_available_resource {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 957.908388] env[62066]: DEBUG nova.compute.manager [req-6edfa8f6-8e3a-4096-a63d-ce199481ef59 req-fa86a79a-ac92-4a90-916f-dbb0b8b49ee0 service nova] [instance: 651e5142-7ab0-4040-97b4-4599bd4e419a] Detach interface failed, port_id=66054023-ce27-4e2d-92f5-18f1dde734d0, reason: Instance 651e5142-7ab0-4040-97b4-4599bd4e419a could not be found. {{(pid=62066) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 957.966180] env[62066]: INFO nova.compute.manager [-] [instance: 651e5142-7ab0-4040-97b4-4599bd4e419a] Took 1.30 seconds to deallocate network for instance. [ 958.074360] env[62066]: DEBUG oslo_vmware.api [None req-c7b14697-dee9-4e9c-8ed4-e9c67e984eb5 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341224, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.212150] env[62066]: DEBUG oslo_vmware.api [None req-d41b2ff6-ff8d-407b-a347-8a3e15a584f9 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1341228, 'name': PowerOnVM_Task, 'duration_secs': 0.443434} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.212524] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-d41b2ff6-ff8d-407b-a347-8a3e15a584f9 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: ca504be5-209c-456c-af00-d403bfa2d634] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 958.212800] env[62066]: INFO nova.compute.manager [None req-d41b2ff6-ff8d-407b-a347-8a3e15a584f9 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: ca504be5-209c-456c-af00-d403bfa2d634] Took 7.21 seconds to spawn the instance on the hypervisor. [ 958.213073] env[62066]: DEBUG nova.compute.manager [None req-d41b2ff6-ff8d-407b-a347-8a3e15a584f9 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: ca504be5-209c-456c-af00-d403bfa2d634] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 958.213927] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4665cbfe-fe59-4f98-b778-a7c7cf798a6a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.250638] env[62066]: DEBUG oslo_vmware.api [None req-96a3683c-7d82-48dc-bf96-f0467366c2eb tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': task-1341226, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.398432] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 958.398684] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 958.398986] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 958.399227] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62066) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 958.400141] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07c59982-fd6a-4c1d-89fb-1e3014898d43 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.408779] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7d6ff92-0ac7-4ea7-b2eb-8937e87859cf {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.422402] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2daf7a6c-312d-4882-9d0c-ed5c8dfb8e52 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.428947] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2389f2cd-3ae4-4d05-a8bd-35f6fce41358 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.460355] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180312MB free_disk=153GB free_vcpus=48 pci_devices=None {{(pid=62066) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 958.460355] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 958.460355] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 958.473333] env[62066]: DEBUG oslo_concurrency.lockutils [None req-980eec8a-3d17-48a4-863b-a262428cc66c tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 958.575122] env[62066]: DEBUG oslo_vmware.api [None req-c7b14697-dee9-4e9c-8ed4-e9c67e984eb5 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341224, 'name': ReconfigVM_Task, 'duration_secs': 1.743154} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.575122] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-c7b14697-dee9-4e9c-8ed4-e9c67e984eb5 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: 5693eea2-1cac-45d5-aa27-58526b65136e] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-286114', 'volume_id': 'e50e4638-0cff-4a5c-b7f9-9f8f39406291', 'name': 'volume-e50e4638-0cff-4a5c-b7f9-9f8f39406291', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '5693eea2-1cac-45d5-aa27-58526b65136e', 'attached_at': '', 'detached_at': '', 'volume_id': 'e50e4638-0cff-4a5c-b7f9-9f8f39406291', 'serial': 'e50e4638-0cff-4a5c-b7f9-9f8f39406291'} {{(pid=62066) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 958.731997] env[62066]: INFO nova.compute.manager [None req-d41b2ff6-ff8d-407b-a347-8a3e15a584f9 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: ca504be5-209c-456c-af00-d403bfa2d634] Took 12.01 seconds to build instance. [ 958.751925] env[62066]: DEBUG oslo_vmware.api [None req-96a3683c-7d82-48dc-bf96-f0467366c2eb tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': task-1341226, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.234761] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d41b2ff6-ff8d-407b-a347-8a3e15a584f9 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Lock "ca504be5-209c-456c-af00-d403bfa2d634" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.518s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 959.253343] env[62066]: DEBUG oslo_vmware.api [None req-96a3683c-7d82-48dc-bf96-f0467366c2eb tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': task-1341226, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.385515] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3dd8531f-78c0-4fc2-af15-28c19b0a2bf2 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Acquiring lock "09e3ae18-1517-470d-8cc9-a8b50ee774c7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 959.385794] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3dd8531f-78c0-4fc2-af15-28c19b0a2bf2 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Lock "09e3ae18-1517-470d-8cc9-a8b50ee774c7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 959.484970] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Instance 02fb3e92-5dd6-4b1e-a6e4-d60d3fc85b07 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 959.485144] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Instance c8e0d47c-4421-4e00-9183-206fceeabc40 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 959.485266] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Instance b718b09b-a7fa-445b-8be3-5b3ebca210a2 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 959.485386] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Instance 8ff6fbb9-c90f-498d-9a85-d220a8c2f794 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 959.485505] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Instance a10a6dc0-7df7-40a1-888a-d1414dbeb1f7 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 959.485644] env[62066]: WARNING nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Instance 651e5142-7ab0-4040-97b4-4599bd4e419a is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 959.485760] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Instance 5693eea2-1cac-45d5-aa27-58526b65136e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 959.622225] env[62066]: DEBUG nova.objects.instance [None req-c7b14697-dee9-4e9c-8ed4-e9c67e984eb5 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Lazy-loading 'flavor' on Instance uuid 5693eea2-1cac-45d5-aa27-58526b65136e {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 959.752577] env[62066]: DEBUG oslo_vmware.api [None req-96a3683c-7d82-48dc-bf96-f0467366c2eb tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': task-1341226, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.888059] env[62066]: DEBUG nova.compute.manager [None req-3dd8531f-78c0-4fc2-af15-28c19b0a2bf2 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 09e3ae18-1517-470d-8cc9-a8b50ee774c7] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 959.989032] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Instance 09e3ae18-1517-470d-8cc9-a8b50ee774c7 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 960.130935] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c7b14697-dee9-4e9c-8ed4-e9c67e984eb5 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Lock "5693eea2-1cac-45d5-aa27-58526b65136e" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 9.286s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 960.135743] env[62066]: DEBUG nova.compute.manager [None req-26ece0ef-0448-44a4-8d3a-1a89aaf96689 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: ca504be5-209c-456c-af00-d403bfa2d634] Stashing vm_state: active {{(pid=62066) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 960.252605] env[62066]: DEBUG oslo_vmware.api [None req-96a3683c-7d82-48dc-bf96-f0467366c2eb tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': task-1341226, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.275444] env[62066]: DEBUG oslo_concurrency.lockutils [None req-bfde70b7-abcf-4102-a176-bfc4580a84a1 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Acquiring lock "5693eea2-1cac-45d5-aa27-58526b65136e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 960.275444] env[62066]: DEBUG oslo_concurrency.lockutils [None req-bfde70b7-abcf-4102-a176-bfc4580a84a1 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Lock "5693eea2-1cac-45d5-aa27-58526b65136e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 960.275610] env[62066]: DEBUG oslo_concurrency.lockutils [None req-bfde70b7-abcf-4102-a176-bfc4580a84a1 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Acquiring lock "5693eea2-1cac-45d5-aa27-58526b65136e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 960.275787] env[62066]: DEBUG oslo_concurrency.lockutils [None req-bfde70b7-abcf-4102-a176-bfc4580a84a1 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Lock "5693eea2-1cac-45d5-aa27-58526b65136e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 960.275956] env[62066]: DEBUG oslo_concurrency.lockutils [None req-bfde70b7-abcf-4102-a176-bfc4580a84a1 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Lock "5693eea2-1cac-45d5-aa27-58526b65136e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 960.277982] env[62066]: INFO nova.compute.manager [None req-bfde70b7-abcf-4102-a176-bfc4580a84a1 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: 5693eea2-1cac-45d5-aa27-58526b65136e] Terminating instance [ 960.279945] env[62066]: DEBUG nova.compute.manager [None req-bfde70b7-abcf-4102-a176-bfc4580a84a1 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: 5693eea2-1cac-45d5-aa27-58526b65136e] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 960.280181] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-bfde70b7-abcf-4102-a176-bfc4580a84a1 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: 5693eea2-1cac-45d5-aa27-58526b65136e] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 960.280440] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9affd464-9b9e-4c50-832c-e5435dba94f5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.287018] env[62066]: DEBUG oslo_vmware.api [None req-bfde70b7-abcf-4102-a176-bfc4580a84a1 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Waiting for the task: (returnval){ [ 960.287018] env[62066]: value = "task-1341229" [ 960.287018] env[62066]: _type = "Task" [ 960.287018] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.294668] env[62066]: DEBUG oslo_vmware.api [None req-bfde70b7-abcf-4102-a176-bfc4580a84a1 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341229, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.412040] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3dd8531f-78c0-4fc2-af15-28c19b0a2bf2 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 960.491509] env[62066]: INFO nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Instance 1253aef0-8d14-4b3f-99eb-ce11746d0cb8 has allocations against this compute host but is not found in the database. [ 960.491710] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Instance ca504be5-209c-456c-af00-d403bfa2d634 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 960.491918] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Total usable vcpus: 48, total allocated vcpus: 7 {{(pid=62066) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 960.492078] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1856MB phys_disk=200GB used_disk=7GB total_vcpus=48 used_vcpus=7 pci_stats=[] {{(pid=62066) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 960.603680] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c7e2b62-1a72-4af3-ba78-ace11a65f016 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.611406] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5332e25-5f51-405e-b178-830cfe67e5d3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.645446] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ec4d4fc-bcab-4af6-bb3f-602d09624ef5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.652710] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46e57067-a477-4028-800d-922d1640ab06 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.657278] env[62066]: DEBUG oslo_concurrency.lockutils [None req-26ece0ef-0448-44a4-8d3a-1a89aaf96689 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 960.668014] env[62066]: DEBUG nova.compute.provider_tree [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 960.754063] env[62066]: DEBUG oslo_vmware.api [None req-96a3683c-7d82-48dc-bf96-f0467366c2eb tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': task-1341226, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.796182] env[62066]: DEBUG oslo_vmware.api [None req-bfde70b7-abcf-4102-a176-bfc4580a84a1 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341229, 'name': PowerOffVM_Task, 'duration_secs': 0.202752} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.796435] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-bfde70b7-abcf-4102-a176-bfc4580a84a1 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: 5693eea2-1cac-45d5-aa27-58526b65136e] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 960.796645] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-bfde70b7-abcf-4102-a176-bfc4580a84a1 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: 5693eea2-1cac-45d5-aa27-58526b65136e] Volume detach. Driver type: vmdk {{(pid=62066) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 960.796891] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-bfde70b7-abcf-4102-a176-bfc4580a84a1 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: 5693eea2-1cac-45d5-aa27-58526b65136e] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-286114', 'volume_id': 'e50e4638-0cff-4a5c-b7f9-9f8f39406291', 'name': 'volume-e50e4638-0cff-4a5c-b7f9-9f8f39406291', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '5693eea2-1cac-45d5-aa27-58526b65136e', 'attached_at': '', 'detached_at': '', 'volume_id': 'e50e4638-0cff-4a5c-b7f9-9f8f39406291', 'serial': 'e50e4638-0cff-4a5c-b7f9-9f8f39406291'} {{(pid=62066) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 960.797742] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe82d6a9-fb7b-4a8c-b38e-e70e7b5e9c36 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.819625] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92d0a726-9194-40c9-bcbd-3429fe55ee62 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.826046] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1ca58ea-27cf-4956-aa0e-9006b1a41bc4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.846248] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d30e7d49-5038-4101-9e30-0124ed046ac1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.860896] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-bfde70b7-abcf-4102-a176-bfc4580a84a1 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] The volume has not been displaced from its original location: [datastore2] volume-e50e4638-0cff-4a5c-b7f9-9f8f39406291/volume-e50e4638-0cff-4a5c-b7f9-9f8f39406291.vmdk. No consolidation needed. {{(pid=62066) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 960.866141] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-bfde70b7-abcf-4102-a176-bfc4580a84a1 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: 5693eea2-1cac-45d5-aa27-58526b65136e] Reconfiguring VM instance instance-0000005e to detach disk 2001 {{(pid=62066) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 960.866451] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bd72f68e-b622-4305-a558-adff5f681d6e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.883692] env[62066]: DEBUG oslo_vmware.api [None req-bfde70b7-abcf-4102-a176-bfc4580a84a1 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Waiting for the task: (returnval){ [ 960.883692] env[62066]: value = "task-1341230" [ 960.883692] env[62066]: _type = "Task" [ 960.883692] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.895073] env[62066]: DEBUG oslo_vmware.api [None req-bfde70b7-abcf-4102-a176-bfc4580a84a1 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341230, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.172135] env[62066]: DEBUG nova.scheduler.client.report [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 961.254259] env[62066]: DEBUG oslo_vmware.api [None req-96a3683c-7d82-48dc-bf96-f0467366c2eb tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': task-1341226, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.393161] env[62066]: DEBUG oslo_vmware.api [None req-bfde70b7-abcf-4102-a176-bfc4580a84a1 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341230, 'name': ReconfigVM_Task, 'duration_secs': 0.185355} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.393444] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-bfde70b7-abcf-4102-a176-bfc4580a84a1 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: 5693eea2-1cac-45d5-aa27-58526b65136e] Reconfigured VM instance instance-0000005e to detach disk 2001 {{(pid=62066) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 961.398145] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d3c45d2a-2a5f-4ff4-95d6-9565744227b2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.412908] env[62066]: DEBUG oslo_vmware.api [None req-bfde70b7-abcf-4102-a176-bfc4580a84a1 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Waiting for the task: (returnval){ [ 961.412908] env[62066]: value = "task-1341231" [ 961.412908] env[62066]: _type = "Task" [ 961.412908] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.420398] env[62066]: DEBUG oslo_vmware.api [None req-bfde70b7-abcf-4102-a176-bfc4580a84a1 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341231, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.508741] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0d13ef41-2caf-426c-8207-6420bd8c108b tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Acquiring lock "c8e0d47c-4421-4e00-9183-206fceeabc40" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 961.509052] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0d13ef41-2caf-426c-8207-6420bd8c108b tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Lock "c8e0d47c-4421-4e00-9183-206fceeabc40" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 961.675830] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62066) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 961.676082] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.216s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 961.676368] env[62066]: DEBUG oslo_concurrency.lockutils [None req-980eec8a-3d17-48a4-863b-a262428cc66c tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.203s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 961.676561] env[62066]: DEBUG oslo_concurrency.lockutils [None req-980eec8a-3d17-48a4-863b-a262428cc66c tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 961.678699] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3dd8531f-78c0-4fc2-af15-28c19b0a2bf2 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.267s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 961.680214] env[62066]: INFO nova.compute.claims [None req-3dd8531f-78c0-4fc2-af15-28c19b0a2bf2 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 09e3ae18-1517-470d-8cc9-a8b50ee774c7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 961.702254] env[62066]: INFO nova.scheduler.client.report [None req-980eec8a-3d17-48a4-863b-a262428cc66c tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Deleted allocations for instance 651e5142-7ab0-4040-97b4-4599bd4e419a [ 961.754911] env[62066]: DEBUG oslo_vmware.api [None req-96a3683c-7d82-48dc-bf96-f0467366c2eb tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': task-1341226, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.924102] env[62066]: DEBUG oslo_vmware.api [None req-bfde70b7-abcf-4102-a176-bfc4580a84a1 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341231, 'name': ReconfigVM_Task, 'duration_secs': 0.421078} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.924236] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-bfde70b7-abcf-4102-a176-bfc4580a84a1 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: 5693eea2-1cac-45d5-aa27-58526b65136e] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-286114', 'volume_id': 'e50e4638-0cff-4a5c-b7f9-9f8f39406291', 'name': 'volume-e50e4638-0cff-4a5c-b7f9-9f8f39406291', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '5693eea2-1cac-45d5-aa27-58526b65136e', 'attached_at': '', 'detached_at': '', 'volume_id': 'e50e4638-0cff-4a5c-b7f9-9f8f39406291', 'serial': 'e50e4638-0cff-4a5c-b7f9-9f8f39406291'} {{(pid=62066) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 961.924494] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-bfde70b7-abcf-4102-a176-bfc4580a84a1 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: 5693eea2-1cac-45d5-aa27-58526b65136e] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 961.925228] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8fbf91b-b5ca-4dc2-9437-fbce785a02a3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.931518] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-bfde70b7-abcf-4102-a176-bfc4580a84a1 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: 5693eea2-1cac-45d5-aa27-58526b65136e] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 961.931734] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-58ee4fa7-3722-470d-9243-d32e7aa4368c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.995389] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-bfde70b7-abcf-4102-a176-bfc4580a84a1 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: 5693eea2-1cac-45d5-aa27-58526b65136e] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 961.995981] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-bfde70b7-abcf-4102-a176-bfc4580a84a1 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: 5693eea2-1cac-45d5-aa27-58526b65136e] Deleting contents of the VM from datastore datastore1 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 961.995981] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-bfde70b7-abcf-4102-a176-bfc4580a84a1 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Deleting the datastore file [datastore1] 5693eea2-1cac-45d5-aa27-58526b65136e {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 961.996185] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0a62e403-4491-47c0-bd04-8f217a8529ec {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.002995] env[62066]: DEBUG oslo_vmware.api [None req-bfde70b7-abcf-4102-a176-bfc4580a84a1 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Waiting for the task: (returnval){ [ 962.002995] env[62066]: value = "task-1341233" [ 962.002995] env[62066]: _type = "Task" [ 962.002995] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.011441] env[62066]: INFO nova.compute.manager [None req-0d13ef41-2caf-426c-8207-6420bd8c108b tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: c8e0d47c-4421-4e00-9183-206fceeabc40] Detaching volume 338bd175-a96d-417e-a5b0-380b7d939a85 [ 962.013186] env[62066]: DEBUG oslo_vmware.api [None req-bfde70b7-abcf-4102-a176-bfc4580a84a1 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341233, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.043745] env[62066]: INFO nova.virt.block_device [None req-0d13ef41-2caf-426c-8207-6420bd8c108b tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: c8e0d47c-4421-4e00-9183-206fceeabc40] Attempting to driver detach volume 338bd175-a96d-417e-a5b0-380b7d939a85 from mountpoint /dev/sdb [ 962.044057] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-0d13ef41-2caf-426c-8207-6420bd8c108b tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: c8e0d47c-4421-4e00-9183-206fceeabc40] Volume detach. Driver type: vmdk {{(pid=62066) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 962.044190] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-0d13ef41-2caf-426c-8207-6420bd8c108b tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: c8e0d47c-4421-4e00-9183-206fceeabc40] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-286099', 'volume_id': '338bd175-a96d-417e-a5b0-380b7d939a85', 'name': 'volume-338bd175-a96d-417e-a5b0-380b7d939a85', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'c8e0d47c-4421-4e00-9183-206fceeabc40', 'attached_at': '', 'detached_at': '', 'volume_id': '338bd175-a96d-417e-a5b0-380b7d939a85', 'serial': '338bd175-a96d-417e-a5b0-380b7d939a85'} {{(pid=62066) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 962.045036] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acdbf9c3-561e-4af6-b307-72eba53c0e90 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.066098] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf7786c8-594f-4052-96db-11a5e8a2acad {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.072767] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e18aaf5-49a5-409f-9696-55640483d03d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.092487] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84fa8129-4ccd-4e73-a7ec-e79633027942 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.107047] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-0d13ef41-2caf-426c-8207-6420bd8c108b tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] The volume has not been displaced from its original location: [datastore2] volume-338bd175-a96d-417e-a5b0-380b7d939a85/volume-338bd175-a96d-417e-a5b0-380b7d939a85.vmdk. No consolidation needed. {{(pid=62066) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 962.112319] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-0d13ef41-2caf-426c-8207-6420bd8c108b tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: c8e0d47c-4421-4e00-9183-206fceeabc40] Reconfiguring VM instance instance-0000004f to detach disk 2001 {{(pid=62066) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 962.112615] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-43931994-c73f-4af3-a315-60ac6200e0c1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.130690] env[62066]: DEBUG oslo_vmware.api [None req-0d13ef41-2caf-426c-8207-6420bd8c108b tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Waiting for the task: (returnval){ [ 962.130690] env[62066]: value = "task-1341234" [ 962.130690] env[62066]: _type = "Task" [ 962.130690] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.139332] env[62066]: DEBUG oslo_vmware.api [None req-0d13ef41-2caf-426c-8207-6420bd8c108b tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Task: {'id': task-1341234, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.208762] env[62066]: DEBUG oslo_concurrency.lockutils [None req-980eec8a-3d17-48a4-863b-a262428cc66c tempest-ImagesTestJSON-474857627 tempest-ImagesTestJSON-474857627-project-member] Lock "651e5142-7ab0-4040-97b4-4599bd4e419a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.706s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 962.255392] env[62066]: DEBUG oslo_vmware.api [None req-96a3683c-7d82-48dc-bf96-f0467366c2eb tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': task-1341226, 'name': ReconfigVM_Task} progress is 18%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.513053] env[62066]: DEBUG oslo_vmware.api [None req-bfde70b7-abcf-4102-a176-bfc4580a84a1 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341233, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.159584} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.513175] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-bfde70b7-abcf-4102-a176-bfc4580a84a1 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 962.513374] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-bfde70b7-abcf-4102-a176-bfc4580a84a1 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: 5693eea2-1cac-45d5-aa27-58526b65136e] Deleted contents of the VM from datastore datastore1 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 962.513557] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-bfde70b7-abcf-4102-a176-bfc4580a84a1 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: 5693eea2-1cac-45d5-aa27-58526b65136e] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 962.513736] env[62066]: INFO nova.compute.manager [None req-bfde70b7-abcf-4102-a176-bfc4580a84a1 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: 5693eea2-1cac-45d5-aa27-58526b65136e] Took 2.23 seconds to destroy the instance on the hypervisor. [ 962.514055] env[62066]: DEBUG oslo.service.loopingcall [None req-bfde70b7-abcf-4102-a176-bfc4580a84a1 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 962.514265] env[62066]: DEBUG nova.compute.manager [-] [instance: 5693eea2-1cac-45d5-aa27-58526b65136e] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 962.514364] env[62066]: DEBUG nova.network.neutron [-] [instance: 5693eea2-1cac-45d5-aa27-58526b65136e] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 962.642207] env[62066]: DEBUG oslo_vmware.api [None req-0d13ef41-2caf-426c-8207-6420bd8c108b tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Task: {'id': task-1341234, 'name': ReconfigVM_Task, 'duration_secs': 0.209071} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.642437] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-0d13ef41-2caf-426c-8207-6420bd8c108b tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: c8e0d47c-4421-4e00-9183-206fceeabc40] Reconfigured VM instance instance-0000004f to detach disk 2001 {{(pid=62066) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 962.646953] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4117faf4-9378-4fe4-a1a2-de96482bd14a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.662305] env[62066]: DEBUG oslo_vmware.api [None req-0d13ef41-2caf-426c-8207-6420bd8c108b tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Waiting for the task: (returnval){ [ 962.662305] env[62066]: value = "task-1341235" [ 962.662305] env[62066]: _type = "Task" [ 962.662305] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.670050] env[62066]: DEBUG oslo_vmware.api [None req-0d13ef41-2caf-426c-8207-6420bd8c108b tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Task: {'id': task-1341235, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.735262] env[62066]: DEBUG nova.compute.manager [req-27a42a72-1ddf-4a8d-8508-c1f0bda0e897 req-7c5747b3-9ffe-4332-b01b-335f4d8d851f service nova] [instance: 5693eea2-1cac-45d5-aa27-58526b65136e] Received event network-vif-deleted-7e38fd56-d921-4784-9206-fdff79574fa4 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 962.735555] env[62066]: INFO nova.compute.manager [req-27a42a72-1ddf-4a8d-8508-c1f0bda0e897 req-7c5747b3-9ffe-4332-b01b-335f4d8d851f service nova] [instance: 5693eea2-1cac-45d5-aa27-58526b65136e] Neutron deleted interface 7e38fd56-d921-4784-9206-fdff79574fa4; detaching it from the instance and deleting it from the info cache [ 962.735640] env[62066]: DEBUG nova.network.neutron [req-27a42a72-1ddf-4a8d-8508-c1f0bda0e897 req-7c5747b3-9ffe-4332-b01b-335f4d8d851f service nova] [instance: 5693eea2-1cac-45d5-aa27-58526b65136e] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 962.758500] env[62066]: DEBUG oslo_vmware.api [None req-96a3683c-7d82-48dc-bf96-f0467366c2eb tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': task-1341226, 'name': ReconfigVM_Task, 'duration_secs': 5.767121} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.758718] env[62066]: DEBUG oslo_concurrency.lockutils [None req-96a3683c-7d82-48dc-bf96-f0467366c2eb tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Releasing lock "b718b09b-a7fa-445b-8be3-5b3ebca210a2" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 962.758942] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-96a3683c-7d82-48dc-bf96-f0467366c2eb tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: b718b09b-a7fa-445b-8be3-5b3ebca210a2] Reconfigured VM to detach interface {{(pid=62066) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 962.816842] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f35bbcdf-630e-46fc-a057-2010198b4902 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.824429] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55ccd9f8-0432-4831-a05f-c2af97746bee {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.854139] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e114911e-2378-4224-b451-f8fc9cf59f80 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.861329] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c7daa21-51eb-4456-b598-32d1f71deee9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.874553] env[62066]: DEBUG nova.compute.provider_tree [None req-3dd8531f-78c0-4fc2-af15-28c19b0a2bf2 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 963.172018] env[62066]: DEBUG oslo_vmware.api [None req-0d13ef41-2caf-426c-8207-6420bd8c108b tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Task: {'id': task-1341235, 'name': ReconfigVM_Task, 'duration_secs': 0.136795} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.173084] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-0d13ef41-2caf-426c-8207-6420bd8c108b tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: c8e0d47c-4421-4e00-9183-206fceeabc40] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-286099', 'volume_id': '338bd175-a96d-417e-a5b0-380b7d939a85', 'name': 'volume-338bd175-a96d-417e-a5b0-380b7d939a85', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'c8e0d47c-4421-4e00-9183-206fceeabc40', 'attached_at': '', 'detached_at': '', 'volume_id': '338bd175-a96d-417e-a5b0-380b7d939a85', 'serial': '338bd175-a96d-417e-a5b0-380b7d939a85'} {{(pid=62066) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 963.213614] env[62066]: DEBUG nova.network.neutron [-] [instance: 5693eea2-1cac-45d5-aa27-58526b65136e] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 963.240739] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5a7493ba-3e6e-485c-a4a6-931bd748071a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.249938] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-432c995a-8a67-4ea2-9e66-f5ccf1362039 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.278785] env[62066]: DEBUG nova.compute.manager [req-27a42a72-1ddf-4a8d-8508-c1f0bda0e897 req-7c5747b3-9ffe-4332-b01b-335f4d8d851f service nova] [instance: 5693eea2-1cac-45d5-aa27-58526b65136e] Detach interface failed, port_id=7e38fd56-d921-4784-9206-fdff79574fa4, reason: Instance 5693eea2-1cac-45d5-aa27-58526b65136e could not be found. {{(pid=62066) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 963.377651] env[62066]: DEBUG nova.scheduler.client.report [None req-3dd8531f-78c0-4fc2-af15-28c19b0a2bf2 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 963.600959] env[62066]: DEBUG nova.compute.manager [req-2e69dbda-bb70-42d3-b5fe-0025cbfc75b7 req-f2890efd-3684-49c2-9b90-7186dc657fa2 service nova] [instance: b718b09b-a7fa-445b-8be3-5b3ebca210a2] Received event network-vif-deleted-c9952f47-6c37-48e3-ad4e-0c2cc1c5bdb0 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 963.601273] env[62066]: INFO nova.compute.manager [req-2e69dbda-bb70-42d3-b5fe-0025cbfc75b7 req-f2890efd-3684-49c2-9b90-7186dc657fa2 service nova] [instance: b718b09b-a7fa-445b-8be3-5b3ebca210a2] Neutron deleted interface c9952f47-6c37-48e3-ad4e-0c2cc1c5bdb0; detaching it from the instance and deleting it from the info cache [ 963.601678] env[62066]: DEBUG nova.network.neutron [req-2e69dbda-bb70-42d3-b5fe-0025cbfc75b7 req-f2890efd-3684-49c2-9b90-7186dc657fa2 service nova] [instance: b718b09b-a7fa-445b-8be3-5b3ebca210a2] Updating instance_info_cache with network_info: [{"id": "3d01745e-0d9e-4bae-ba4d-242380edbab6", "address": "fa:16:3e:ed:53:4c", "network": {"id": "449dfe4e-bb49-43b6-9ba7-b57af74ebfb3", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-282029676-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.224", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0d4e33a99c7741fb8cdd97f4ec5dbbd8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db68bd64-5b56-49af-a075-13dcf85cb2e0", "external-id": "nsx-vlan-transportzone-590", "segmentation_id": 590, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3d01745e-0d", "ovs_interfaceid": "3d01745e-0d9e-4bae-ba4d-242380edbab6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "26c5cd4b-0e0d-4e99-a7fe-7a46764dc642", "address": "fa:16:3e:ed:67:f7", "network": {"id": "449dfe4e-bb49-43b6-9ba7-b57af74ebfb3", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-282029676-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0d4e33a99c7741fb8cdd97f4ec5dbbd8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db68bd64-5b56-49af-a075-13dcf85cb2e0", "external-id": "nsx-vlan-transportzone-590", "segmentation_id": 590, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap26c5cd4b-0e", "ovs_interfaceid": "26c5cd4b-0e0d-4e99-a7fe-7a46764dc642", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 963.715447] env[62066]: DEBUG nova.objects.instance [None req-0d13ef41-2caf-426c-8207-6420bd8c108b tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Lazy-loading 'flavor' on Instance uuid c8e0d47c-4421-4e00-9183-206fceeabc40 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 963.716869] env[62066]: INFO nova.compute.manager [-] [instance: 5693eea2-1cac-45d5-aa27-58526b65136e] Took 1.20 seconds to deallocate network for instance. [ 963.884157] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3dd8531f-78c0-4fc2-af15-28c19b0a2bf2 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.205s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 963.884742] env[62066]: DEBUG nova.compute.manager [None req-3dd8531f-78c0-4fc2-af15-28c19b0a2bf2 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 09e3ae18-1517-470d-8cc9-a8b50ee774c7] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 963.888073] env[62066]: DEBUG oslo_concurrency.lockutils [None req-26ece0ef-0448-44a4-8d3a-1a89aaf96689 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 3.231s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 964.001960] env[62066]: DEBUG oslo_concurrency.lockutils [None req-96a3683c-7d82-48dc-bf96-f0467366c2eb tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Acquiring lock "refresh_cache-b718b09b-a7fa-445b-8be3-5b3ebca210a2" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 964.002172] env[62066]: DEBUG oslo_concurrency.lockutils [None req-96a3683c-7d82-48dc-bf96-f0467366c2eb tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Acquired lock "refresh_cache-b718b09b-a7fa-445b-8be3-5b3ebca210a2" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 964.002295] env[62066]: DEBUG nova.network.neutron [None req-96a3683c-7d82-48dc-bf96-f0467366c2eb tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: b718b09b-a7fa-445b-8be3-5b3ebca210a2] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 964.057443] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ea8fb2df-bddb-407b-bcac-dde5dd08fd22 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Acquiring lock "8ff6fbb9-c90f-498d-9a85-d220a8c2f794" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 964.057875] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ea8fb2df-bddb-407b-bcac-dde5dd08fd22 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Lock "8ff6fbb9-c90f-498d-9a85-d220a8c2f794" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 964.058219] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ea8fb2df-bddb-407b-bcac-dde5dd08fd22 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Acquiring lock "8ff6fbb9-c90f-498d-9a85-d220a8c2f794-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 964.058520] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ea8fb2df-bddb-407b-bcac-dde5dd08fd22 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Lock "8ff6fbb9-c90f-498d-9a85-d220a8c2f794-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 964.058802] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ea8fb2df-bddb-407b-bcac-dde5dd08fd22 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Lock "8ff6fbb9-c90f-498d-9a85-d220a8c2f794-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 964.060915] env[62066]: INFO nova.compute.manager [None req-ea8fb2df-bddb-407b-bcac-dde5dd08fd22 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] [instance: 8ff6fbb9-c90f-498d-9a85-d220a8c2f794] Terminating instance [ 964.065600] env[62066]: DEBUG nova.compute.manager [None req-ea8fb2df-bddb-407b-bcac-dde5dd08fd22 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] [instance: 8ff6fbb9-c90f-498d-9a85-d220a8c2f794] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 964.065600] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-ea8fb2df-bddb-407b-bcac-dde5dd08fd22 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] [instance: 8ff6fbb9-c90f-498d-9a85-d220a8c2f794] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 964.065600] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47d0d315-26c8-441c-9c7b-b61939e6bd6b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.072035] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-ea8fb2df-bddb-407b-bcac-dde5dd08fd22 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] [instance: 8ff6fbb9-c90f-498d-9a85-d220a8c2f794] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 964.072394] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2864f533-2ca4-494e-8f93-134fa763375f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.078743] env[62066]: DEBUG oslo_vmware.api [None req-ea8fb2df-bddb-407b-bcac-dde5dd08fd22 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Waiting for the task: (returnval){ [ 964.078743] env[62066]: value = "task-1341236" [ 964.078743] env[62066]: _type = "Task" [ 964.078743] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.087078] env[62066]: DEBUG oslo_vmware.api [None req-ea8fb2df-bddb-407b-bcac-dde5dd08fd22 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Task: {'id': task-1341236, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.105270] env[62066]: DEBUG oslo_concurrency.lockutils [req-2e69dbda-bb70-42d3-b5fe-0025cbfc75b7 req-f2890efd-3684-49c2-9b90-7186dc657fa2 service nova] Acquiring lock "b718b09b-a7fa-445b-8be3-5b3ebca210a2" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 964.105374] env[62066]: DEBUG oslo_concurrency.lockutils [req-2e69dbda-bb70-42d3-b5fe-0025cbfc75b7 req-f2890efd-3684-49c2-9b90-7186dc657fa2 service nova] Acquired lock "b718b09b-a7fa-445b-8be3-5b3ebca210a2" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 964.106238] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a41fa179-0c8e-49ec-9fbb-df2d4ca676af {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.126892] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72130d3e-77cc-42bf-8193-102a35a102cc {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.156176] env[62066]: DEBUG nova.virt.vmwareapi.vmops [req-2e69dbda-bb70-42d3-b5fe-0025cbfc75b7 req-f2890efd-3684-49c2-9b90-7186dc657fa2 service nova] [instance: b718b09b-a7fa-445b-8be3-5b3ebca210a2] Reconfiguring VM to detach interface {{(pid=62066) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 964.156542] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-902fc082-8cee-41ea-9643-3444ec533202 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.171671] env[62066]: DEBUG oslo_concurrency.lockutils [None req-2552ba89-1024-40da-89b2-2a360324a01f tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Acquiring lock "b718b09b-a7fa-445b-8be3-5b3ebca210a2" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 964.175235] env[62066]: DEBUG oslo_vmware.api [req-2e69dbda-bb70-42d3-b5fe-0025cbfc75b7 req-f2890efd-3684-49c2-9b90-7186dc657fa2 service nova] Waiting for the task: (returnval){ [ 964.175235] env[62066]: value = "task-1341237" [ 964.175235] env[62066]: _type = "Task" [ 964.175235] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.183984] env[62066]: DEBUG oslo_vmware.api [req-2e69dbda-bb70-42d3-b5fe-0025cbfc75b7 req-f2890efd-3684-49c2-9b90-7186dc657fa2 service nova] Task: {'id': task-1341237, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.272072] env[62066]: INFO nova.compute.manager [None req-bfde70b7-abcf-4102-a176-bfc4580a84a1 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: 5693eea2-1cac-45d5-aa27-58526b65136e] Took 0.55 seconds to detach 1 volumes for instance. [ 964.393417] env[62066]: DEBUG nova.compute.utils [None req-3dd8531f-78c0-4fc2-af15-28c19b0a2bf2 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 964.396916] env[62066]: INFO nova.compute.claims [None req-26ece0ef-0448-44a4-8d3a-1a89aaf96689 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: ca504be5-209c-456c-af00-d403bfa2d634] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 964.407023] env[62066]: DEBUG nova.compute.manager [None req-3dd8531f-78c0-4fc2-af15-28c19b0a2bf2 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 09e3ae18-1517-470d-8cc9-a8b50ee774c7] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 964.407023] env[62066]: DEBUG nova.network.neutron [None req-3dd8531f-78c0-4fc2-af15-28c19b0a2bf2 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 09e3ae18-1517-470d-8cc9-a8b50ee774c7] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 964.456475] env[62066]: DEBUG nova.policy [None req-3dd8531f-78c0-4fc2-af15-28c19b0a2bf2 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8910a229218b4ec5ad72c893badfc598', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6497ab02f327476d8ff81c2ecc0371e5', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 964.589011] env[62066]: DEBUG oslo_vmware.api [None req-ea8fb2df-bddb-407b-bcac-dde5dd08fd22 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Task: {'id': task-1341236, 'name': PowerOffVM_Task, 'duration_secs': 0.186094} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.589655] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-ea8fb2df-bddb-407b-bcac-dde5dd08fd22 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] [instance: 8ff6fbb9-c90f-498d-9a85-d220a8c2f794] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 964.589832] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-ea8fb2df-bddb-407b-bcac-dde5dd08fd22 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] [instance: 8ff6fbb9-c90f-498d-9a85-d220a8c2f794] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 964.590162] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-af9f1e47-ce8a-405f-9d8c-93e444acb10f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.673897] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-ea8fb2df-bddb-407b-bcac-dde5dd08fd22 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] [instance: 8ff6fbb9-c90f-498d-9a85-d220a8c2f794] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 964.674867] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-ea8fb2df-bddb-407b-bcac-dde5dd08fd22 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] [instance: 8ff6fbb9-c90f-498d-9a85-d220a8c2f794] Deleting contents of the VM from datastore datastore1 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 964.674867] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-ea8fb2df-bddb-407b-bcac-dde5dd08fd22 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Deleting the datastore file [datastore1] 8ff6fbb9-c90f-498d-9a85-d220a8c2f794 {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 964.674867] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7386c878-b46c-42bf-9b17-fe6d9d512aae {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.695849] env[62066]: DEBUG oslo_vmware.api [None req-ea8fb2df-bddb-407b-bcac-dde5dd08fd22 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Waiting for the task: (returnval){ [ 964.695849] env[62066]: value = "task-1341239" [ 964.695849] env[62066]: _type = "Task" [ 964.695849] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.696099] env[62066]: DEBUG oslo_vmware.api [req-2e69dbda-bb70-42d3-b5fe-0025cbfc75b7 req-f2890efd-3684-49c2-9b90-7186dc657fa2 service nova] Task: {'id': task-1341237, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.709718] env[62066]: DEBUG oslo_vmware.api [None req-ea8fb2df-bddb-407b-bcac-dde5dd08fd22 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Task: {'id': task-1341239, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.726400] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0d13ef41-2caf-426c-8207-6420bd8c108b tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Lock "c8e0d47c-4421-4e00-9183-206fceeabc40" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.217s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 964.736906] env[62066]: DEBUG oslo_concurrency.lockutils [None req-41711bcc-fddb-4aba-b7e4-374d02404b25 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Acquiring lock "c8e0d47c-4421-4e00-9183-206fceeabc40" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 964.736906] env[62066]: DEBUG oslo_concurrency.lockutils [None req-41711bcc-fddb-4aba-b7e4-374d02404b25 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Lock "c8e0d47c-4421-4e00-9183-206fceeabc40" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 964.736906] env[62066]: DEBUG nova.compute.manager [None req-41711bcc-fddb-4aba-b7e4-374d02404b25 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: c8e0d47c-4421-4e00-9183-206fceeabc40] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 964.737667] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2090fa7b-8fcc-4786-9c6f-6d399564b9eb {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.745794] env[62066]: DEBUG nova.compute.manager [None req-41711bcc-fddb-4aba-b7e4-374d02404b25 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: c8e0d47c-4421-4e00-9183-206fceeabc40] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62066) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 964.746371] env[62066]: DEBUG nova.objects.instance [None req-41711bcc-fddb-4aba-b7e4-374d02404b25 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Lazy-loading 'flavor' on Instance uuid c8e0d47c-4421-4e00-9183-206fceeabc40 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 964.769603] env[62066]: DEBUG nova.compute.manager [req-93968833-5618-4e60-b150-1a661ce4b6b4 req-0dde6c1f-b6d4-4fb5-ba44-c829d1d16ec3 service nova] [instance: b718b09b-a7fa-445b-8be3-5b3ebca210a2] Received event network-vif-deleted-26c5cd4b-0e0d-4e99-a7fe-7a46764dc642 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 964.769821] env[62066]: INFO nova.compute.manager [req-93968833-5618-4e60-b150-1a661ce4b6b4 req-0dde6c1f-b6d4-4fb5-ba44-c829d1d16ec3 service nova] [instance: b718b09b-a7fa-445b-8be3-5b3ebca210a2] Neutron deleted interface 26c5cd4b-0e0d-4e99-a7fe-7a46764dc642; detaching it from the instance and deleting it from the info cache [ 964.770656] env[62066]: DEBUG nova.network.neutron [req-93968833-5618-4e60-b150-1a661ce4b6b4 req-0dde6c1f-b6d4-4fb5-ba44-c829d1d16ec3 service nova] [instance: b718b09b-a7fa-445b-8be3-5b3ebca210a2] Updating instance_info_cache with network_info: [{"id": "3d01745e-0d9e-4bae-ba4d-242380edbab6", "address": "fa:16:3e:ed:53:4c", "network": {"id": "449dfe4e-bb49-43b6-9ba7-b57af74ebfb3", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-282029676-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.224", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0d4e33a99c7741fb8cdd97f4ec5dbbd8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db68bd64-5b56-49af-a075-13dcf85cb2e0", "external-id": "nsx-vlan-transportzone-590", "segmentation_id": 590, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3d01745e-0d", "ovs_interfaceid": "3d01745e-0d9e-4bae-ba4d-242380edbab6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 964.779037] env[62066]: INFO nova.network.neutron [None req-96a3683c-7d82-48dc-bf96-f0467366c2eb tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: b718b09b-a7fa-445b-8be3-5b3ebca210a2] Port 26c5cd4b-0e0d-4e99-a7fe-7a46764dc642 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 964.779037] env[62066]: DEBUG nova.network.neutron [None req-96a3683c-7d82-48dc-bf96-f0467366c2eb tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: b718b09b-a7fa-445b-8be3-5b3ebca210a2] Updating instance_info_cache with network_info: [{"id": "3d01745e-0d9e-4bae-ba4d-242380edbab6", "address": "fa:16:3e:ed:53:4c", "network": {"id": "449dfe4e-bb49-43b6-9ba7-b57af74ebfb3", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-282029676-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.224", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0d4e33a99c7741fb8cdd97f4ec5dbbd8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db68bd64-5b56-49af-a075-13dcf85cb2e0", "external-id": "nsx-vlan-transportzone-590", "segmentation_id": 590, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3d01745e-0d", "ovs_interfaceid": "3d01745e-0d9e-4bae-ba4d-242380edbab6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 964.780025] env[62066]: DEBUG oslo_concurrency.lockutils [None req-bfde70b7-abcf-4102-a176-bfc4580a84a1 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 964.904489] env[62066]: DEBUG nova.compute.manager [None req-3dd8531f-78c0-4fc2-af15-28c19b0a2bf2 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 09e3ae18-1517-470d-8cc9-a8b50ee774c7] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 964.909547] env[62066]: INFO nova.compute.resource_tracker [None req-26ece0ef-0448-44a4-8d3a-1a89aaf96689 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: ca504be5-209c-456c-af00-d403bfa2d634] Updating resource usage from migration 1253aef0-8d14-4b3f-99eb-ce11746d0cb8 [ 964.964958] env[62066]: DEBUG nova.network.neutron [None req-3dd8531f-78c0-4fc2-af15-28c19b0a2bf2 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 09e3ae18-1517-470d-8cc9-a8b50ee774c7] Successfully created port: f62ec0e9-6137-4099-8d1b-62a38a3a9938 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 965.054426] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b3dcef5-3680-4944-964c-9e2af92c3e48 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.062347] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f51bf7a0-7b97-47a4-b9b1-02aaffef9222 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.094567] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c66eb558-0aff-4fb1-9017-66abe2f45210 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.102930] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e13e235-53f9-48d5-b57c-8f30d96726eb {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.116932] env[62066]: DEBUG nova.compute.provider_tree [None req-26ece0ef-0448-44a4-8d3a-1a89aaf96689 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 965.187017] env[62066]: DEBUG oslo_vmware.api [req-2e69dbda-bb70-42d3-b5fe-0025cbfc75b7 req-f2890efd-3684-49c2-9b90-7186dc657fa2 service nova] Task: {'id': task-1341237, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.210102] env[62066]: DEBUG oslo_vmware.api [None req-ea8fb2df-bddb-407b-bcac-dde5dd08fd22 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Task: {'id': task-1341239, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.134548} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.210642] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-ea8fb2df-bddb-407b-bcac-dde5dd08fd22 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 965.210972] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-ea8fb2df-bddb-407b-bcac-dde5dd08fd22 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] [instance: 8ff6fbb9-c90f-498d-9a85-d220a8c2f794] Deleted contents of the VM from datastore datastore1 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 965.211353] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-ea8fb2df-bddb-407b-bcac-dde5dd08fd22 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] [instance: 8ff6fbb9-c90f-498d-9a85-d220a8c2f794] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 965.211666] env[62066]: INFO nova.compute.manager [None req-ea8fb2df-bddb-407b-bcac-dde5dd08fd22 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] [instance: 8ff6fbb9-c90f-498d-9a85-d220a8c2f794] Took 1.15 seconds to destroy the instance on the hypervisor. [ 965.212060] env[62066]: DEBUG oslo.service.loopingcall [None req-ea8fb2df-bddb-407b-bcac-dde5dd08fd22 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 965.212402] env[62066]: DEBUG nova.compute.manager [-] [instance: 8ff6fbb9-c90f-498d-9a85-d220a8c2f794] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 965.212645] env[62066]: DEBUG nova.network.neutron [-] [instance: 8ff6fbb9-c90f-498d-9a85-d220a8c2f794] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 965.252697] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-41711bcc-fddb-4aba-b7e4-374d02404b25 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: c8e0d47c-4421-4e00-9183-206fceeabc40] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 965.252998] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b71c5cbe-035e-4893-8eef-b455429f5346 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.261461] env[62066]: DEBUG oslo_vmware.api [None req-41711bcc-fddb-4aba-b7e4-374d02404b25 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Waiting for the task: (returnval){ [ 965.261461] env[62066]: value = "task-1341240" [ 965.261461] env[62066]: _type = "Task" [ 965.261461] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.271057] env[62066]: DEBUG oslo_vmware.api [None req-41711bcc-fddb-4aba-b7e4-374d02404b25 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Task: {'id': task-1341240, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.273512] env[62066]: DEBUG oslo_concurrency.lockutils [req-93968833-5618-4e60-b150-1a661ce4b6b4 req-0dde6c1f-b6d4-4fb5-ba44-c829d1d16ec3 service nova] Acquiring lock "b718b09b-a7fa-445b-8be3-5b3ebca210a2" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 965.282189] env[62066]: DEBUG oslo_concurrency.lockutils [None req-96a3683c-7d82-48dc-bf96-f0467366c2eb tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Releasing lock "refresh_cache-b718b09b-a7fa-445b-8be3-5b3ebca210a2" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 965.619961] env[62066]: DEBUG nova.scheduler.client.report [None req-26ece0ef-0448-44a4-8d3a-1a89aaf96689 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 965.687802] env[62066]: DEBUG oslo_vmware.api [req-2e69dbda-bb70-42d3-b5fe-0025cbfc75b7 req-f2890efd-3684-49c2-9b90-7186dc657fa2 service nova] Task: {'id': task-1341237, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.761555] env[62066]: DEBUG nova.compute.manager [req-64fc33f1-68cd-4f99-8edc-c3872e35913d req-758bfa9b-349f-4be2-a213-3ab90c259669 service nova] [instance: 8ff6fbb9-c90f-498d-9a85-d220a8c2f794] Received event network-vif-deleted-8c324833-7bf9-4007-85aa-5b16871f63c7 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 965.761764] env[62066]: INFO nova.compute.manager [req-64fc33f1-68cd-4f99-8edc-c3872e35913d req-758bfa9b-349f-4be2-a213-3ab90c259669 service nova] [instance: 8ff6fbb9-c90f-498d-9a85-d220a8c2f794] Neutron deleted interface 8c324833-7bf9-4007-85aa-5b16871f63c7; detaching it from the instance and deleting it from the info cache [ 965.761938] env[62066]: DEBUG nova.network.neutron [req-64fc33f1-68cd-4f99-8edc-c3872e35913d req-758bfa9b-349f-4be2-a213-3ab90c259669 service nova] [instance: 8ff6fbb9-c90f-498d-9a85-d220a8c2f794] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 965.774169] env[62066]: DEBUG oslo_vmware.api [None req-41711bcc-fddb-4aba-b7e4-374d02404b25 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Task: {'id': task-1341240, 'name': PowerOffVM_Task, 'duration_secs': 0.334486} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.774956] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-41711bcc-fddb-4aba-b7e4-374d02404b25 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: c8e0d47c-4421-4e00-9183-206fceeabc40] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 965.775294] env[62066]: DEBUG nova.compute.manager [None req-41711bcc-fddb-4aba-b7e4-374d02404b25 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: c8e0d47c-4421-4e00-9183-206fceeabc40] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 965.776627] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-022427e2-1fcf-44b6-ba51-6f47009686bc {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.787961] env[62066]: DEBUG oslo_concurrency.lockutils [None req-96a3683c-7d82-48dc-bf96-f0467366c2eb tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Lock "interface-b718b09b-a7fa-445b-8be3-5b3ebca210a2-26c5cd4b-0e0d-4e99-a7fe-7a46764dc642" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 9.633s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 965.919352] env[62066]: DEBUG nova.compute.manager [None req-3dd8531f-78c0-4fc2-af15-28c19b0a2bf2 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 09e3ae18-1517-470d-8cc9-a8b50ee774c7] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 965.945335] env[62066]: DEBUG nova.virt.hardware [None req-3dd8531f-78c0-4fc2-af15-28c19b0a2bf2 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-23T13:40:41Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-23T13:40:23Z,direct_url=,disk_format='vmdk',id=50ff584c-3b50-4395-af07-3e66769bc9f7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='eb52f7069a374c61ae946f052007c6d9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-23T13:40:24Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 965.945596] env[62066]: DEBUG nova.virt.hardware [None req-3dd8531f-78c0-4fc2-af15-28c19b0a2bf2 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 965.945758] env[62066]: DEBUG nova.virt.hardware [None req-3dd8531f-78c0-4fc2-af15-28c19b0a2bf2 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 965.945945] env[62066]: DEBUG nova.virt.hardware [None req-3dd8531f-78c0-4fc2-af15-28c19b0a2bf2 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 965.946121] env[62066]: DEBUG nova.virt.hardware [None req-3dd8531f-78c0-4fc2-af15-28c19b0a2bf2 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 965.946279] env[62066]: DEBUG nova.virt.hardware [None req-3dd8531f-78c0-4fc2-af15-28c19b0a2bf2 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 965.946489] env[62066]: DEBUG nova.virt.hardware [None req-3dd8531f-78c0-4fc2-af15-28c19b0a2bf2 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 965.946652] env[62066]: DEBUG nova.virt.hardware [None req-3dd8531f-78c0-4fc2-af15-28c19b0a2bf2 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 965.946823] env[62066]: DEBUG nova.virt.hardware [None req-3dd8531f-78c0-4fc2-af15-28c19b0a2bf2 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 965.946989] env[62066]: DEBUG nova.virt.hardware [None req-3dd8531f-78c0-4fc2-af15-28c19b0a2bf2 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 965.947185] env[62066]: DEBUG nova.virt.hardware [None req-3dd8531f-78c0-4fc2-af15-28c19b0a2bf2 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 965.948059] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3026e01a-9026-40c1-80c3-96a7ff745d2a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.956690] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e1e3a07-cf6a-4c83-992d-3a8becf534d2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.126465] env[62066]: DEBUG oslo_concurrency.lockutils [None req-26ece0ef-0448-44a4-8d3a-1a89aaf96689 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.238s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 966.126697] env[62066]: INFO nova.compute.manager [None req-26ece0ef-0448-44a4-8d3a-1a89aaf96689 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: ca504be5-209c-456c-af00-d403bfa2d634] Migrating [ 966.133141] env[62066]: DEBUG oslo_concurrency.lockutils [None req-bfde70b7-abcf-4102-a176-bfc4580a84a1 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.353s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 966.133362] env[62066]: DEBUG nova.objects.instance [None req-bfde70b7-abcf-4102-a176-bfc4580a84a1 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Lazy-loading 'resources' on Instance uuid 5693eea2-1cac-45d5-aa27-58526b65136e {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 966.188220] env[62066]: DEBUG oslo_vmware.api [req-2e69dbda-bb70-42d3-b5fe-0025cbfc75b7 req-f2890efd-3684-49c2-9b90-7186dc657fa2 service nova] Task: {'id': task-1341237, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.235773] env[62066]: DEBUG nova.network.neutron [-] [instance: 8ff6fbb9-c90f-498d-9a85-d220a8c2f794] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 966.264641] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4c68cd06-2d7e-4005-a6fa-676fa3bba183 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.274930] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e29078b-93d8-4be4-bc4a-250ff95dfa30 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.293200] env[62066]: DEBUG oslo_concurrency.lockutils [None req-41711bcc-fddb-4aba-b7e4-374d02404b25 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Lock "c8e0d47c-4421-4e00-9183-206fceeabc40" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.556s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 966.307485] env[62066]: DEBUG nova.compute.manager [req-64fc33f1-68cd-4f99-8edc-c3872e35913d req-758bfa9b-349f-4be2-a213-3ab90c259669 service nova] [instance: 8ff6fbb9-c90f-498d-9a85-d220a8c2f794] Detach interface failed, port_id=8c324833-7bf9-4007-85aa-5b16871f63c7, reason: Instance 8ff6fbb9-c90f-498d-9a85-d220a8c2f794 could not be found. {{(pid=62066) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 966.645236] env[62066]: DEBUG oslo_concurrency.lockutils [None req-26ece0ef-0448-44a4-8d3a-1a89aaf96689 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Acquiring lock "refresh_cache-ca504be5-209c-456c-af00-d403bfa2d634" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 966.645491] env[62066]: DEBUG oslo_concurrency.lockutils [None req-26ece0ef-0448-44a4-8d3a-1a89aaf96689 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Acquired lock "refresh_cache-ca504be5-209c-456c-af00-d403bfa2d634" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 966.645682] env[62066]: DEBUG nova.network.neutron [None req-26ece0ef-0448-44a4-8d3a-1a89aaf96689 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: ca504be5-209c-456c-af00-d403bfa2d634] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 966.690244] env[62066]: DEBUG oslo_vmware.api [req-2e69dbda-bb70-42d3-b5fe-0025cbfc75b7 req-f2890efd-3684-49c2-9b90-7186dc657fa2 service nova] Task: {'id': task-1341237, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.711688] env[62066]: DEBUG nova.network.neutron [None req-3dd8531f-78c0-4fc2-af15-28c19b0a2bf2 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 09e3ae18-1517-470d-8cc9-a8b50ee774c7] Successfully updated port: f62ec0e9-6137-4099-8d1b-62a38a3a9938 {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 966.738780] env[62066]: INFO nova.compute.manager [-] [instance: 8ff6fbb9-c90f-498d-9a85-d220a8c2f794] Took 1.53 seconds to deallocate network for instance. [ 966.790833] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02698791-8327-413f-81ab-78e39d8a7695 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.797190] env[62066]: DEBUG nova.compute.manager [req-5aa53442-aca7-47d9-b3d0-f3563081bf98 req-6ec7d12b-b8e8-479b-a3ad-87fb88a0d4db service nova] [instance: 09e3ae18-1517-470d-8cc9-a8b50ee774c7] Received event network-vif-plugged-f62ec0e9-6137-4099-8d1b-62a38a3a9938 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 966.797415] env[62066]: DEBUG oslo_concurrency.lockutils [req-5aa53442-aca7-47d9-b3d0-f3563081bf98 req-6ec7d12b-b8e8-479b-a3ad-87fb88a0d4db service nova] Acquiring lock "09e3ae18-1517-470d-8cc9-a8b50ee774c7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 966.797625] env[62066]: DEBUG oslo_concurrency.lockutils [req-5aa53442-aca7-47d9-b3d0-f3563081bf98 req-6ec7d12b-b8e8-479b-a3ad-87fb88a0d4db service nova] Lock "09e3ae18-1517-470d-8cc9-a8b50ee774c7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 966.797795] env[62066]: DEBUG oslo_concurrency.lockutils [req-5aa53442-aca7-47d9-b3d0-f3563081bf98 req-6ec7d12b-b8e8-479b-a3ad-87fb88a0d4db service nova] Lock "09e3ae18-1517-470d-8cc9-a8b50ee774c7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 966.797967] env[62066]: DEBUG nova.compute.manager [req-5aa53442-aca7-47d9-b3d0-f3563081bf98 req-6ec7d12b-b8e8-479b-a3ad-87fb88a0d4db service nova] [instance: 09e3ae18-1517-470d-8cc9-a8b50ee774c7] No waiting events found dispatching network-vif-plugged-f62ec0e9-6137-4099-8d1b-62a38a3a9938 {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 966.798152] env[62066]: WARNING nova.compute.manager [req-5aa53442-aca7-47d9-b3d0-f3563081bf98 req-6ec7d12b-b8e8-479b-a3ad-87fb88a0d4db service nova] [instance: 09e3ae18-1517-470d-8cc9-a8b50ee774c7] Received unexpected event network-vif-plugged-f62ec0e9-6137-4099-8d1b-62a38a3a9938 for instance with vm_state building and task_state spawning. [ 966.798337] env[62066]: DEBUG nova.compute.manager [req-5aa53442-aca7-47d9-b3d0-f3563081bf98 req-6ec7d12b-b8e8-479b-a3ad-87fb88a0d4db service nova] [instance: 09e3ae18-1517-470d-8cc9-a8b50ee774c7] Received event network-changed-f62ec0e9-6137-4099-8d1b-62a38a3a9938 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 966.798539] env[62066]: DEBUG nova.compute.manager [req-5aa53442-aca7-47d9-b3d0-f3563081bf98 req-6ec7d12b-b8e8-479b-a3ad-87fb88a0d4db service nova] [instance: 09e3ae18-1517-470d-8cc9-a8b50ee774c7] Refreshing instance network info cache due to event network-changed-f62ec0e9-6137-4099-8d1b-62a38a3a9938. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 966.798688] env[62066]: DEBUG oslo_concurrency.lockutils [req-5aa53442-aca7-47d9-b3d0-f3563081bf98 req-6ec7d12b-b8e8-479b-a3ad-87fb88a0d4db service nova] Acquiring lock "refresh_cache-09e3ae18-1517-470d-8cc9-a8b50ee774c7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 966.798828] env[62066]: DEBUG oslo_concurrency.lockutils [req-5aa53442-aca7-47d9-b3d0-f3563081bf98 req-6ec7d12b-b8e8-479b-a3ad-87fb88a0d4db service nova] Acquired lock "refresh_cache-09e3ae18-1517-470d-8cc9-a8b50ee774c7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 966.798986] env[62066]: DEBUG nova.network.neutron [req-5aa53442-aca7-47d9-b3d0-f3563081bf98 req-6ec7d12b-b8e8-479b-a3ad-87fb88a0d4db service nova] [instance: 09e3ae18-1517-470d-8cc9-a8b50ee774c7] Refreshing network info cache for port f62ec0e9-6137-4099-8d1b-62a38a3a9938 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 966.803236] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d18d423f-db72-4c3c-b5a3-4953a8f33614 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.832926] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c522dbed-e62f-4b71-80a8-df0346ddb8aa {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.840140] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-622d29ec-bb0f-4455-9c1c-760f8e378257 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.853367] env[62066]: DEBUG nova.compute.provider_tree [None req-bfde70b7-abcf-4102-a176-bfc4580a84a1 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 967.191846] env[62066]: DEBUG oslo_vmware.api [req-2e69dbda-bb70-42d3-b5fe-0025cbfc75b7 req-f2890efd-3684-49c2-9b90-7186dc657fa2 service nova] Task: {'id': task-1341237, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.213644] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3dd8531f-78c0-4fc2-af15-28c19b0a2bf2 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Acquiring lock "refresh_cache-09e3ae18-1517-470d-8cc9-a8b50ee774c7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 967.239820] env[62066]: DEBUG nova.objects.instance [None req-7e06b327-3961-4087-9770-cd5ec70f0e21 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Lazy-loading 'flavor' on Instance uuid c8e0d47c-4421-4e00-9183-206fceeabc40 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 967.249652] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ea8fb2df-bddb-407b-bcac-dde5dd08fd22 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 967.343326] env[62066]: DEBUG nova.network.neutron [req-5aa53442-aca7-47d9-b3d0-f3563081bf98 req-6ec7d12b-b8e8-479b-a3ad-87fb88a0d4db service nova] [instance: 09e3ae18-1517-470d-8cc9-a8b50ee774c7] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 967.356537] env[62066]: DEBUG nova.scheduler.client.report [None req-bfde70b7-abcf-4102-a176-bfc4580a84a1 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 967.413358] env[62066]: DEBUG nova.network.neutron [None req-26ece0ef-0448-44a4-8d3a-1a89aaf96689 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: ca504be5-209c-456c-af00-d403bfa2d634] Updating instance_info_cache with network_info: [{"id": "11deacd3-d3a5-49a4-98f3-3226eb532e76", "address": "fa:16:3e:3a:b5:74", "network": {"id": "1c6c51e4-96f8-4175-b7f5-96b8a42fcb13", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-2006089809-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d609babc987a42e2a8ddb4bfb9c3b103", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a965790c-2d2f-4c2a-9ee7-745f4d53039b", "external-id": "nsx-vlan-transportzone-708", "segmentation_id": 708, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap11deacd3-d3", "ovs_interfaceid": "11deacd3-d3a5-49a4-98f3-3226eb532e76", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 967.438772] env[62066]: DEBUG nova.network.neutron [req-5aa53442-aca7-47d9-b3d0-f3563081bf98 req-6ec7d12b-b8e8-479b-a3ad-87fb88a0d4db service nova] [instance: 09e3ae18-1517-470d-8cc9-a8b50ee774c7] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 967.690237] env[62066]: DEBUG oslo_vmware.api [req-2e69dbda-bb70-42d3-b5fe-0025cbfc75b7 req-f2890efd-3684-49c2-9b90-7186dc657fa2 service nova] Task: {'id': task-1341237, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.746057] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7e06b327-3961-4087-9770-cd5ec70f0e21 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Acquiring lock "refresh_cache-c8e0d47c-4421-4e00-9183-206fceeabc40" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 967.746057] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7e06b327-3961-4087-9770-cd5ec70f0e21 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Acquired lock "refresh_cache-c8e0d47c-4421-4e00-9183-206fceeabc40" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 967.746299] env[62066]: DEBUG nova.network.neutron [None req-7e06b327-3961-4087-9770-cd5ec70f0e21 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: c8e0d47c-4421-4e00-9183-206fceeabc40] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 967.746501] env[62066]: DEBUG nova.objects.instance [None req-7e06b327-3961-4087-9770-cd5ec70f0e21 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Lazy-loading 'info_cache' on Instance uuid c8e0d47c-4421-4e00-9183-206fceeabc40 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 967.865033] env[62066]: DEBUG oslo_concurrency.lockutils [None req-bfde70b7-abcf-4102-a176-bfc4580a84a1 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.732s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 967.867395] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ea8fb2df-bddb-407b-bcac-dde5dd08fd22 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.618s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 967.867635] env[62066]: DEBUG nova.objects.instance [None req-ea8fb2df-bddb-407b-bcac-dde5dd08fd22 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Lazy-loading 'resources' on Instance uuid 8ff6fbb9-c90f-498d-9a85-d220a8c2f794 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 967.889263] env[62066]: INFO nova.scheduler.client.report [None req-bfde70b7-abcf-4102-a176-bfc4580a84a1 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Deleted allocations for instance 5693eea2-1cac-45d5-aa27-58526b65136e [ 967.918444] env[62066]: DEBUG oslo_concurrency.lockutils [None req-26ece0ef-0448-44a4-8d3a-1a89aaf96689 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Releasing lock "refresh_cache-ca504be5-209c-456c-af00-d403bfa2d634" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 967.940896] env[62066]: DEBUG oslo_concurrency.lockutils [req-5aa53442-aca7-47d9-b3d0-f3563081bf98 req-6ec7d12b-b8e8-479b-a3ad-87fb88a0d4db service nova] Releasing lock "refresh_cache-09e3ae18-1517-470d-8cc9-a8b50ee774c7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 967.941285] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3dd8531f-78c0-4fc2-af15-28c19b0a2bf2 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Acquired lock "refresh_cache-09e3ae18-1517-470d-8cc9-a8b50ee774c7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 967.941447] env[62066]: DEBUG nova.network.neutron [None req-3dd8531f-78c0-4fc2-af15-28c19b0a2bf2 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 09e3ae18-1517-470d-8cc9-a8b50ee774c7] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 968.190844] env[62066]: DEBUG oslo_vmware.api [req-2e69dbda-bb70-42d3-b5fe-0025cbfc75b7 req-f2890efd-3684-49c2-9b90-7186dc657fa2 service nova] Task: {'id': task-1341237, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.249888] env[62066]: DEBUG nova.objects.base [None req-7e06b327-3961-4087-9770-cd5ec70f0e21 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Object Instance lazy-loaded attributes: flavor,info_cache {{(pid=62066) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 968.397288] env[62066]: DEBUG oslo_concurrency.lockutils [None req-bfde70b7-abcf-4102-a176-bfc4580a84a1 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Lock "5693eea2-1cac-45d5-aa27-58526b65136e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.122s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 968.468104] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df4d50da-65d1-4091-a822-c2019d68d2d5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.472345] env[62066]: DEBUG nova.network.neutron [None req-3dd8531f-78c0-4fc2-af15-28c19b0a2bf2 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 09e3ae18-1517-470d-8cc9-a8b50ee774c7] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 968.476884] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50c2f0ab-c2e3-438c-a82f-af6e4d3d6000 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.511114] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a93825b-0cc4-44d8-ab9d-25ed7de242b4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.518883] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29bc4899-05a1-4634-a901-0a20252f5730 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.532313] env[62066]: DEBUG nova.compute.provider_tree [None req-ea8fb2df-bddb-407b-bcac-dde5dd08fd22 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 968.626335] env[62066]: DEBUG nova.network.neutron [None req-3dd8531f-78c0-4fc2-af15-28c19b0a2bf2 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 09e3ae18-1517-470d-8cc9-a8b50ee774c7] Updating instance_info_cache with network_info: [{"id": "f62ec0e9-6137-4099-8d1b-62a38a3a9938", "address": "fa:16:3e:3d:1c:9e", "network": {"id": "9f3396ce-ee48-4eaf-8cc7-3e1711f37537", "bridge": "br-int", "label": "tempest-ServersTestJSON-2129775093-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6497ab02f327476d8ff81c2ecc0371e5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7ab8d568-adb0-4f3b-b6cc-68413e6546ae", "external-id": "nsx-vlan-transportzone-86", "segmentation_id": 86, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf62ec0e9-61", "ovs_interfaceid": "f62ec0e9-6137-4099-8d1b-62a38a3a9938", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 968.691962] env[62066]: DEBUG oslo_vmware.api [req-2e69dbda-bb70-42d3-b5fe-0025cbfc75b7 req-f2890efd-3684-49c2-9b90-7186dc657fa2 service nova] Task: {'id': task-1341237, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.968508] env[62066]: DEBUG nova.network.neutron [None req-7e06b327-3961-4087-9770-cd5ec70f0e21 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: c8e0d47c-4421-4e00-9183-206fceeabc40] Updating instance_info_cache with network_info: [{"id": "bb9924e0-8b18-43b2-bda0-3e409fa16718", "address": "fa:16:3e:b1:e4:41", "network": {"id": "5b89b2d4-71dd-4cd1-9edb-f4040be26540", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-283918837-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.165", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2252d0f6f18a47fd8a1ddad8020a6db7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "352165bb-004f-4180-9627-3a275dbe18af", "external-id": "nsx-vlan-transportzone-926", "segmentation_id": 926, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbb9924e0-8b", "ovs_interfaceid": "bb9924e0-8b18-43b2-bda0-3e409fa16718", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 969.037225] env[62066]: DEBUG nova.scheduler.client.report [None req-ea8fb2df-bddb-407b-bcac-dde5dd08fd22 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 969.129230] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3dd8531f-78c0-4fc2-af15-28c19b0a2bf2 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Releasing lock "refresh_cache-09e3ae18-1517-470d-8cc9-a8b50ee774c7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 969.129554] env[62066]: DEBUG nova.compute.manager [None req-3dd8531f-78c0-4fc2-af15-28c19b0a2bf2 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 09e3ae18-1517-470d-8cc9-a8b50ee774c7] Instance network_info: |[{"id": "f62ec0e9-6137-4099-8d1b-62a38a3a9938", "address": "fa:16:3e:3d:1c:9e", "network": {"id": "9f3396ce-ee48-4eaf-8cc7-3e1711f37537", "bridge": "br-int", "label": "tempest-ServersTestJSON-2129775093-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6497ab02f327476d8ff81c2ecc0371e5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7ab8d568-adb0-4f3b-b6cc-68413e6546ae", "external-id": "nsx-vlan-transportzone-86", "segmentation_id": 86, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf62ec0e9-61", "ovs_interfaceid": "f62ec0e9-6137-4099-8d1b-62a38a3a9938", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 969.129997] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-3dd8531f-78c0-4fc2-af15-28c19b0a2bf2 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 09e3ae18-1517-470d-8cc9-a8b50ee774c7] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:3d:1c:9e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7ab8d568-adb0-4f3b-b6cc-68413e6546ae', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f62ec0e9-6137-4099-8d1b-62a38a3a9938', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 969.137690] env[62066]: DEBUG oslo.service.loopingcall [None req-3dd8531f-78c0-4fc2-af15-28c19b0a2bf2 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 969.137908] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 09e3ae18-1517-470d-8cc9-a8b50ee774c7] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 969.138145] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-27269b96-eaa2-46bd-8350-39afff1c8453 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.158896] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 969.158896] env[62066]: value = "task-1341241" [ 969.158896] env[62066]: _type = "Task" [ 969.158896] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.166488] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1341241, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.192196] env[62066]: DEBUG oslo_vmware.api [req-2e69dbda-bb70-42d3-b5fe-0025cbfc75b7 req-f2890efd-3684-49c2-9b90-7186dc657fa2 service nova] Task: {'id': task-1341237, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.432835] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d68692f2-3d91-4d76-945c-47eb9393478a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.453224] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-26ece0ef-0448-44a4-8d3a-1a89aaf96689 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: ca504be5-209c-456c-af00-d403bfa2d634] Updating instance 'ca504be5-209c-456c-af00-d403bfa2d634' progress to 0 {{(pid=62066) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 969.474429] env[62066]: DEBUG oslo_concurrency.lockutils [None req-7e06b327-3961-4087-9770-cd5ec70f0e21 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Releasing lock "refresh_cache-c8e0d47c-4421-4e00-9183-206fceeabc40" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 969.543354] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ea8fb2df-bddb-407b-bcac-dde5dd08fd22 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.675s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 969.559112] env[62066]: INFO nova.scheduler.client.report [None req-ea8fb2df-bddb-407b-bcac-dde5dd08fd22 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Deleted allocations for instance 8ff6fbb9-c90f-498d-9a85-d220a8c2f794 [ 969.668599] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1341241, 'name': CreateVM_Task, 'duration_secs': 0.273523} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.668782] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 09e3ae18-1517-470d-8cc9-a8b50ee774c7] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 969.669494] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3dd8531f-78c0-4fc2-af15-28c19b0a2bf2 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 969.669688] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3dd8531f-78c0-4fc2-af15-28c19b0a2bf2 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Acquired lock "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 969.670015] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3dd8531f-78c0-4fc2-af15-28c19b0a2bf2 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 969.670539] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c39276ae-c5b8-4c56-8dcd-0c84862a378c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.674524] env[62066]: DEBUG oslo_vmware.api [None req-3dd8531f-78c0-4fc2-af15-28c19b0a2bf2 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Waiting for the task: (returnval){ [ 969.674524] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]529c02b9-cbe3-4c7a-9161-2d864ff59306" [ 969.674524] env[62066]: _type = "Task" [ 969.674524] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.681670] env[62066]: DEBUG oslo_vmware.api [None req-3dd8531f-78c0-4fc2-af15-28c19b0a2bf2 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]529c02b9-cbe3-4c7a-9161-2d864ff59306, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.690571] env[62066]: DEBUG oslo_vmware.api [req-2e69dbda-bb70-42d3-b5fe-0025cbfc75b7 req-f2890efd-3684-49c2-9b90-7186dc657fa2 service nova] Task: {'id': task-1341237, 'name': ReconfigVM_Task} progress is 18%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.957905] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-26ece0ef-0448-44a4-8d3a-1a89aaf96689 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: ca504be5-209c-456c-af00-d403bfa2d634] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 969.958225] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b41bdc4d-2fde-4438-804a-4e1e0d715411 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.965180] env[62066]: DEBUG oslo_vmware.api [None req-26ece0ef-0448-44a4-8d3a-1a89aaf96689 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Waiting for the task: (returnval){ [ 969.965180] env[62066]: value = "task-1341242" [ 969.965180] env[62066]: _type = "Task" [ 969.965180] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.973147] env[62066]: DEBUG oslo_vmware.api [None req-26ece0ef-0448-44a4-8d3a-1a89aaf96689 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1341242, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.977055] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e06b327-3961-4087-9770-cd5ec70f0e21 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: c8e0d47c-4421-4e00-9183-206fceeabc40] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 969.977347] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-73506b45-9ced-4a3f-b2e5-1370c31caf28 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.983024] env[62066]: DEBUG oslo_vmware.api [None req-7e06b327-3961-4087-9770-cd5ec70f0e21 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Waiting for the task: (returnval){ [ 969.983024] env[62066]: value = "task-1341243" [ 969.983024] env[62066]: _type = "Task" [ 969.983024] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.990404] env[62066]: DEBUG oslo_vmware.api [None req-7e06b327-3961-4087-9770-cd5ec70f0e21 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Task: {'id': task-1341243, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.067828] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ea8fb2df-bddb-407b-bcac-dde5dd08fd22 tempest-AttachVolumeShelveTestJSON-1753114425 tempest-AttachVolumeShelveTestJSON-1753114425-project-member] Lock "8ff6fbb9-c90f-498d-9a85-d220a8c2f794" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.010s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 970.079323] env[62066]: DEBUG oslo_concurrency.lockutils [None req-20e007b5-b757-49e8-8380-d9dd27ea229d tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Acquiring lock "a2c5277e-af49-4b31-8480-a74d354c7383" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 970.079570] env[62066]: DEBUG oslo_concurrency.lockutils [None req-20e007b5-b757-49e8-8380-d9dd27ea229d tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Lock "a2c5277e-af49-4b31-8480-a74d354c7383" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 970.184977] env[62066]: DEBUG oslo_vmware.api [None req-3dd8531f-78c0-4fc2-af15-28c19b0a2bf2 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]529c02b9-cbe3-4c7a-9161-2d864ff59306, 'name': SearchDatastore_Task, 'duration_secs': 0.008849} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.188045] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3dd8531f-78c0-4fc2-af15-28c19b0a2bf2 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Releasing lock "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 970.188312] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-3dd8531f-78c0-4fc2-af15-28c19b0a2bf2 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 09e3ae18-1517-470d-8cc9-a8b50ee774c7] Processing image 50ff584c-3b50-4395-af07-3e66769bc9f7 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 970.188580] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3dd8531f-78c0-4fc2-af15-28c19b0a2bf2 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 970.188739] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3dd8531f-78c0-4fc2-af15-28c19b0a2bf2 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Acquired lock "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 970.188923] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-3dd8531f-78c0-4fc2-af15-28c19b0a2bf2 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 970.189210] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-fe49c187-5b97-4b67-b774-9d297ff90d9b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.196078] env[62066]: DEBUG oslo_vmware.api [req-2e69dbda-bb70-42d3-b5fe-0025cbfc75b7 req-f2890efd-3684-49c2-9b90-7186dc657fa2 service nova] Task: {'id': task-1341237, 'name': ReconfigVM_Task, 'duration_secs': 5.723971} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.197234] env[62066]: DEBUG oslo_concurrency.lockutils [req-2e69dbda-bb70-42d3-b5fe-0025cbfc75b7 req-f2890efd-3684-49c2-9b90-7186dc657fa2 service nova] Releasing lock "b718b09b-a7fa-445b-8be3-5b3ebca210a2" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 970.197553] env[62066]: DEBUG nova.virt.vmwareapi.vmops [req-2e69dbda-bb70-42d3-b5fe-0025cbfc75b7 req-f2890efd-3684-49c2-9b90-7186dc657fa2 service nova] [instance: b718b09b-a7fa-445b-8be3-5b3ebca210a2] Reconfigured VM to detach interface {{(pid=62066) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 970.198240] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-3dd8531f-78c0-4fc2-af15-28c19b0a2bf2 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 970.198439] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-3dd8531f-78c0-4fc2-af15-28c19b0a2bf2 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 970.199171] env[62066]: DEBUG oslo_concurrency.lockutils [None req-2552ba89-1024-40da-89b2-2a360324a01f tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Lock "b718b09b-a7fa-445b-8be3-5b3ebca210a2" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 6.028s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 970.199387] env[62066]: DEBUG oslo_concurrency.lockutils [None req-2552ba89-1024-40da-89b2-2a360324a01f tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Acquiring lock "b718b09b-a7fa-445b-8be3-5b3ebca210a2-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 970.199586] env[62066]: DEBUG oslo_concurrency.lockutils [None req-2552ba89-1024-40da-89b2-2a360324a01f tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Lock "b718b09b-a7fa-445b-8be3-5b3ebca210a2-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 970.199755] env[62066]: DEBUG oslo_concurrency.lockutils [None req-2552ba89-1024-40da-89b2-2a360324a01f tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Lock "b718b09b-a7fa-445b-8be3-5b3ebca210a2-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 970.201368] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c81b0606-d718-438e-aeaf-166e2d7639ac {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.203646] env[62066]: INFO nova.compute.manager [None req-2552ba89-1024-40da-89b2-2a360324a01f tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: b718b09b-a7fa-445b-8be3-5b3ebca210a2] Terminating instance [ 970.205573] env[62066]: DEBUG nova.compute.manager [None req-2552ba89-1024-40da-89b2-2a360324a01f tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: b718b09b-a7fa-445b-8be3-5b3ebca210a2] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 970.205766] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-2552ba89-1024-40da-89b2-2a360324a01f tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: b718b09b-a7fa-445b-8be3-5b3ebca210a2] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 970.206497] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae31ba66-c8bf-4c19-a93b-671613204b47 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.209832] env[62066]: DEBUG oslo_vmware.api [None req-3dd8531f-78c0-4fc2-af15-28c19b0a2bf2 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Waiting for the task: (returnval){ [ 970.209832] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]52203546-7b40-ac3c-3813-97f6c3bbe077" [ 970.209832] env[62066]: _type = "Task" [ 970.209832] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.215038] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-2552ba89-1024-40da-89b2-2a360324a01f tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: b718b09b-a7fa-445b-8be3-5b3ebca210a2] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 970.215523] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c4acc8b4-84eb-476b-8067-84af577b7480 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.220465] env[62066]: DEBUG oslo_vmware.api [None req-3dd8531f-78c0-4fc2-af15-28c19b0a2bf2 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52203546-7b40-ac3c-3813-97f6c3bbe077, 'name': SearchDatastore_Task, 'duration_secs': 0.008115} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.222032] env[62066]: DEBUG oslo_vmware.api [None req-2552ba89-1024-40da-89b2-2a360324a01f tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Waiting for the task: (returnval){ [ 970.222032] env[62066]: value = "task-1341244" [ 970.222032] env[62066]: _type = "Task" [ 970.222032] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.222222] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4a5ae932-dee6-43de-bb6e-3d40b5d3080d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.229551] env[62066]: DEBUG oslo_vmware.api [None req-3dd8531f-78c0-4fc2-af15-28c19b0a2bf2 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Waiting for the task: (returnval){ [ 970.229551] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]528b65e7-7211-c0e6-2b1c-ef6cf058c302" [ 970.229551] env[62066]: _type = "Task" [ 970.229551] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.232635] env[62066]: DEBUG oslo_vmware.api [None req-2552ba89-1024-40da-89b2-2a360324a01f tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': task-1341244, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.239979] env[62066]: DEBUG oslo_vmware.api [None req-3dd8531f-78c0-4fc2-af15-28c19b0a2bf2 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]528b65e7-7211-c0e6-2b1c-ef6cf058c302, 'name': SearchDatastore_Task, 'duration_secs': 0.008325} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.240240] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3dd8531f-78c0-4fc2-af15-28c19b0a2bf2 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Releasing lock "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 970.240516] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-3dd8531f-78c0-4fc2-af15-28c19b0a2bf2 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk to [datastore1] 09e3ae18-1517-470d-8cc9-a8b50ee774c7/09e3ae18-1517-470d-8cc9-a8b50ee774c7.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 970.240758] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e33baff0-b94b-45c8-8b2d-f14d9d0bddc9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.247157] env[62066]: DEBUG oslo_vmware.api [None req-3dd8531f-78c0-4fc2-af15-28c19b0a2bf2 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Waiting for the task: (returnval){ [ 970.247157] env[62066]: value = "task-1341245" [ 970.247157] env[62066]: _type = "Task" [ 970.247157] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.254321] env[62066]: DEBUG oslo_vmware.api [None req-3dd8531f-78c0-4fc2-af15-28c19b0a2bf2 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': task-1341245, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.478032] env[62066]: DEBUG oslo_vmware.api [None req-26ece0ef-0448-44a4-8d3a-1a89aaf96689 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1341242, 'name': PowerOffVM_Task, 'duration_secs': 0.205228} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.478480] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-26ece0ef-0448-44a4-8d3a-1a89aaf96689 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: ca504be5-209c-456c-af00-d403bfa2d634] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 970.478755] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-26ece0ef-0448-44a4-8d3a-1a89aaf96689 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: ca504be5-209c-456c-af00-d403bfa2d634] Updating instance 'ca504be5-209c-456c-af00-d403bfa2d634' progress to 17 {{(pid=62066) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 970.493766] env[62066]: DEBUG oslo_vmware.api [None req-7e06b327-3961-4087-9770-cd5ec70f0e21 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Task: {'id': task-1341243, 'name': PowerOnVM_Task, 'duration_secs': 0.401803} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.494306] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e06b327-3961-4087-9770-cd5ec70f0e21 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: c8e0d47c-4421-4e00-9183-206fceeabc40] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 970.494306] env[62066]: DEBUG nova.compute.manager [None req-7e06b327-3961-4087-9770-cd5ec70f0e21 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: c8e0d47c-4421-4e00-9183-206fceeabc40] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 970.495124] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d2b5d47-e933-4623-984a-5b83ba03295a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.582040] env[62066]: DEBUG nova.compute.manager [None req-20e007b5-b757-49e8-8380-d9dd27ea229d tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: a2c5277e-af49-4b31-8480-a74d354c7383] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 970.734560] env[62066]: DEBUG oslo_vmware.api [None req-2552ba89-1024-40da-89b2-2a360324a01f tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': task-1341244, 'name': PowerOffVM_Task, 'duration_secs': 0.197413} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.735655] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-2552ba89-1024-40da-89b2-2a360324a01f tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: b718b09b-a7fa-445b-8be3-5b3ebca210a2] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 970.735655] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-2552ba89-1024-40da-89b2-2a360324a01f tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: b718b09b-a7fa-445b-8be3-5b3ebca210a2] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 970.735655] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1536b0ed-7872-4012-8b98-9e96a030d5f6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.757510] env[62066]: DEBUG oslo_vmware.api [None req-3dd8531f-78c0-4fc2-af15-28c19b0a2bf2 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': task-1341245, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.494829} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.757815] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-3dd8531f-78c0-4fc2-af15-28c19b0a2bf2 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk to [datastore1] 09e3ae18-1517-470d-8cc9-a8b50ee774c7/09e3ae18-1517-470d-8cc9-a8b50ee774c7.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 970.758057] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-3dd8531f-78c0-4fc2-af15-28c19b0a2bf2 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 09e3ae18-1517-470d-8cc9-a8b50ee774c7] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 970.758331] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6990c81f-aff2-4e2b-bdcc-eb16ed776bc5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.764449] env[62066]: DEBUG oslo_vmware.api [None req-3dd8531f-78c0-4fc2-af15-28c19b0a2bf2 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Waiting for the task: (returnval){ [ 970.764449] env[62066]: value = "task-1341247" [ 970.764449] env[62066]: _type = "Task" [ 970.764449] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.771950] env[62066]: DEBUG oslo_vmware.api [None req-3dd8531f-78c0-4fc2-af15-28c19b0a2bf2 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': task-1341247, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.819981] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-2552ba89-1024-40da-89b2-2a360324a01f tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: b718b09b-a7fa-445b-8be3-5b3ebca210a2] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 970.820323] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-2552ba89-1024-40da-89b2-2a360324a01f tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: b718b09b-a7fa-445b-8be3-5b3ebca210a2] Deleting contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 970.820593] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-2552ba89-1024-40da-89b2-2a360324a01f tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Deleting the datastore file [datastore2] b718b09b-a7fa-445b-8be3-5b3ebca210a2 {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 970.820939] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-517f25bd-772c-445f-a10a-c3625cb3e355 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.827783] env[62066]: DEBUG oslo_vmware.api [None req-2552ba89-1024-40da-89b2-2a360324a01f tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Waiting for the task: (returnval){ [ 970.827783] env[62066]: value = "task-1341248" [ 970.827783] env[62066]: _type = "Task" [ 970.827783] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.835995] env[62066]: DEBUG oslo_vmware.api [None req-2552ba89-1024-40da-89b2-2a360324a01f tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': task-1341248, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.985334] env[62066]: DEBUG nova.virt.hardware [None req-26ece0ef-0448-44a4-8d3a-1a89aaf96689 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-23T13:40:43Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=50ff584c-3b50-4395-af07-3e66769bc9f7,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 970.985669] env[62066]: DEBUG nova.virt.hardware [None req-26ece0ef-0448-44a4-8d3a-1a89aaf96689 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 970.985953] env[62066]: DEBUG nova.virt.hardware [None req-26ece0ef-0448-44a4-8d3a-1a89aaf96689 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 970.986216] env[62066]: DEBUG nova.virt.hardware [None req-26ece0ef-0448-44a4-8d3a-1a89aaf96689 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 970.986465] env[62066]: DEBUG nova.virt.hardware [None req-26ece0ef-0448-44a4-8d3a-1a89aaf96689 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 970.986698] env[62066]: DEBUG nova.virt.hardware [None req-26ece0ef-0448-44a4-8d3a-1a89aaf96689 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 970.987019] env[62066]: DEBUG nova.virt.hardware [None req-26ece0ef-0448-44a4-8d3a-1a89aaf96689 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 970.987300] env[62066]: DEBUG nova.virt.hardware [None req-26ece0ef-0448-44a4-8d3a-1a89aaf96689 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 970.987512] env[62066]: DEBUG nova.virt.hardware [None req-26ece0ef-0448-44a4-8d3a-1a89aaf96689 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 970.987705] env[62066]: DEBUG nova.virt.hardware [None req-26ece0ef-0448-44a4-8d3a-1a89aaf96689 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 970.987893] env[62066]: DEBUG nova.virt.hardware [None req-26ece0ef-0448-44a4-8d3a-1a89aaf96689 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 970.992901] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-625b6253-d28e-4fa4-a022-06b47e3de61c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.011895] env[62066]: DEBUG oslo_vmware.api [None req-26ece0ef-0448-44a4-8d3a-1a89aaf96689 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Waiting for the task: (returnval){ [ 971.011895] env[62066]: value = "task-1341249" [ 971.011895] env[62066]: _type = "Task" [ 971.011895] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.022447] env[62066]: DEBUG oslo_vmware.api [None req-26ece0ef-0448-44a4-8d3a-1a89aaf96689 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1341249, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.105311] env[62066]: DEBUG oslo_concurrency.lockutils [None req-20e007b5-b757-49e8-8380-d9dd27ea229d tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 971.105595] env[62066]: DEBUG oslo_concurrency.lockutils [None req-20e007b5-b757-49e8-8380-d9dd27ea229d tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 971.107624] env[62066]: INFO nova.compute.claims [None req-20e007b5-b757-49e8-8380-d9dd27ea229d tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: a2c5277e-af49-4b31-8480-a74d354c7383] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 971.277216] env[62066]: DEBUG oslo_vmware.api [None req-3dd8531f-78c0-4fc2-af15-28c19b0a2bf2 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': task-1341247, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.060413} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.277530] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-3dd8531f-78c0-4fc2-af15-28c19b0a2bf2 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 09e3ae18-1517-470d-8cc9-a8b50ee774c7] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 971.278340] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64e1a6d4-7946-41d3-810c-fa9deb8166e5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.303725] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-3dd8531f-78c0-4fc2-af15-28c19b0a2bf2 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 09e3ae18-1517-470d-8cc9-a8b50ee774c7] Reconfiguring VM instance instance-00000060 to attach disk [datastore1] 09e3ae18-1517-470d-8cc9-a8b50ee774c7/09e3ae18-1517-470d-8cc9-a8b50ee774c7.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 971.303918] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cea7191a-115e-4901-a9ad-f46fa44d51ae {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.323241] env[62066]: DEBUG oslo_vmware.api [None req-3dd8531f-78c0-4fc2-af15-28c19b0a2bf2 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Waiting for the task: (returnval){ [ 971.323241] env[62066]: value = "task-1341250" [ 971.323241] env[62066]: _type = "Task" [ 971.323241] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.331345] env[62066]: DEBUG oslo_vmware.api [None req-3dd8531f-78c0-4fc2-af15-28c19b0a2bf2 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': task-1341250, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.338660] env[62066]: DEBUG oslo_vmware.api [None req-2552ba89-1024-40da-89b2-2a360324a01f tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': task-1341248, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.343348} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.338904] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-2552ba89-1024-40da-89b2-2a360324a01f tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 971.339105] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-2552ba89-1024-40da-89b2-2a360324a01f tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: b718b09b-a7fa-445b-8be3-5b3ebca210a2] Deleted contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 971.339287] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-2552ba89-1024-40da-89b2-2a360324a01f tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: b718b09b-a7fa-445b-8be3-5b3ebca210a2] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 971.339467] env[62066]: INFO nova.compute.manager [None req-2552ba89-1024-40da-89b2-2a360324a01f tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: b718b09b-a7fa-445b-8be3-5b3ebca210a2] Took 1.13 seconds to destroy the instance on the hypervisor. [ 971.339726] env[62066]: DEBUG oslo.service.loopingcall [None req-2552ba89-1024-40da-89b2-2a360324a01f tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 971.339929] env[62066]: DEBUG nova.compute.manager [-] [instance: b718b09b-a7fa-445b-8be3-5b3ebca210a2] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 971.340036] env[62066]: DEBUG nova.network.neutron [-] [instance: b718b09b-a7fa-445b-8be3-5b3ebca210a2] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 971.522508] env[62066]: DEBUG oslo_vmware.api [None req-26ece0ef-0448-44a4-8d3a-1a89aaf96689 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1341249, 'name': ReconfigVM_Task, 'duration_secs': 0.189453} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.522850] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-26ece0ef-0448-44a4-8d3a-1a89aaf96689 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: ca504be5-209c-456c-af00-d403bfa2d634] Updating instance 'ca504be5-209c-456c-af00-d403bfa2d634' progress to 33 {{(pid=62066) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 971.794516] env[62066]: DEBUG nova.compute.manager [req-cd68fa36-d2b6-44dd-bc72-088cadbecacb req-afd87fba-09c1-4ea8-b57a-431198a2c0ae service nova] [instance: b718b09b-a7fa-445b-8be3-5b3ebca210a2] Received event network-vif-deleted-3d01745e-0d9e-4bae-ba4d-242380edbab6 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 971.794726] env[62066]: INFO nova.compute.manager [req-cd68fa36-d2b6-44dd-bc72-088cadbecacb req-afd87fba-09c1-4ea8-b57a-431198a2c0ae service nova] [instance: b718b09b-a7fa-445b-8be3-5b3ebca210a2] Neutron deleted interface 3d01745e-0d9e-4bae-ba4d-242380edbab6; detaching it from the instance and deleting it from the info cache [ 971.794907] env[62066]: DEBUG nova.network.neutron [req-cd68fa36-d2b6-44dd-bc72-088cadbecacb req-afd87fba-09c1-4ea8-b57a-431198a2c0ae service nova] [instance: b718b09b-a7fa-445b-8be3-5b3ebca210a2] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 971.832798] env[62066]: DEBUG oslo_vmware.api [None req-3dd8531f-78c0-4fc2-af15-28c19b0a2bf2 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': task-1341250, 'name': ReconfigVM_Task, 'duration_secs': 0.277992} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.833099] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-3dd8531f-78c0-4fc2-af15-28c19b0a2bf2 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 09e3ae18-1517-470d-8cc9-a8b50ee774c7] Reconfigured VM instance instance-00000060 to attach disk [datastore1] 09e3ae18-1517-470d-8cc9-a8b50ee774c7/09e3ae18-1517-470d-8cc9-a8b50ee774c7.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 971.834034] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-37c5520b-42f2-4e7a-be11-be7ac6d8b9a1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.840373] env[62066]: DEBUG oslo_vmware.api [None req-3dd8531f-78c0-4fc2-af15-28c19b0a2bf2 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Waiting for the task: (returnval){ [ 971.840373] env[62066]: value = "task-1341251" [ 971.840373] env[62066]: _type = "Task" [ 971.840373] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.847765] env[62066]: DEBUG oslo_vmware.api [None req-3dd8531f-78c0-4fc2-af15-28c19b0a2bf2 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': task-1341251, 'name': Rename_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.030024] env[62066]: DEBUG nova.virt.hardware [None req-26ece0ef-0448-44a4-8d3a-1a89aaf96689 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-23T13:40:41Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=50ff584c-3b50-4395-af07-3e66769bc9f7,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 972.030024] env[62066]: DEBUG nova.virt.hardware [None req-26ece0ef-0448-44a4-8d3a-1a89aaf96689 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 972.030024] env[62066]: DEBUG nova.virt.hardware [None req-26ece0ef-0448-44a4-8d3a-1a89aaf96689 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 972.030236] env[62066]: DEBUG nova.virt.hardware [None req-26ece0ef-0448-44a4-8d3a-1a89aaf96689 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 972.030236] env[62066]: DEBUG nova.virt.hardware [None req-26ece0ef-0448-44a4-8d3a-1a89aaf96689 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 972.030371] env[62066]: DEBUG nova.virt.hardware [None req-26ece0ef-0448-44a4-8d3a-1a89aaf96689 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 972.030663] env[62066]: DEBUG nova.virt.hardware [None req-26ece0ef-0448-44a4-8d3a-1a89aaf96689 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 972.030831] env[62066]: DEBUG nova.virt.hardware [None req-26ece0ef-0448-44a4-8d3a-1a89aaf96689 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 972.030995] env[62066]: DEBUG nova.virt.hardware [None req-26ece0ef-0448-44a4-8d3a-1a89aaf96689 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 972.031179] env[62066]: DEBUG nova.virt.hardware [None req-26ece0ef-0448-44a4-8d3a-1a89aaf96689 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 972.031355] env[62066]: DEBUG nova.virt.hardware [None req-26ece0ef-0448-44a4-8d3a-1a89aaf96689 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 972.036773] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-26ece0ef-0448-44a4-8d3a-1a89aaf96689 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: ca504be5-209c-456c-af00-d403bfa2d634] Reconfiguring VM instance instance-0000005f to detach disk 2000 {{(pid=62066) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 972.037086] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-87ff6605-c553-4df3-a146-d7e5e61dcc97 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.057770] env[62066]: DEBUG oslo_vmware.api [None req-26ece0ef-0448-44a4-8d3a-1a89aaf96689 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Waiting for the task: (returnval){ [ 972.057770] env[62066]: value = "task-1341252" [ 972.057770] env[62066]: _type = "Task" [ 972.057770] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.066774] env[62066]: DEBUG oslo_vmware.api [None req-26ece0ef-0448-44a4-8d3a-1a89aaf96689 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1341252, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.154186] env[62066]: DEBUG oslo_concurrency.lockutils [None req-2da977d2-62cd-4e3c-b78e-c62af59abb54 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Acquiring lock "a10a6dc0-7df7-40a1-888a-d1414dbeb1f7" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 972.154455] env[62066]: DEBUG oslo_concurrency.lockutils [None req-2da977d2-62cd-4e3c-b78e-c62af59abb54 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Lock "a10a6dc0-7df7-40a1-888a-d1414dbeb1f7" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 972.225605] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-730add6c-3964-46cd-a9c3-5da2af470bd4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.232739] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44c8dac3-455b-427c-a492-bc9c934edc22 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.263623] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd1577e6-775b-4861-8bcf-026284225569 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.270542] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca3b5436-f0b0-4300-8ad5-951c5fc19176 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.274305] env[62066]: DEBUG nova.network.neutron [-] [instance: b718b09b-a7fa-445b-8be3-5b3ebca210a2] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 972.284597] env[62066]: DEBUG nova.compute.provider_tree [None req-20e007b5-b757-49e8-8380-d9dd27ea229d tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 972.297952] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3ac1ab90-f337-481c-abba-b7e0044bbea2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.307088] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eefd87bf-3bcb-41e1-b73b-ea7103b86235 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.331937] env[62066]: DEBUG nova.compute.manager [req-cd68fa36-d2b6-44dd-bc72-088cadbecacb req-afd87fba-09c1-4ea8-b57a-431198a2c0ae service nova] [instance: b718b09b-a7fa-445b-8be3-5b3ebca210a2] Detach interface failed, port_id=3d01745e-0d9e-4bae-ba4d-242380edbab6, reason: Instance b718b09b-a7fa-445b-8be3-5b3ebca210a2 could not be found. {{(pid=62066) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 972.349261] env[62066]: DEBUG oslo_vmware.api [None req-3dd8531f-78c0-4fc2-af15-28c19b0a2bf2 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': task-1341251, 'name': Rename_Task, 'duration_secs': 0.143309} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.349650] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-3dd8531f-78c0-4fc2-af15-28c19b0a2bf2 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 09e3ae18-1517-470d-8cc9-a8b50ee774c7] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 972.350045] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ccd03533-3fb2-448a-ad93-57479b911df1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.357516] env[62066]: DEBUG oslo_vmware.api [None req-3dd8531f-78c0-4fc2-af15-28c19b0a2bf2 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Waiting for the task: (returnval){ [ 972.357516] env[62066]: value = "task-1341253" [ 972.357516] env[62066]: _type = "Task" [ 972.357516] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.366073] env[62066]: DEBUG oslo_vmware.api [None req-3dd8531f-78c0-4fc2-af15-28c19b0a2bf2 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': task-1341253, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.568803] env[62066]: DEBUG oslo_vmware.api [None req-26ece0ef-0448-44a4-8d3a-1a89aaf96689 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1341252, 'name': ReconfigVM_Task, 'duration_secs': 0.158523} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.569120] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-26ece0ef-0448-44a4-8d3a-1a89aaf96689 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: ca504be5-209c-456c-af00-d403bfa2d634] Reconfigured VM instance instance-0000005f to detach disk 2000 {{(pid=62066) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 972.569904] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0df1989c-b2e6-4743-8d4a-c2b573b214fb {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.594249] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-26ece0ef-0448-44a4-8d3a-1a89aaf96689 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: ca504be5-209c-456c-af00-d403bfa2d634] Reconfiguring VM instance instance-0000005f to attach disk [datastore2] ca504be5-209c-456c-af00-d403bfa2d634/ca504be5-209c-456c-af00-d403bfa2d634.vmdk or device None with type thin {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 972.594700] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3b6167b1-e079-4b2b-b320-e6416b2af4fe {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.613271] env[62066]: DEBUG oslo_vmware.api [None req-26ece0ef-0448-44a4-8d3a-1a89aaf96689 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Waiting for the task: (returnval){ [ 972.613271] env[62066]: value = "task-1341254" [ 972.613271] env[62066]: _type = "Task" [ 972.613271] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.621745] env[62066]: DEBUG oslo_vmware.api [None req-26ece0ef-0448-44a4-8d3a-1a89aaf96689 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1341254, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.658033] env[62066]: DEBUG nova.compute.utils [None req-2da977d2-62cd-4e3c-b78e-c62af59abb54 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 972.776811] env[62066]: INFO nova.compute.manager [-] [instance: b718b09b-a7fa-445b-8be3-5b3ebca210a2] Took 1.44 seconds to deallocate network for instance. [ 972.787963] env[62066]: DEBUG nova.scheduler.client.report [None req-20e007b5-b757-49e8-8380-d9dd27ea229d tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 972.867956] env[62066]: DEBUG oslo_vmware.api [None req-3dd8531f-78c0-4fc2-af15-28c19b0a2bf2 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': task-1341253, 'name': PowerOnVM_Task, 'duration_secs': 0.456813} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.868650] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-3dd8531f-78c0-4fc2-af15-28c19b0a2bf2 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 09e3ae18-1517-470d-8cc9-a8b50ee774c7] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 972.868987] env[62066]: INFO nova.compute.manager [None req-3dd8531f-78c0-4fc2-af15-28c19b0a2bf2 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 09e3ae18-1517-470d-8cc9-a8b50ee774c7] Took 6.95 seconds to spawn the instance on the hypervisor. [ 972.869333] env[62066]: DEBUG nova.compute.manager [None req-3dd8531f-78c0-4fc2-af15-28c19b0a2bf2 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 09e3ae18-1517-470d-8cc9-a8b50ee774c7] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 972.870272] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b7bfe67-9f56-4938-8708-b683c95b0089 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.123551] env[62066]: DEBUG oslo_vmware.api [None req-26ece0ef-0448-44a4-8d3a-1a89aaf96689 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1341254, 'name': ReconfigVM_Task, 'duration_secs': 0.261175} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.127418] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-26ece0ef-0448-44a4-8d3a-1a89aaf96689 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: ca504be5-209c-456c-af00-d403bfa2d634] Reconfigured VM instance instance-0000005f to attach disk [datastore2] ca504be5-209c-456c-af00-d403bfa2d634/ca504be5-209c-456c-af00-d403bfa2d634.vmdk or device None with type thin {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 973.127709] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-26ece0ef-0448-44a4-8d3a-1a89aaf96689 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: ca504be5-209c-456c-af00-d403bfa2d634] Updating instance 'ca504be5-209c-456c-af00-d403bfa2d634' progress to 50 {{(pid=62066) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 973.162283] env[62066]: DEBUG oslo_concurrency.lockutils [None req-2da977d2-62cd-4e3c-b78e-c62af59abb54 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Lock "a10a6dc0-7df7-40a1-888a-d1414dbeb1f7" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 973.285034] env[62066]: DEBUG oslo_concurrency.lockutils [None req-2552ba89-1024-40da-89b2-2a360324a01f tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 973.294736] env[62066]: DEBUG oslo_concurrency.lockutils [None req-20e007b5-b757-49e8-8380-d9dd27ea229d tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.188s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 973.294736] env[62066]: DEBUG nova.compute.manager [None req-20e007b5-b757-49e8-8380-d9dd27ea229d tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: a2c5277e-af49-4b31-8480-a74d354c7383] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 973.297556] env[62066]: DEBUG oslo_concurrency.lockutils [None req-2552ba89-1024-40da-89b2-2a360324a01f tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.013s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 973.297855] env[62066]: DEBUG nova.objects.instance [None req-2552ba89-1024-40da-89b2-2a360324a01f tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Lazy-loading 'resources' on Instance uuid b718b09b-a7fa-445b-8be3-5b3ebca210a2 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 973.389793] env[62066]: INFO nova.compute.manager [None req-3dd8531f-78c0-4fc2-af15-28c19b0a2bf2 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 09e3ae18-1517-470d-8cc9-a8b50ee774c7] Took 13.00 seconds to build instance. [ 973.634923] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0b99704-4de2-4c85-a67a-cf2ee61640a0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.654350] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6047004e-93a6-47b0-a6da-3434a2a773dd {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.676164] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-26ece0ef-0448-44a4-8d3a-1a89aaf96689 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: ca504be5-209c-456c-af00-d403bfa2d634] Updating instance 'ca504be5-209c-456c-af00-d403bfa2d634' progress to 67 {{(pid=62066) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 973.801014] env[62066]: DEBUG nova.compute.utils [None req-20e007b5-b757-49e8-8380-d9dd27ea229d tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 973.807073] env[62066]: DEBUG nova.compute.manager [None req-20e007b5-b757-49e8-8380-d9dd27ea229d tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: a2c5277e-af49-4b31-8480-a74d354c7383] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 973.807258] env[62066]: DEBUG nova.network.neutron [None req-20e007b5-b757-49e8-8380-d9dd27ea229d tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: a2c5277e-af49-4b31-8480-a74d354c7383] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 973.866857] env[62066]: DEBUG nova.policy [None req-20e007b5-b757-49e8-8380-d9dd27ea229d tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '588e53cee85f4ab484b76e7a59fcbe78', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8475ad5a900548cba568360999c846ea', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 973.893521] env[62066]: DEBUG oslo_concurrency.lockutils [None req-3dd8531f-78c0-4fc2-af15-28c19b0a2bf2 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Lock "09e3ae18-1517-470d-8cc9-a8b50ee774c7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.508s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 973.939225] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce0d14a4-7bf7-42e8-a3ae-78bedb19a526 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.949926] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76fbd95c-77fb-4073-8420-d2fe91a2c4f9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.981835] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b99bb348-c1ca-4f31-b0cc-303613a1ccbb {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.989882] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8ee0f5b-16fb-4b6e-8125-a4e247f392bf {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.003039] env[62066]: DEBUG nova.compute.provider_tree [None req-2552ba89-1024-40da-89b2-2a360324a01f tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 974.160338] env[62066]: DEBUG nova.network.neutron [None req-20e007b5-b757-49e8-8380-d9dd27ea229d tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: a2c5277e-af49-4b31-8480-a74d354c7383] Successfully created port: 370027d1-b4b7-4b8c-8258-8a5350637a2c {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 974.224766] env[62066]: DEBUG nova.network.neutron [None req-26ece0ef-0448-44a4-8d3a-1a89aaf96689 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: ca504be5-209c-456c-af00-d403bfa2d634] Port 11deacd3-d3a5-49a4-98f3-3226eb532e76 binding to destination host cpu-1 is already ACTIVE {{(pid=62066) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 974.239117] env[62066]: DEBUG oslo_concurrency.lockutils [None req-471055b2-0bc6-4354-a54a-bec6a9fb9819 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Acquiring lock "09e3ae18-1517-470d-8cc9-a8b50ee774c7" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 974.239546] env[62066]: DEBUG oslo_concurrency.lockutils [None req-471055b2-0bc6-4354-a54a-bec6a9fb9819 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Lock "09e3ae18-1517-470d-8cc9-a8b50ee774c7" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 974.239899] env[62066]: DEBUG oslo_concurrency.lockutils [None req-471055b2-0bc6-4354-a54a-bec6a9fb9819 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Acquiring lock "09e3ae18-1517-470d-8cc9-a8b50ee774c7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 974.240209] env[62066]: DEBUG oslo_concurrency.lockutils [None req-471055b2-0bc6-4354-a54a-bec6a9fb9819 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Lock "09e3ae18-1517-470d-8cc9-a8b50ee774c7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 974.240513] env[62066]: DEBUG oslo_concurrency.lockutils [None req-471055b2-0bc6-4354-a54a-bec6a9fb9819 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Lock "09e3ae18-1517-470d-8cc9-a8b50ee774c7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 974.244325] env[62066]: DEBUG oslo_concurrency.lockutils [None req-2da977d2-62cd-4e3c-b78e-c62af59abb54 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Acquiring lock "a10a6dc0-7df7-40a1-888a-d1414dbeb1f7" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 974.244657] env[62066]: DEBUG oslo_concurrency.lockutils [None req-2da977d2-62cd-4e3c-b78e-c62af59abb54 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Lock "a10a6dc0-7df7-40a1-888a-d1414dbeb1f7" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 974.244996] env[62066]: INFO nova.compute.manager [None req-2da977d2-62cd-4e3c-b78e-c62af59abb54 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] [instance: a10a6dc0-7df7-40a1-888a-d1414dbeb1f7] Attaching volume ace70cc7-8451-40d6-9484-e28c1c57ff86 to /dev/sdb [ 974.247025] env[62066]: INFO nova.compute.manager [None req-471055b2-0bc6-4354-a54a-bec6a9fb9819 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 09e3ae18-1517-470d-8cc9-a8b50ee774c7] Terminating instance [ 974.248824] env[62066]: DEBUG nova.compute.manager [None req-471055b2-0bc6-4354-a54a-bec6a9fb9819 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 09e3ae18-1517-470d-8cc9-a8b50ee774c7] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 974.249140] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-471055b2-0bc6-4354-a54a-bec6a9fb9819 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 09e3ae18-1517-470d-8cc9-a8b50ee774c7] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 974.250481] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4d2c11c-538a-420a-8273-dde30d1dd2b7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.261494] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-471055b2-0bc6-4354-a54a-bec6a9fb9819 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 09e3ae18-1517-470d-8cc9-a8b50ee774c7] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 974.262408] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-858baef3-fdc3-470b-afdd-c752b4f62349 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.267968] env[62066]: DEBUG oslo_vmware.api [None req-471055b2-0bc6-4354-a54a-bec6a9fb9819 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Waiting for the task: (returnval){ [ 974.267968] env[62066]: value = "task-1341255" [ 974.267968] env[62066]: _type = "Task" [ 974.267968] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.275623] env[62066]: DEBUG oslo_vmware.api [None req-471055b2-0bc6-4354-a54a-bec6a9fb9819 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': task-1341255, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.290096] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02bc5135-8dab-45bb-8dc4-0215dfe403e8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.296591] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3a3dec6-15ae-4bfd-9ad4-bb13b0b8df11 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.309358] env[62066]: DEBUG nova.compute.manager [None req-20e007b5-b757-49e8-8380-d9dd27ea229d tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: a2c5277e-af49-4b31-8480-a74d354c7383] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 974.313206] env[62066]: DEBUG nova.virt.block_device [None req-2da977d2-62cd-4e3c-b78e-c62af59abb54 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] [instance: a10a6dc0-7df7-40a1-888a-d1414dbeb1f7] Updating existing volume attachment record: 266e8cd4-a2cc-410a-89d3-058ce9916b00 {{(pid=62066) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 974.506352] env[62066]: DEBUG nova.scheduler.client.report [None req-2552ba89-1024-40da-89b2-2a360324a01f tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 974.778689] env[62066]: DEBUG oslo_vmware.api [None req-471055b2-0bc6-4354-a54a-bec6a9fb9819 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': task-1341255, 'name': PowerOffVM_Task, 'duration_secs': 0.208571} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.779036] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-471055b2-0bc6-4354-a54a-bec6a9fb9819 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 09e3ae18-1517-470d-8cc9-a8b50ee774c7] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 974.779235] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-471055b2-0bc6-4354-a54a-bec6a9fb9819 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 09e3ae18-1517-470d-8cc9-a8b50ee774c7] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 974.779536] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4bb164a4-f74f-4a6a-a6c8-7eae6f349b5a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.846065] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-471055b2-0bc6-4354-a54a-bec6a9fb9819 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 09e3ae18-1517-470d-8cc9-a8b50ee774c7] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 974.846065] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-471055b2-0bc6-4354-a54a-bec6a9fb9819 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 09e3ae18-1517-470d-8cc9-a8b50ee774c7] Deleting contents of the VM from datastore datastore1 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 974.846065] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-471055b2-0bc6-4354-a54a-bec6a9fb9819 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Deleting the datastore file [datastore1] 09e3ae18-1517-470d-8cc9-a8b50ee774c7 {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 974.846065] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8ad164fe-b1ff-46b0-9874-c7d8f896207f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.853428] env[62066]: DEBUG oslo_vmware.api [None req-471055b2-0bc6-4354-a54a-bec6a9fb9819 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Waiting for the task: (returnval){ [ 974.853428] env[62066]: value = "task-1341258" [ 974.853428] env[62066]: _type = "Task" [ 974.853428] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.861034] env[62066]: DEBUG oslo_vmware.api [None req-471055b2-0bc6-4354-a54a-bec6a9fb9819 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': task-1341258, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.011912] env[62066]: DEBUG oslo_concurrency.lockutils [None req-2552ba89-1024-40da-89b2-2a360324a01f tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.714s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 975.035896] env[62066]: INFO nova.scheduler.client.report [None req-2552ba89-1024-40da-89b2-2a360324a01f tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Deleted allocations for instance b718b09b-a7fa-445b-8be3-5b3ebca210a2 [ 975.243896] env[62066]: DEBUG oslo_concurrency.lockutils [None req-26ece0ef-0448-44a4-8d3a-1a89aaf96689 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Acquiring lock "ca504be5-209c-456c-af00-d403bfa2d634-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 975.244165] env[62066]: DEBUG oslo_concurrency.lockutils [None req-26ece0ef-0448-44a4-8d3a-1a89aaf96689 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Lock "ca504be5-209c-456c-af00-d403bfa2d634-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 975.244349] env[62066]: DEBUG oslo_concurrency.lockutils [None req-26ece0ef-0448-44a4-8d3a-1a89aaf96689 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Lock "ca504be5-209c-456c-af00-d403bfa2d634-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 975.319982] env[62066]: DEBUG nova.compute.manager [None req-20e007b5-b757-49e8-8380-d9dd27ea229d tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: a2c5277e-af49-4b31-8480-a74d354c7383] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 975.343813] env[62066]: DEBUG nova.virt.hardware [None req-20e007b5-b757-49e8-8380-d9dd27ea229d tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-23T13:40:41Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-23T13:40:23Z,direct_url=,disk_format='vmdk',id=50ff584c-3b50-4395-af07-3e66769bc9f7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='eb52f7069a374c61ae946f052007c6d9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-23T13:40:24Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 975.344077] env[62066]: DEBUG nova.virt.hardware [None req-20e007b5-b757-49e8-8380-d9dd27ea229d tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 975.344242] env[62066]: DEBUG nova.virt.hardware [None req-20e007b5-b757-49e8-8380-d9dd27ea229d tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 975.344429] env[62066]: DEBUG nova.virt.hardware [None req-20e007b5-b757-49e8-8380-d9dd27ea229d tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 975.344579] env[62066]: DEBUG nova.virt.hardware [None req-20e007b5-b757-49e8-8380-d9dd27ea229d tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 975.344732] env[62066]: DEBUG nova.virt.hardware [None req-20e007b5-b757-49e8-8380-d9dd27ea229d tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 975.344945] env[62066]: DEBUG nova.virt.hardware [None req-20e007b5-b757-49e8-8380-d9dd27ea229d tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 975.345143] env[62066]: DEBUG nova.virt.hardware [None req-20e007b5-b757-49e8-8380-d9dd27ea229d tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 975.345325] env[62066]: DEBUG nova.virt.hardware [None req-20e007b5-b757-49e8-8380-d9dd27ea229d tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 975.345496] env[62066]: DEBUG nova.virt.hardware [None req-20e007b5-b757-49e8-8380-d9dd27ea229d tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 975.345675] env[62066]: DEBUG nova.virt.hardware [None req-20e007b5-b757-49e8-8380-d9dd27ea229d tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 975.346608] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-366227e2-2161-43d3-b2ea-319a1745798e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.359533] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc954739-6da3-433d-9289-6bbf8e9860d0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.067593] env[62066]: DEBUG nova.network.neutron [None req-20e007b5-b757-49e8-8380-d9dd27ea229d tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: a2c5277e-af49-4b31-8480-a74d354c7383] Successfully updated port: 370027d1-b4b7-4b8c-8258-8a5350637a2c {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 976.074722] env[62066]: DEBUG nova.compute.manager [req-296dccde-f2fa-406e-9e09-9f3f37911a4a req-4de1a3d2-367f-4768-87c1-e793fed61680 service nova] [instance: a2c5277e-af49-4b31-8480-a74d354c7383] Received event network-vif-plugged-370027d1-b4b7-4b8c-8258-8a5350637a2c {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 976.074722] env[62066]: DEBUG oslo_concurrency.lockutils [req-296dccde-f2fa-406e-9e09-9f3f37911a4a req-4de1a3d2-367f-4768-87c1-e793fed61680 service nova] Acquiring lock "a2c5277e-af49-4b31-8480-a74d354c7383-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 976.074722] env[62066]: DEBUG oslo_concurrency.lockutils [req-296dccde-f2fa-406e-9e09-9f3f37911a4a req-4de1a3d2-367f-4768-87c1-e793fed61680 service nova] Lock "a2c5277e-af49-4b31-8480-a74d354c7383-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 976.074722] env[62066]: DEBUG oslo_concurrency.lockutils [req-296dccde-f2fa-406e-9e09-9f3f37911a4a req-4de1a3d2-367f-4768-87c1-e793fed61680 service nova] Lock "a2c5277e-af49-4b31-8480-a74d354c7383-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 976.074722] env[62066]: DEBUG nova.compute.manager [req-296dccde-f2fa-406e-9e09-9f3f37911a4a req-4de1a3d2-367f-4768-87c1-e793fed61680 service nova] [instance: a2c5277e-af49-4b31-8480-a74d354c7383] No waiting events found dispatching network-vif-plugged-370027d1-b4b7-4b8c-8258-8a5350637a2c {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 976.074722] env[62066]: WARNING nova.compute.manager [req-296dccde-f2fa-406e-9e09-9f3f37911a4a req-4de1a3d2-367f-4768-87c1-e793fed61680 service nova] [instance: a2c5277e-af49-4b31-8480-a74d354c7383] Received unexpected event network-vif-plugged-370027d1-b4b7-4b8c-8258-8a5350637a2c for instance with vm_state building and task_state spawning. [ 976.074722] env[62066]: DEBUG oslo_concurrency.lockutils [None req-2552ba89-1024-40da-89b2-2a360324a01f tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Lock "b718b09b-a7fa-445b-8be3-5b3ebca210a2" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.875s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 976.076509] env[62066]: DEBUG oslo_concurrency.lockutils [req-93968833-5618-4e60-b150-1a661ce4b6b4 req-0dde6c1f-b6d4-4fb5-ba44-c829d1d16ec3 service nova] Acquired lock "b718b09b-a7fa-445b-8be3-5b3ebca210a2" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 976.076957] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-940b8c30-6aad-4d66-abcd-87e197bd3ae9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.090314] env[62066]: DEBUG oslo_vmware.api [None req-471055b2-0bc6-4354-a54a-bec6a9fb9819 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': task-1341258, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.152946} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.092166] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-471055b2-0bc6-4354-a54a-bec6a9fb9819 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 976.092405] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-471055b2-0bc6-4354-a54a-bec6a9fb9819 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 09e3ae18-1517-470d-8cc9-a8b50ee774c7] Deleted contents of the VM from datastore datastore1 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 976.092603] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-471055b2-0bc6-4354-a54a-bec6a9fb9819 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 09e3ae18-1517-470d-8cc9-a8b50ee774c7] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 976.092781] env[62066]: INFO nova.compute.manager [None req-471055b2-0bc6-4354-a54a-bec6a9fb9819 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 09e3ae18-1517-470d-8cc9-a8b50ee774c7] Took 1.84 seconds to destroy the instance on the hypervisor. [ 976.093014] env[62066]: DEBUG oslo.service.loopingcall [None req-471055b2-0bc6-4354-a54a-bec6a9fb9819 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 976.093424] env[62066]: DEBUG nova.compute.manager [-] [instance: 09e3ae18-1517-470d-8cc9-a8b50ee774c7] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 976.093520] env[62066]: DEBUG nova.network.neutron [-] [instance: 09e3ae18-1517-470d-8cc9-a8b50ee774c7] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 976.096818] env[62066]: WARNING suds.client [-] Web service reported a SOAP processing fault using an unexpected HTTP status code 200. Reporting as an internal server error. [ 976.096981] env[62066]: DEBUG oslo_vmware.api [-] Fault list: [ManagedObjectNotFound] {{(pid=62066) _invoke_api /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:337}} [ 976.097534] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-56e3fa68-7bb1-4686-b529-8ea6b26d846a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.106076] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ab929c9-e69f-4cec-9e0a-e3cbe574f5d6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.131973] env[62066]: ERROR root [req-93968833-5618-4e60-b150-1a661ce4b6b4 req-0dde6c1f-b6d4-4fb5-ba44-c829d1d16ec3 service nova] Original exception being dropped: ['Traceback (most recent call last):\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py", line 377, in request_handler\n response = request(managed_object, **kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 586, in __call__\n return client.invoke(args, kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 728, in invoke\n result = self.send(soapenv, timeout=timeout)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 777, in send\n return self.process_reply(reply.message, None, None)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 840, in process_reply\n raise WebFault(fault, replyroot)\n', "suds.WebFault: Server raised fault: 'The object 'vim.VirtualMachine:vm-286090' has already been deleted or has not been completely created'\n", '\nDuring handling of the above exception, another exception occurred:\n\n', 'Traceback (most recent call last):\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 301, in _invoke_api\n return api_method(*args, **kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/vim_util.py", line 480, in get_object_property\n props = get_object_properties(vim, moref, [property_name],\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/vim_util.py", line 360, in get_object_properties\n retrieve_result = vim.RetrievePropertiesEx(\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py", line 413, in request_handler\n raise exceptions.VimFaultException(fault_list, fault_string,\n', "oslo_vmware.exceptions.VimFaultException: The object 'vim.VirtualMachine:vm-286090' has already been deleted or has not been completely created\nCause: Server raised fault: 'The object 'vim.VirtualMachine:vm-286090' has already been deleted or has not been completely created'\nFaults: [ManagedObjectNotFound]\nDetails: {'obj': 'vm-286090'}\n", '\nDuring handling of the above exception, another exception occurred:\n\n', 'Traceback (most recent call last):\n', ' File "/opt/stack/nova/nova/virt/vmwareapi/session.py", line 123, in _call_method\n return self.invoke_api(module, method, self.vim, *args,\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 358, in invoke_api\n return _invoke_api(module, method, *args, **kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 122, in func\n return evt.wait()\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait\n result = hub.switch()\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch\n return self.greenlet.switch()\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/common/loopingcall.py", line 122, in _inner\n idle = self.f(*self.args, **self.kw)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 96, in _func\n result = f(*args, **kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 341, in _invoke_api\n raise clazz(str(excep),\n', "oslo_vmware.exceptions.ManagedObjectNotFoundException: The object 'vim.VirtualMachine:vm-286090' has already been deleted or has not been completely created\nCause: Server raised fault: 'The object 'vim.VirtualMachine:vm-286090' has already been deleted or has not been completely created'\nFaults: [ManagedObjectNotFound]\nDetails: {'obj': 'vm-286090'}\n"]: nova.exception.InstanceNotFound: Instance b718b09b-a7fa-445b-8be3-5b3ebca210a2 could not be found. [ 976.132326] env[62066]: DEBUG oslo_concurrency.lockutils [req-93968833-5618-4e60-b150-1a661ce4b6b4 req-0dde6c1f-b6d4-4fb5-ba44-c829d1d16ec3 service nova] Releasing lock "b718b09b-a7fa-445b-8be3-5b3ebca210a2" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 976.132418] env[62066]: DEBUG nova.compute.manager [req-93968833-5618-4e60-b150-1a661ce4b6b4 req-0dde6c1f-b6d4-4fb5-ba44-c829d1d16ec3 service nova] [instance: b718b09b-a7fa-445b-8be3-5b3ebca210a2] Detach interface failed, port_id=26c5cd4b-0e0d-4e99-a7fe-7a46764dc642, reason: Instance b718b09b-a7fa-445b-8be3-5b3ebca210a2 could not be found. {{(pid=62066) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 976.582030] env[62066]: DEBUG oslo_concurrency.lockutils [None req-20e007b5-b757-49e8-8380-d9dd27ea229d tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Acquiring lock "refresh_cache-a2c5277e-af49-4b31-8480-a74d354c7383" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 976.582030] env[62066]: DEBUG oslo_concurrency.lockutils [None req-20e007b5-b757-49e8-8380-d9dd27ea229d tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Acquired lock "refresh_cache-a2c5277e-af49-4b31-8480-a74d354c7383" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 976.582218] env[62066]: DEBUG nova.network.neutron [None req-20e007b5-b757-49e8-8380-d9dd27ea229d tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: a2c5277e-af49-4b31-8480-a74d354c7383] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 976.632443] env[62066]: DEBUG oslo_concurrency.lockutils [None req-26ece0ef-0448-44a4-8d3a-1a89aaf96689 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Acquiring lock "refresh_cache-ca504be5-209c-456c-af00-d403bfa2d634" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 976.632680] env[62066]: DEBUG oslo_concurrency.lockutils [None req-26ece0ef-0448-44a4-8d3a-1a89aaf96689 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Acquired lock "refresh_cache-ca504be5-209c-456c-af00-d403bfa2d634" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 976.632809] env[62066]: DEBUG nova.network.neutron [None req-26ece0ef-0448-44a4-8d3a-1a89aaf96689 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: ca504be5-209c-456c-af00-d403bfa2d634] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 976.833188] env[62066]: DEBUG nova.network.neutron [-] [instance: 09e3ae18-1517-470d-8cc9-a8b50ee774c7] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 977.116292] env[62066]: DEBUG nova.network.neutron [None req-20e007b5-b757-49e8-8380-d9dd27ea229d tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: a2c5277e-af49-4b31-8480-a74d354c7383] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 977.179384] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c39c0227-c481-4e93-a315-7377013f468c tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Acquiring lock "c8482d92-0180-44ee-a4f1-bf84786dad43" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 977.179660] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c39c0227-c481-4e93-a315-7377013f468c tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Lock "c8482d92-0180-44ee-a4f1-bf84786dad43" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 977.287875] env[62066]: DEBUG nova.network.neutron [None req-20e007b5-b757-49e8-8380-d9dd27ea229d tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: a2c5277e-af49-4b31-8480-a74d354c7383] Updating instance_info_cache with network_info: [{"id": "370027d1-b4b7-4b8c-8258-8a5350637a2c", "address": "fa:16:3e:42:e1:91", "network": {"id": "7617f887-bdd0-400a-8b5f-606d7785df78", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-558766873-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8475ad5a900548cba568360999c846ea", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2b6a4065-12af-4fb9-ac47-ec9143f7297e", "external-id": "nsx-vlan-transportzone-95", "segmentation_id": 95, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap370027d1-b4", "ovs_interfaceid": "370027d1-b4b7-4b8c-8258-8a5350637a2c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 977.336535] env[62066]: INFO nova.compute.manager [-] [instance: 09e3ae18-1517-470d-8cc9-a8b50ee774c7] Took 1.24 seconds to deallocate network for instance. [ 977.436269] env[62066]: DEBUG nova.network.neutron [None req-26ece0ef-0448-44a4-8d3a-1a89aaf96689 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: ca504be5-209c-456c-af00-d403bfa2d634] Updating instance_info_cache with network_info: [{"id": "11deacd3-d3a5-49a4-98f3-3226eb532e76", "address": "fa:16:3e:3a:b5:74", "network": {"id": "1c6c51e4-96f8-4175-b7f5-96b8a42fcb13", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-2006089809-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d609babc987a42e2a8ddb4bfb9c3b103", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a965790c-2d2f-4c2a-9ee7-745f4d53039b", "external-id": "nsx-vlan-transportzone-708", "segmentation_id": 708, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap11deacd3-d3", "ovs_interfaceid": "11deacd3-d3a5-49a4-98f3-3226eb532e76", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 977.610097] env[62066]: DEBUG nova.compute.manager [req-c7c4fe93-3c7f-4ff2-a027-a406553837cd req-62198f4d-b5e7-4b78-8645-e511885dcb52 service nova] [instance: a2c5277e-af49-4b31-8480-a74d354c7383] Received event network-changed-370027d1-b4b7-4b8c-8258-8a5350637a2c {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 977.610318] env[62066]: DEBUG nova.compute.manager [req-c7c4fe93-3c7f-4ff2-a027-a406553837cd req-62198f4d-b5e7-4b78-8645-e511885dcb52 service nova] [instance: a2c5277e-af49-4b31-8480-a74d354c7383] Refreshing instance network info cache due to event network-changed-370027d1-b4b7-4b8c-8258-8a5350637a2c. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 977.610566] env[62066]: DEBUG oslo_concurrency.lockutils [req-c7c4fe93-3c7f-4ff2-a027-a406553837cd req-62198f4d-b5e7-4b78-8645-e511885dcb52 service nova] Acquiring lock "refresh_cache-a2c5277e-af49-4b31-8480-a74d354c7383" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 977.682203] env[62066]: DEBUG nova.compute.manager [None req-c39c0227-c481-4e93-a315-7377013f468c tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: c8482d92-0180-44ee-a4f1-bf84786dad43] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 977.790886] env[62066]: DEBUG oslo_concurrency.lockutils [None req-20e007b5-b757-49e8-8380-d9dd27ea229d tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Releasing lock "refresh_cache-a2c5277e-af49-4b31-8480-a74d354c7383" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 977.791226] env[62066]: DEBUG nova.compute.manager [None req-20e007b5-b757-49e8-8380-d9dd27ea229d tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: a2c5277e-af49-4b31-8480-a74d354c7383] Instance network_info: |[{"id": "370027d1-b4b7-4b8c-8258-8a5350637a2c", "address": "fa:16:3e:42:e1:91", "network": {"id": "7617f887-bdd0-400a-8b5f-606d7785df78", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-558766873-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8475ad5a900548cba568360999c846ea", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2b6a4065-12af-4fb9-ac47-ec9143f7297e", "external-id": "nsx-vlan-transportzone-95", "segmentation_id": 95, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap370027d1-b4", "ovs_interfaceid": "370027d1-b4b7-4b8c-8258-8a5350637a2c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 977.791552] env[62066]: DEBUG oslo_concurrency.lockutils [req-c7c4fe93-3c7f-4ff2-a027-a406553837cd req-62198f4d-b5e7-4b78-8645-e511885dcb52 service nova] Acquired lock "refresh_cache-a2c5277e-af49-4b31-8480-a74d354c7383" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 977.791738] env[62066]: DEBUG nova.network.neutron [req-c7c4fe93-3c7f-4ff2-a027-a406553837cd req-62198f4d-b5e7-4b78-8645-e511885dcb52 service nova] [instance: a2c5277e-af49-4b31-8480-a74d354c7383] Refreshing network info cache for port 370027d1-b4b7-4b8c-8258-8a5350637a2c {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 977.792981] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-20e007b5-b757-49e8-8380-d9dd27ea229d tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: a2c5277e-af49-4b31-8480-a74d354c7383] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:42:e1:91', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '2b6a4065-12af-4fb9-ac47-ec9143f7297e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '370027d1-b4b7-4b8c-8258-8a5350637a2c', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 977.800431] env[62066]: DEBUG oslo.service.loopingcall [None req-20e007b5-b757-49e8-8380-d9dd27ea229d tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 977.803386] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a2c5277e-af49-4b31-8480-a74d354c7383] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 977.803835] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-95eba901-323e-4c87-ba62-6f3862f9520f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.823698] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 977.823698] env[62066]: value = "task-1341260" [ 977.823698] env[62066]: _type = "Task" [ 977.823698] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.831294] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1341260, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.842605] env[62066]: DEBUG oslo_concurrency.lockutils [None req-471055b2-0bc6-4354-a54a-bec6a9fb9819 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 977.842820] env[62066]: DEBUG oslo_concurrency.lockutils [None req-471055b2-0bc6-4354-a54a-bec6a9fb9819 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 977.843054] env[62066]: DEBUG nova.objects.instance [None req-471055b2-0bc6-4354-a54a-bec6a9fb9819 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Lazy-loading 'resources' on Instance uuid 09e3ae18-1517-470d-8cc9-a8b50ee774c7 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 977.938850] env[62066]: DEBUG oslo_concurrency.lockutils [None req-26ece0ef-0448-44a4-8d3a-1a89aaf96689 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Releasing lock "refresh_cache-ca504be5-209c-456c-af00-d403bfa2d634" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 978.013685] env[62066]: DEBUG nova.network.neutron [req-c7c4fe93-3c7f-4ff2-a027-a406553837cd req-62198f4d-b5e7-4b78-8645-e511885dcb52 service nova] [instance: a2c5277e-af49-4b31-8480-a74d354c7383] Updated VIF entry in instance network info cache for port 370027d1-b4b7-4b8c-8258-8a5350637a2c. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 978.014097] env[62066]: DEBUG nova.network.neutron [req-c7c4fe93-3c7f-4ff2-a027-a406553837cd req-62198f4d-b5e7-4b78-8645-e511885dcb52 service nova] [instance: a2c5277e-af49-4b31-8480-a74d354c7383] Updating instance_info_cache with network_info: [{"id": "370027d1-b4b7-4b8c-8258-8a5350637a2c", "address": "fa:16:3e:42:e1:91", "network": {"id": "7617f887-bdd0-400a-8b5f-606d7785df78", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-558766873-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8475ad5a900548cba568360999c846ea", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2b6a4065-12af-4fb9-ac47-ec9143f7297e", "external-id": "nsx-vlan-transportzone-95", "segmentation_id": 95, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap370027d1-b4", "ovs_interfaceid": "370027d1-b4b7-4b8c-8258-8a5350637a2c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 978.205622] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c39c0227-c481-4e93-a315-7377013f468c tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 978.334749] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1341260, 'name': CreateVM_Task, 'duration_secs': 0.287279} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.334940] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a2c5277e-af49-4b31-8480-a74d354c7383] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 978.335640] env[62066]: DEBUG oslo_concurrency.lockutils [None req-20e007b5-b757-49e8-8380-d9dd27ea229d tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 978.335894] env[62066]: DEBUG oslo_concurrency.lockutils [None req-20e007b5-b757-49e8-8380-d9dd27ea229d tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Acquired lock "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 978.336244] env[62066]: DEBUG oslo_concurrency.lockutils [None req-20e007b5-b757-49e8-8380-d9dd27ea229d tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 978.336506] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-87c0f85f-5b13-479d-b00c-d261e6e444b6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.341045] env[62066]: DEBUG oslo_vmware.api [None req-20e007b5-b757-49e8-8380-d9dd27ea229d tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Waiting for the task: (returnval){ [ 978.341045] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]52bdb2ab-87fb-73b0-e7ea-b2d37e0bf056" [ 978.341045] env[62066]: _type = "Task" [ 978.341045] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.350662] env[62066]: DEBUG oslo_vmware.api [None req-20e007b5-b757-49e8-8380-d9dd27ea229d tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52bdb2ab-87fb-73b0-e7ea-b2d37e0bf056, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.450997] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d937d175-1a4d-41fe-b4f1-516e13c637fe {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.459513] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b82cd16-31dd-4802-8638-277ceaa83353 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.462612] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac2faeb2-8c10-4faf-aabe-d7ac390be972 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.502576] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22df2c2d-ca44-4d2d-b80e-6832bfd220dc {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.505449] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edd0eb72-93e4-4daf-ade6-e652a304593a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.511499] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-26ece0ef-0448-44a4-8d3a-1a89aaf96689 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: ca504be5-209c-456c-af00-d403bfa2d634] Updating instance 'ca504be5-209c-456c-af00-d403bfa2d634' progress to 83 {{(pid=62066) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 978.516686] env[62066]: DEBUG oslo_concurrency.lockutils [req-c7c4fe93-3c7f-4ff2-a027-a406553837cd req-62198f4d-b5e7-4b78-8645-e511885dcb52 service nova] Releasing lock "refresh_cache-a2c5277e-af49-4b31-8480-a74d354c7383" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 978.516920] env[62066]: DEBUG nova.compute.manager [req-c7c4fe93-3c7f-4ff2-a027-a406553837cd req-62198f4d-b5e7-4b78-8645-e511885dcb52 service nova] [instance: 09e3ae18-1517-470d-8cc9-a8b50ee774c7] Received event network-vif-deleted-f62ec0e9-6137-4099-8d1b-62a38a3a9938 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 978.518406] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca31b7e1-bde3-4a39-9f53-f9ddbf40448d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.530615] env[62066]: DEBUG nova.compute.provider_tree [None req-471055b2-0bc6-4354-a54a-bec6a9fb9819 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 978.851779] env[62066]: DEBUG oslo_vmware.api [None req-20e007b5-b757-49e8-8380-d9dd27ea229d tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52bdb2ab-87fb-73b0-e7ea-b2d37e0bf056, 'name': SearchDatastore_Task, 'duration_secs': 0.011289} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.852274] env[62066]: DEBUG oslo_concurrency.lockutils [None req-20e007b5-b757-49e8-8380-d9dd27ea229d tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Releasing lock "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 978.852543] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-20e007b5-b757-49e8-8380-d9dd27ea229d tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: a2c5277e-af49-4b31-8480-a74d354c7383] Processing image 50ff584c-3b50-4395-af07-3e66769bc9f7 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 978.852872] env[62066]: DEBUG oslo_concurrency.lockutils [None req-20e007b5-b757-49e8-8380-d9dd27ea229d tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 978.853152] env[62066]: DEBUG oslo_concurrency.lockutils [None req-20e007b5-b757-49e8-8380-d9dd27ea229d tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Acquired lock "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 978.853472] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-20e007b5-b757-49e8-8380-d9dd27ea229d tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 978.853787] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9b000645-125e-4c50-b078-94fa4c4ec173 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.859089] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-2da977d2-62cd-4e3c-b78e-c62af59abb54 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] [instance: a10a6dc0-7df7-40a1-888a-d1414dbeb1f7] Volume attach. Driver type: vmdk {{(pid=62066) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 978.859317] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-2da977d2-62cd-4e3c-b78e-c62af59abb54 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] [instance: a10a6dc0-7df7-40a1-888a-d1414dbeb1f7] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-286117', 'volume_id': 'ace70cc7-8451-40d6-9484-e28c1c57ff86', 'name': 'volume-ace70cc7-8451-40d6-9484-e28c1c57ff86', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'a10a6dc0-7df7-40a1-888a-d1414dbeb1f7', 'attached_at': '', 'detached_at': '', 'volume_id': 'ace70cc7-8451-40d6-9484-e28c1c57ff86', 'serial': 'ace70cc7-8451-40d6-9484-e28c1c57ff86'} {{(pid=62066) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 978.860099] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f900d4da-2093-4b48-a337-5b943a71bc25 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.863294] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-20e007b5-b757-49e8-8380-d9dd27ea229d tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 978.863468] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-20e007b5-b757-49e8-8380-d9dd27ea229d tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 978.864455] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-12a28c2b-e82c-475e-a71b-691a2eca86e7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.878546] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d50c5bba-eaef-475d-b314-5aa30ea2fc69 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.881868] env[62066]: DEBUG oslo_vmware.api [None req-20e007b5-b757-49e8-8380-d9dd27ea229d tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Waiting for the task: (returnval){ [ 978.881868] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]52b33592-4575-7a57-31b5-c4b5b5c70064" [ 978.881868] env[62066]: _type = "Task" [ 978.881868] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.904203] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-2da977d2-62cd-4e3c-b78e-c62af59abb54 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] [instance: a10a6dc0-7df7-40a1-888a-d1414dbeb1f7] Reconfiguring VM instance instance-0000005b to attach disk [datastore2] volume-ace70cc7-8451-40d6-9484-e28c1c57ff86/volume-ace70cc7-8451-40d6-9484-e28c1c57ff86.vmdk or device None with type thin {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 978.904844] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ecbe028a-fe77-4175-879d-450993f0f83d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.920524] env[62066]: DEBUG oslo_vmware.api [None req-20e007b5-b757-49e8-8380-d9dd27ea229d tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52b33592-4575-7a57-31b5-c4b5b5c70064, 'name': SearchDatastore_Task, 'duration_secs': 0.008566} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.921611] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-05be3392-d0ea-47b4-b892-4669dad84ade {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.925383] env[62066]: DEBUG oslo_vmware.api [None req-2da977d2-62cd-4e3c-b78e-c62af59abb54 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Waiting for the task: (returnval){ [ 978.925383] env[62066]: value = "task-1341261" [ 978.925383] env[62066]: _type = "Task" [ 978.925383] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.926520] env[62066]: DEBUG oslo_vmware.api [None req-20e007b5-b757-49e8-8380-d9dd27ea229d tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Waiting for the task: (returnval){ [ 978.926520] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]52f17a01-f02d-42fd-f68a-c4a22614b7eb" [ 978.926520] env[62066]: _type = "Task" [ 978.926520] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.937117] env[62066]: DEBUG oslo_vmware.api [None req-2da977d2-62cd-4e3c-b78e-c62af59abb54 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Task: {'id': task-1341261, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.941175] env[62066]: DEBUG oslo_vmware.api [None req-20e007b5-b757-49e8-8380-d9dd27ea229d tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52f17a01-f02d-42fd-f68a-c4a22614b7eb, 'name': SearchDatastore_Task, 'duration_secs': 0.007938} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.941418] env[62066]: DEBUG oslo_concurrency.lockutils [None req-20e007b5-b757-49e8-8380-d9dd27ea229d tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Releasing lock "[datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 978.941675] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-20e007b5-b757-49e8-8380-d9dd27ea229d tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk to [datastore1] a2c5277e-af49-4b31-8480-a74d354c7383/a2c5277e-af49-4b31-8480-a74d354c7383.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 978.941921] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6d2804b2-5adf-4f1e-8552-981a70e37c28 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.947149] env[62066]: DEBUG oslo_vmware.api [None req-20e007b5-b757-49e8-8380-d9dd27ea229d tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Waiting for the task: (returnval){ [ 978.947149] env[62066]: value = "task-1341262" [ 978.947149] env[62066]: _type = "Task" [ 978.947149] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.954607] env[62066]: DEBUG oslo_vmware.api [None req-20e007b5-b757-49e8-8380-d9dd27ea229d tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341262, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.019748] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-26ece0ef-0448-44a4-8d3a-1a89aaf96689 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: ca504be5-209c-456c-af00-d403bfa2d634] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 979.020082] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8a3dfef7-7ace-48c7-a3cd-02970aa7544a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.025788] env[62066]: DEBUG oslo_vmware.api [None req-26ece0ef-0448-44a4-8d3a-1a89aaf96689 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Waiting for the task: (returnval){ [ 979.025788] env[62066]: value = "task-1341263" [ 979.025788] env[62066]: _type = "Task" [ 979.025788] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.033627] env[62066]: DEBUG nova.scheduler.client.report [None req-471055b2-0bc6-4354-a54a-bec6a9fb9819 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 979.036626] env[62066]: DEBUG oslo_vmware.api [None req-26ece0ef-0448-44a4-8d3a-1a89aaf96689 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1341263, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.435431] env[62066]: DEBUG oslo_vmware.api [None req-2da977d2-62cd-4e3c-b78e-c62af59abb54 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Task: {'id': task-1341261, 'name': ReconfigVM_Task, 'duration_secs': 0.472041} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.435759] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-2da977d2-62cd-4e3c-b78e-c62af59abb54 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] [instance: a10a6dc0-7df7-40a1-888a-d1414dbeb1f7] Reconfigured VM instance instance-0000005b to attach disk [datastore2] volume-ace70cc7-8451-40d6-9484-e28c1c57ff86/volume-ace70cc7-8451-40d6-9484-e28c1c57ff86.vmdk or device None with type thin {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 979.440465] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b3eddf31-5d55-4834-ace0-ffa425d70dbb {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.458745] env[62066]: DEBUG oslo_vmware.api [None req-20e007b5-b757-49e8-8380-d9dd27ea229d tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341262, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.460124} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.460027] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-20e007b5-b757-49e8-8380-d9dd27ea229d tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk to [datastore1] a2c5277e-af49-4b31-8480-a74d354c7383/a2c5277e-af49-4b31-8480-a74d354c7383.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 979.460266] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-20e007b5-b757-49e8-8380-d9dd27ea229d tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: a2c5277e-af49-4b31-8480-a74d354c7383] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 979.460616] env[62066]: DEBUG oslo_vmware.api [None req-2da977d2-62cd-4e3c-b78e-c62af59abb54 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Waiting for the task: (returnval){ [ 979.460616] env[62066]: value = "task-1341264" [ 979.460616] env[62066]: _type = "Task" [ 979.460616] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.460822] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6b70e4f2-fd3d-42c7-b823-83843fa3e0f6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.470776] env[62066]: DEBUG oslo_vmware.api [None req-2da977d2-62cd-4e3c-b78e-c62af59abb54 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Task: {'id': task-1341264, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.471961] env[62066]: DEBUG oslo_vmware.api [None req-20e007b5-b757-49e8-8380-d9dd27ea229d tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Waiting for the task: (returnval){ [ 979.471961] env[62066]: value = "task-1341265" [ 979.471961] env[62066]: _type = "Task" [ 979.471961] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.480571] env[62066]: DEBUG oslo_vmware.api [None req-20e007b5-b757-49e8-8380-d9dd27ea229d tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341265, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.536687] env[62066]: DEBUG oslo_vmware.api [None req-26ece0ef-0448-44a4-8d3a-1a89aaf96689 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1341263, 'name': PowerOnVM_Task, 'duration_secs': 0.469382} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.536981] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-26ece0ef-0448-44a4-8d3a-1a89aaf96689 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: ca504be5-209c-456c-af00-d403bfa2d634] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 979.537183] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-26ece0ef-0448-44a4-8d3a-1a89aaf96689 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: ca504be5-209c-456c-af00-d403bfa2d634] Updating instance 'ca504be5-209c-456c-af00-d403bfa2d634' progress to 100 {{(pid=62066) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 979.541080] env[62066]: DEBUG oslo_concurrency.lockutils [None req-471055b2-0bc6-4354-a54a-bec6a9fb9819 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.698s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 979.543313] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c39c0227-c481-4e93-a315-7377013f468c tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.338s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 979.544814] env[62066]: INFO nova.compute.claims [None req-c39c0227-c481-4e93-a315-7377013f468c tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: c8482d92-0180-44ee-a4f1-bf84786dad43] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 979.566351] env[62066]: INFO nova.scheduler.client.report [None req-471055b2-0bc6-4354-a54a-bec6a9fb9819 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Deleted allocations for instance 09e3ae18-1517-470d-8cc9-a8b50ee774c7 [ 979.972306] env[62066]: DEBUG oslo_vmware.api [None req-2da977d2-62cd-4e3c-b78e-c62af59abb54 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Task: {'id': task-1341264, 'name': ReconfigVM_Task, 'duration_secs': 0.13306} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.972677] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-2da977d2-62cd-4e3c-b78e-c62af59abb54 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] [instance: a10a6dc0-7df7-40a1-888a-d1414dbeb1f7] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-286117', 'volume_id': 'ace70cc7-8451-40d6-9484-e28c1c57ff86', 'name': 'volume-ace70cc7-8451-40d6-9484-e28c1c57ff86', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'a10a6dc0-7df7-40a1-888a-d1414dbeb1f7', 'attached_at': '', 'detached_at': '', 'volume_id': 'ace70cc7-8451-40d6-9484-e28c1c57ff86', 'serial': 'ace70cc7-8451-40d6-9484-e28c1c57ff86'} {{(pid=62066) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 979.981976] env[62066]: DEBUG oslo_vmware.api [None req-20e007b5-b757-49e8-8380-d9dd27ea229d tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341265, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068254} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.982270] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-20e007b5-b757-49e8-8380-d9dd27ea229d tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: a2c5277e-af49-4b31-8480-a74d354c7383] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 979.983059] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3f00c66-5fc0-4a93-bfe1-63fc8804714d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.005395] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-20e007b5-b757-49e8-8380-d9dd27ea229d tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: a2c5277e-af49-4b31-8480-a74d354c7383] Reconfiguring VM instance instance-00000061 to attach disk [datastore1] a2c5277e-af49-4b31-8480-a74d354c7383/a2c5277e-af49-4b31-8480-a74d354c7383.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 980.005916] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2212d420-8f24-464e-b9c7-35f8c1b7d5e3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.025762] env[62066]: DEBUG oslo_vmware.api [None req-20e007b5-b757-49e8-8380-d9dd27ea229d tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Waiting for the task: (returnval){ [ 980.025762] env[62066]: value = "task-1341266" [ 980.025762] env[62066]: _type = "Task" [ 980.025762] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.033987] env[62066]: DEBUG oslo_vmware.api [None req-20e007b5-b757-49e8-8380-d9dd27ea229d tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341266, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.076846] env[62066]: DEBUG oslo_concurrency.lockutils [None req-471055b2-0bc6-4354-a54a-bec6a9fb9819 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Lock "09e3ae18-1517-470d-8cc9-a8b50ee774c7" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.837s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 980.536465] env[62066]: DEBUG oslo_vmware.api [None req-20e007b5-b757-49e8-8380-d9dd27ea229d tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341266, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.651227] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad0ab184-6fd9-4cec-892e-de6761f3dc6b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.658670] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a181e06-a7f8-4ac4-a089-6d9489788ac6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.690621] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7261b2c0-ee7e-44f8-bcb4-a70613956e97 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.698393] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a543f243-d613-4bcd-9b77-621ae08c7d8f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.712090] env[62066]: DEBUG nova.compute.provider_tree [None req-c39c0227-c481-4e93-a315-7377013f468c tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 981.010670] env[62066]: DEBUG nova.objects.instance [None req-2da977d2-62cd-4e3c-b78e-c62af59abb54 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Lazy-loading 'flavor' on Instance uuid a10a6dc0-7df7-40a1-888a-d1414dbeb1f7 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 981.037720] env[62066]: DEBUG oslo_vmware.api [None req-20e007b5-b757-49e8-8380-d9dd27ea229d tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341266, 'name': ReconfigVM_Task, 'duration_secs': 0.654984} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.038754] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-20e007b5-b757-49e8-8380-d9dd27ea229d tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: a2c5277e-af49-4b31-8480-a74d354c7383] Reconfigured VM instance instance-00000061 to attach disk [datastore1] a2c5277e-af49-4b31-8480-a74d354c7383/a2c5277e-af49-4b31-8480-a74d354c7383.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 981.039393] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-46aabceb-5794-47cc-b780-b3f0516be568 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.045383] env[62066]: DEBUG oslo_vmware.api [None req-20e007b5-b757-49e8-8380-d9dd27ea229d tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Waiting for the task: (returnval){ [ 981.045383] env[62066]: value = "task-1341267" [ 981.045383] env[62066]: _type = "Task" [ 981.045383] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.065013] env[62066]: DEBUG oslo_vmware.api [None req-20e007b5-b757-49e8-8380-d9dd27ea229d tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341267, 'name': Rename_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.215524] env[62066]: DEBUG nova.scheduler.client.report [None req-c39c0227-c481-4e93-a315-7377013f468c tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 981.273497] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b907d10e-116c-4b88-b193-0e79f2464198 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Acquiring lock "ca504be5-209c-456c-af00-d403bfa2d634" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 981.273894] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b907d10e-116c-4b88-b193-0e79f2464198 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Lock "ca504be5-209c-456c-af00-d403bfa2d634" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 981.274099] env[62066]: DEBUG nova.compute.manager [None req-b907d10e-116c-4b88-b193-0e79f2464198 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: ca504be5-209c-456c-af00-d403bfa2d634] Going to confirm migration 2 {{(pid=62066) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 981.516252] env[62066]: DEBUG oslo_concurrency.lockutils [None req-2da977d2-62cd-4e3c-b78e-c62af59abb54 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Lock "a10a6dc0-7df7-40a1-888a-d1414dbeb1f7" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.271s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 981.555524] env[62066]: DEBUG oslo_vmware.api [None req-20e007b5-b757-49e8-8380-d9dd27ea229d tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341267, 'name': Rename_Task, 'duration_secs': 0.127739} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.555904] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-20e007b5-b757-49e8-8380-d9dd27ea229d tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: a2c5277e-af49-4b31-8480-a74d354c7383] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 981.556163] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f03d2f50-a8e4-4449-9c51-e7e2c3206f44 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.562600] env[62066]: DEBUG oslo_vmware.api [None req-20e007b5-b757-49e8-8380-d9dd27ea229d tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Waiting for the task: (returnval){ [ 981.562600] env[62066]: value = "task-1341268" [ 981.562600] env[62066]: _type = "Task" [ 981.562600] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.569835] env[62066]: DEBUG oslo_vmware.api [None req-20e007b5-b757-49e8-8380-d9dd27ea229d tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341268, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.720757] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c39c0227-c481-4e93-a315-7377013f468c tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.177s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 981.721375] env[62066]: DEBUG nova.compute.manager [None req-c39c0227-c481-4e93-a315-7377013f468c tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: c8482d92-0180-44ee-a4f1-bf84786dad43] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 981.737250] env[62066]: DEBUG oslo_concurrency.lockutils [None req-71dcf218-2242-47f7-84f9-80568ee1ecb7 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Acquiring lock "a10a6dc0-7df7-40a1-888a-d1414dbeb1f7" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 981.737499] env[62066]: DEBUG oslo_concurrency.lockutils [None req-71dcf218-2242-47f7-84f9-80568ee1ecb7 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Lock "a10a6dc0-7df7-40a1-888a-d1414dbeb1f7" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 981.818349] env[62066]: DEBUG oslo_concurrency.lockutils [None req-6ecf159e-1d26-4f9a-b3e2-f54f67ef8e8e tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Acquiring lock "a4ef97ad-bbb5-47fa-b5c1-2867a0d56435" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 981.818901] env[62066]: DEBUG oslo_concurrency.lockutils [None req-6ecf159e-1d26-4f9a-b3e2-f54f67ef8e8e tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Lock "a4ef97ad-bbb5-47fa-b5c1-2867a0d56435" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 981.840980] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b907d10e-116c-4b88-b193-0e79f2464198 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Acquiring lock "refresh_cache-ca504be5-209c-456c-af00-d403bfa2d634" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 981.841212] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b907d10e-116c-4b88-b193-0e79f2464198 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Acquired lock "refresh_cache-ca504be5-209c-456c-af00-d403bfa2d634" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 981.841393] env[62066]: DEBUG nova.network.neutron [None req-b907d10e-116c-4b88-b193-0e79f2464198 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: ca504be5-209c-456c-af00-d403bfa2d634] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 981.841596] env[62066]: DEBUG nova.objects.instance [None req-b907d10e-116c-4b88-b193-0e79f2464198 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Lazy-loading 'info_cache' on Instance uuid ca504be5-209c-456c-af00-d403bfa2d634 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 982.071799] env[62066]: DEBUG oslo_vmware.api [None req-20e007b5-b757-49e8-8380-d9dd27ea229d tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341268, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.226778] env[62066]: DEBUG nova.compute.utils [None req-c39c0227-c481-4e93-a315-7377013f468c tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 982.228322] env[62066]: DEBUG nova.compute.manager [None req-c39c0227-c481-4e93-a315-7377013f468c tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: c8482d92-0180-44ee-a4f1-bf84786dad43] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 982.228507] env[62066]: DEBUG nova.network.neutron [None req-c39c0227-c481-4e93-a315-7377013f468c tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: c8482d92-0180-44ee-a4f1-bf84786dad43] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 982.240221] env[62066]: INFO nova.compute.manager [None req-71dcf218-2242-47f7-84f9-80568ee1ecb7 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] [instance: a10a6dc0-7df7-40a1-888a-d1414dbeb1f7] Detaching volume ace70cc7-8451-40d6-9484-e28c1c57ff86 [ 982.277623] env[62066]: INFO nova.virt.block_device [None req-71dcf218-2242-47f7-84f9-80568ee1ecb7 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] [instance: a10a6dc0-7df7-40a1-888a-d1414dbeb1f7] Attempting to driver detach volume ace70cc7-8451-40d6-9484-e28c1c57ff86 from mountpoint /dev/sdb [ 982.277879] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-71dcf218-2242-47f7-84f9-80568ee1ecb7 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] [instance: a10a6dc0-7df7-40a1-888a-d1414dbeb1f7] Volume detach. Driver type: vmdk {{(pid=62066) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 982.278084] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-71dcf218-2242-47f7-84f9-80568ee1ecb7 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] [instance: a10a6dc0-7df7-40a1-888a-d1414dbeb1f7] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-286117', 'volume_id': 'ace70cc7-8451-40d6-9484-e28c1c57ff86', 'name': 'volume-ace70cc7-8451-40d6-9484-e28c1c57ff86', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'a10a6dc0-7df7-40a1-888a-d1414dbeb1f7', 'attached_at': '', 'detached_at': '', 'volume_id': 'ace70cc7-8451-40d6-9484-e28c1c57ff86', 'serial': 'ace70cc7-8451-40d6-9484-e28c1c57ff86'} {{(pid=62066) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 982.279169] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c233744-6e0a-4b0b-aa93-8973daaa4d3f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.283505] env[62066]: DEBUG nova.policy [None req-c39c0227-c481-4e93-a315-7377013f468c tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '747026d272d4467bbf14a0c30589c3b7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0d4e33a99c7741fb8cdd97f4ec5dbbd8', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 982.302817] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5d40e33-ec66-4f02-b26d-f6cccd52cc28 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.309619] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be5c3663-a9d6-4265-b467-871d7da9ae4d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.332191] env[62066]: DEBUG nova.compute.manager [None req-6ecf159e-1d26-4f9a-b3e2-f54f67ef8e8e tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: a4ef97ad-bbb5-47fa-b5c1-2867a0d56435] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 982.335572] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35f7806c-fa77-4675-9421-fbc866b9c266 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.352742] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-71dcf218-2242-47f7-84f9-80568ee1ecb7 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] The volume has not been displaced from its original location: [datastore2] volume-ace70cc7-8451-40d6-9484-e28c1c57ff86/volume-ace70cc7-8451-40d6-9484-e28c1c57ff86.vmdk. No consolidation needed. {{(pid=62066) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 982.357849] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-71dcf218-2242-47f7-84f9-80568ee1ecb7 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] [instance: a10a6dc0-7df7-40a1-888a-d1414dbeb1f7] Reconfiguring VM instance instance-0000005b to detach disk 2001 {{(pid=62066) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 982.358387] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-08ea23b5-e68d-44b0-be74-46d30727c715 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.376958] env[62066]: DEBUG oslo_vmware.api [None req-71dcf218-2242-47f7-84f9-80568ee1ecb7 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Waiting for the task: (returnval){ [ 982.376958] env[62066]: value = "task-1341269" [ 982.376958] env[62066]: _type = "Task" [ 982.376958] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.384690] env[62066]: DEBUG oslo_vmware.api [None req-71dcf218-2242-47f7-84f9-80568ee1ecb7 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Task: {'id': task-1341269, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.551381] env[62066]: DEBUG nova.network.neutron [None req-c39c0227-c481-4e93-a315-7377013f468c tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: c8482d92-0180-44ee-a4f1-bf84786dad43] Successfully created port: 2f2f1786-934d-478f-be39-7509cfacf86f {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 982.574483] env[62066]: DEBUG oslo_vmware.api [None req-20e007b5-b757-49e8-8380-d9dd27ea229d tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341268, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.732179] env[62066]: DEBUG nova.compute.manager [None req-c39c0227-c481-4e93-a315-7377013f468c tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: c8482d92-0180-44ee-a4f1-bf84786dad43] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 982.856121] env[62066]: DEBUG oslo_concurrency.lockutils [None req-6ecf159e-1d26-4f9a-b3e2-f54f67ef8e8e tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 982.856393] env[62066]: DEBUG oslo_concurrency.lockutils [None req-6ecf159e-1d26-4f9a-b3e2-f54f67ef8e8e tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 982.857840] env[62066]: INFO nova.compute.claims [None req-6ecf159e-1d26-4f9a-b3e2-f54f67ef8e8e tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: a4ef97ad-bbb5-47fa-b5c1-2867a0d56435] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 982.887500] env[62066]: DEBUG oslo_vmware.api [None req-71dcf218-2242-47f7-84f9-80568ee1ecb7 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Task: {'id': task-1341269, 'name': ReconfigVM_Task, 'duration_secs': 0.214828} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 982.887837] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-71dcf218-2242-47f7-84f9-80568ee1ecb7 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] [instance: a10a6dc0-7df7-40a1-888a-d1414dbeb1f7] Reconfigured VM instance instance-0000005b to detach disk 2001 {{(pid=62066) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 982.895118] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0ea76913-ec40-4e2e-8e2d-e718cf6ec7b4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.912200] env[62066]: DEBUG oslo_vmware.api [None req-71dcf218-2242-47f7-84f9-80568ee1ecb7 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Waiting for the task: (returnval){ [ 982.912200] env[62066]: value = "task-1341270" [ 982.912200] env[62066]: _type = "Task" [ 982.912200] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.921018] env[62066]: DEBUG oslo_vmware.api [None req-71dcf218-2242-47f7-84f9-80568ee1ecb7 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Task: {'id': task-1341270, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.073815] env[62066]: DEBUG oslo_vmware.api [None req-20e007b5-b757-49e8-8380-d9dd27ea229d tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341268, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.091922] env[62066]: DEBUG nova.network.neutron [None req-b907d10e-116c-4b88-b193-0e79f2464198 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: ca504be5-209c-456c-af00-d403bfa2d634] Updating instance_info_cache with network_info: [{"id": "11deacd3-d3a5-49a4-98f3-3226eb532e76", "address": "fa:16:3e:3a:b5:74", "network": {"id": "1c6c51e4-96f8-4175-b7f5-96b8a42fcb13", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-2006089809-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d609babc987a42e2a8ddb4bfb9c3b103", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a965790c-2d2f-4c2a-9ee7-745f4d53039b", "external-id": "nsx-vlan-transportzone-708", "segmentation_id": 708, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap11deacd3-d3", "ovs_interfaceid": "11deacd3-d3a5-49a4-98f3-3226eb532e76", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 983.422153] env[62066]: DEBUG oslo_vmware.api [None req-71dcf218-2242-47f7-84f9-80568ee1ecb7 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Task: {'id': task-1341270, 'name': ReconfigVM_Task, 'duration_secs': 0.13645} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.422482] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-71dcf218-2242-47f7-84f9-80568ee1ecb7 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] [instance: a10a6dc0-7df7-40a1-888a-d1414dbeb1f7] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-286117', 'volume_id': 'ace70cc7-8451-40d6-9484-e28c1c57ff86', 'name': 'volume-ace70cc7-8451-40d6-9484-e28c1c57ff86', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'a10a6dc0-7df7-40a1-888a-d1414dbeb1f7', 'attached_at': '', 'detached_at': '', 'volume_id': 'ace70cc7-8451-40d6-9484-e28c1c57ff86', 'serial': 'ace70cc7-8451-40d6-9484-e28c1c57ff86'} {{(pid=62066) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 983.578991] env[62066]: DEBUG oslo_vmware.api [None req-20e007b5-b757-49e8-8380-d9dd27ea229d tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341268, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.596289] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b907d10e-116c-4b88-b193-0e79f2464198 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Releasing lock "refresh_cache-ca504be5-209c-456c-af00-d403bfa2d634" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 983.596561] env[62066]: DEBUG nova.objects.instance [None req-b907d10e-116c-4b88-b193-0e79f2464198 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Lazy-loading 'migration_context' on Instance uuid ca504be5-209c-456c-af00-d403bfa2d634 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 983.742057] env[62066]: DEBUG nova.compute.manager [None req-c39c0227-c481-4e93-a315-7377013f468c tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: c8482d92-0180-44ee-a4f1-bf84786dad43] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 983.766007] env[62066]: DEBUG nova.virt.hardware [None req-c39c0227-c481-4e93-a315-7377013f468c tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-23T13:40:41Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-23T13:40:23Z,direct_url=,disk_format='vmdk',id=50ff584c-3b50-4395-af07-3e66769bc9f7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='eb52f7069a374c61ae946f052007c6d9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-23T13:40:24Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 983.766279] env[62066]: DEBUG nova.virt.hardware [None req-c39c0227-c481-4e93-a315-7377013f468c tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 983.766440] env[62066]: DEBUG nova.virt.hardware [None req-c39c0227-c481-4e93-a315-7377013f468c tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 983.766626] env[62066]: DEBUG nova.virt.hardware [None req-c39c0227-c481-4e93-a315-7377013f468c tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 983.766774] env[62066]: DEBUG nova.virt.hardware [None req-c39c0227-c481-4e93-a315-7377013f468c tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 983.766922] env[62066]: DEBUG nova.virt.hardware [None req-c39c0227-c481-4e93-a315-7377013f468c tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 983.767146] env[62066]: DEBUG nova.virt.hardware [None req-c39c0227-c481-4e93-a315-7377013f468c tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 983.767309] env[62066]: DEBUG nova.virt.hardware [None req-c39c0227-c481-4e93-a315-7377013f468c tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 983.767476] env[62066]: DEBUG nova.virt.hardware [None req-c39c0227-c481-4e93-a315-7377013f468c tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 983.767639] env[62066]: DEBUG nova.virt.hardware [None req-c39c0227-c481-4e93-a315-7377013f468c tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 983.767816] env[62066]: DEBUG nova.virt.hardware [None req-c39c0227-c481-4e93-a315-7377013f468c tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 983.768674] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbac815c-3e1f-4e67-9eac-971a9aad1a49 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.776217] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d47ca48c-b9c6-48fc-a6a3-b7eebf0f3f46 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.920899] env[62066]: DEBUG nova.compute.manager [req-c641b093-2f14-4189-a6a1-99049d991fb3 req-d8422c77-1719-43af-ae81-87143a1f5558 service nova] [instance: c8482d92-0180-44ee-a4f1-bf84786dad43] Received event network-vif-plugged-2f2f1786-934d-478f-be39-7509cfacf86f {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 983.921300] env[62066]: DEBUG oslo_concurrency.lockutils [req-c641b093-2f14-4189-a6a1-99049d991fb3 req-d8422c77-1719-43af-ae81-87143a1f5558 service nova] Acquiring lock "c8482d92-0180-44ee-a4f1-bf84786dad43-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 983.921559] env[62066]: DEBUG oslo_concurrency.lockutils [req-c641b093-2f14-4189-a6a1-99049d991fb3 req-d8422c77-1719-43af-ae81-87143a1f5558 service nova] Lock "c8482d92-0180-44ee-a4f1-bf84786dad43-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 983.921746] env[62066]: DEBUG oslo_concurrency.lockutils [req-c641b093-2f14-4189-a6a1-99049d991fb3 req-d8422c77-1719-43af-ae81-87143a1f5558 service nova] Lock "c8482d92-0180-44ee-a4f1-bf84786dad43-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 983.921931] env[62066]: DEBUG nova.compute.manager [req-c641b093-2f14-4189-a6a1-99049d991fb3 req-d8422c77-1719-43af-ae81-87143a1f5558 service nova] [instance: c8482d92-0180-44ee-a4f1-bf84786dad43] No waiting events found dispatching network-vif-plugged-2f2f1786-934d-478f-be39-7509cfacf86f {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 983.922109] env[62066]: WARNING nova.compute.manager [req-c641b093-2f14-4189-a6a1-99049d991fb3 req-d8422c77-1719-43af-ae81-87143a1f5558 service nova] [instance: c8482d92-0180-44ee-a4f1-bf84786dad43] Received unexpected event network-vif-plugged-2f2f1786-934d-478f-be39-7509cfacf86f for instance with vm_state building and task_state spawning. [ 983.965579] env[62066]: DEBUG nova.objects.instance [None req-71dcf218-2242-47f7-84f9-80568ee1ecb7 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Lazy-loading 'flavor' on Instance uuid a10a6dc0-7df7-40a1-888a-d1414dbeb1f7 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 983.980162] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-893a67c7-854f-42ac-9523-06ad71f208ac {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.988493] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-239044f2-c38f-413b-9ce4-47e5090164c5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.018694] env[62066]: DEBUG nova.network.neutron [None req-c39c0227-c481-4e93-a315-7377013f468c tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: c8482d92-0180-44ee-a4f1-bf84786dad43] Successfully updated port: 2f2f1786-934d-478f-be39-7509cfacf86f {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 984.020424] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7a985d6-c02b-4520-b736-5293e6de08aa {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.027517] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f028f9eb-2dfb-425d-a1be-159da62d44b1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.042775] env[62066]: DEBUG nova.compute.provider_tree [None req-6ecf159e-1d26-4f9a-b3e2-f54f67ef8e8e tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 984.075258] env[62066]: DEBUG oslo_vmware.api [None req-20e007b5-b757-49e8-8380-d9dd27ea229d tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341268, 'name': PowerOnVM_Task, 'duration_secs': 2.043866} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.075258] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-20e007b5-b757-49e8-8380-d9dd27ea229d tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: a2c5277e-af49-4b31-8480-a74d354c7383] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 984.075258] env[62066]: INFO nova.compute.manager [None req-20e007b5-b757-49e8-8380-d9dd27ea229d tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: a2c5277e-af49-4b31-8480-a74d354c7383] Took 8.75 seconds to spawn the instance on the hypervisor. [ 984.075258] env[62066]: DEBUG nova.compute.manager [None req-20e007b5-b757-49e8-8380-d9dd27ea229d tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: a2c5277e-af49-4b31-8480-a74d354c7383] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 984.075452] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c42ba506-b480-447e-ae33-032b2696d3c7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.099193] env[62066]: DEBUG nova.objects.base [None req-b907d10e-116c-4b88-b193-0e79f2464198 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Object Instance lazy-loaded attributes: info_cache,migration_context {{(pid=62066) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 984.100023] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69417620-f63a-4dab-a8f9-a0715c2e64e8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.119169] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3ddcb33a-9ab7-45c6-b553-2bbd79efc0fa {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.124519] env[62066]: DEBUG oslo_vmware.api [None req-b907d10e-116c-4b88-b193-0e79f2464198 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Waiting for the task: (returnval){ [ 984.124519] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]524ad106-4342-c50f-78d0-438300613351" [ 984.124519] env[62066]: _type = "Task" [ 984.124519] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.132639] env[62066]: DEBUG oslo_vmware.api [None req-b907d10e-116c-4b88-b193-0e79f2464198 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]524ad106-4342-c50f-78d0-438300613351, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.524027] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c39c0227-c481-4e93-a315-7377013f468c tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Acquiring lock "refresh_cache-c8482d92-0180-44ee-a4f1-bf84786dad43" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 984.524164] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c39c0227-c481-4e93-a315-7377013f468c tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Acquired lock "refresh_cache-c8482d92-0180-44ee-a4f1-bf84786dad43" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 984.524291] env[62066]: DEBUG nova.network.neutron [None req-c39c0227-c481-4e93-a315-7377013f468c tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: c8482d92-0180-44ee-a4f1-bf84786dad43] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 984.545755] env[62066]: DEBUG nova.scheduler.client.report [None req-6ecf159e-1d26-4f9a-b3e2-f54f67ef8e8e tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 984.593031] env[62066]: INFO nova.compute.manager [None req-20e007b5-b757-49e8-8380-d9dd27ea229d tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: a2c5277e-af49-4b31-8480-a74d354c7383] Took 13.51 seconds to build instance. [ 984.635135] env[62066]: DEBUG oslo_vmware.api [None req-b907d10e-116c-4b88-b193-0e79f2464198 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]524ad106-4342-c50f-78d0-438300613351, 'name': SearchDatastore_Task, 'duration_secs': 0.007164} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.635448] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b907d10e-116c-4b88-b193-0e79f2464198 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 984.972669] env[62066]: DEBUG oslo_concurrency.lockutils [None req-71dcf218-2242-47f7-84f9-80568ee1ecb7 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Lock "a10a6dc0-7df7-40a1-888a-d1414dbeb1f7" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.235s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 985.051137] env[62066]: DEBUG oslo_concurrency.lockutils [None req-6ecf159e-1d26-4f9a-b3e2-f54f67ef8e8e tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.195s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 985.051695] env[62066]: DEBUG nova.compute.manager [None req-6ecf159e-1d26-4f9a-b3e2-f54f67ef8e8e tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: a4ef97ad-bbb5-47fa-b5c1-2867a0d56435] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 985.054797] env[62066]: DEBUG nova.network.neutron [None req-c39c0227-c481-4e93-a315-7377013f468c tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: c8482d92-0180-44ee-a4f1-bf84786dad43] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 985.056641] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b907d10e-116c-4b88-b193-0e79f2464198 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.421s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 985.097195] env[62066]: DEBUG oslo_concurrency.lockutils [None req-20e007b5-b757-49e8-8380-d9dd27ea229d tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Lock "a2c5277e-af49-4b31-8480-a74d354c7383" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.017s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 985.198109] env[62066]: DEBUG nova.network.neutron [None req-c39c0227-c481-4e93-a315-7377013f468c tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: c8482d92-0180-44ee-a4f1-bf84786dad43] Updating instance_info_cache with network_info: [{"id": "2f2f1786-934d-478f-be39-7509cfacf86f", "address": "fa:16:3e:24:41:2e", "network": {"id": "449dfe4e-bb49-43b6-9ba7-b57af74ebfb3", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-282029676-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0d4e33a99c7741fb8cdd97f4ec5dbbd8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db68bd64-5b56-49af-a075-13dcf85cb2e0", "external-id": "nsx-vlan-transportzone-590", "segmentation_id": 590, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2f2f1786-93", "ovs_interfaceid": "2f2f1786-934d-478f-be39-7509cfacf86f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 985.560150] env[62066]: DEBUG nova.compute.utils [None req-6ecf159e-1d26-4f9a-b3e2-f54f67ef8e8e tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 985.565554] env[62066]: DEBUG nova.compute.manager [None req-6ecf159e-1d26-4f9a-b3e2-f54f67ef8e8e tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: a4ef97ad-bbb5-47fa-b5c1-2867a0d56435] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 985.565554] env[62066]: DEBUG nova.network.neutron [None req-6ecf159e-1d26-4f9a-b3e2-f54f67ef8e8e tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: a4ef97ad-bbb5-47fa-b5c1-2867a0d56435] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 985.636861] env[62066]: DEBUG nova.policy [None req-6ecf159e-1d26-4f9a-b3e2-f54f67ef8e8e tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8910a229218b4ec5ad72c893badfc598', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6497ab02f327476d8ff81c2ecc0371e5', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 985.687362] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e193830b-b220-44b0-a9ba-5fe9ff44c8b7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.697520] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ec72bd7-f8ce-46dd-93cb-14ba448d2f6d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.701122] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c39c0227-c481-4e93-a315-7377013f468c tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Releasing lock "refresh_cache-c8482d92-0180-44ee-a4f1-bf84786dad43" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 985.701418] env[62066]: DEBUG nova.compute.manager [None req-c39c0227-c481-4e93-a315-7377013f468c tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: c8482d92-0180-44ee-a4f1-bf84786dad43] Instance network_info: |[{"id": "2f2f1786-934d-478f-be39-7509cfacf86f", "address": "fa:16:3e:24:41:2e", "network": {"id": "449dfe4e-bb49-43b6-9ba7-b57af74ebfb3", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-282029676-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0d4e33a99c7741fb8cdd97f4ec5dbbd8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db68bd64-5b56-49af-a075-13dcf85cb2e0", "external-id": "nsx-vlan-transportzone-590", "segmentation_id": 590, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2f2f1786-93", "ovs_interfaceid": "2f2f1786-934d-478f-be39-7509cfacf86f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 985.702071] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-c39c0227-c481-4e93-a315-7377013f468c tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: c8482d92-0180-44ee-a4f1-bf84786dad43] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:24:41:2e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'db68bd64-5b56-49af-a075-13dcf85cb2e0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2f2f1786-934d-478f-be39-7509cfacf86f', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 985.709903] env[62066]: DEBUG oslo.service.loopingcall [None req-c39c0227-c481-4e93-a315-7377013f468c tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 985.711041] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c8482d92-0180-44ee-a4f1-bf84786dad43] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 985.711605] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-14191113-6643-49c2-86ff-81d76ca061aa {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.751783] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8abe888a-12ff-4162-b06f-487c49360990 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.755927] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 985.755927] env[62066]: value = "task-1341271" [ 985.755927] env[62066]: _type = "Task" [ 985.755927] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.761734] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a1246c3-0ff9-4cd3-bf4e-6f9812556105 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.769278] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1341271, 'name': CreateVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.778787] env[62066]: DEBUG nova.compute.provider_tree [None req-b907d10e-116c-4b88-b193-0e79f2464198 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 985.952850] env[62066]: DEBUG nova.compute.manager [req-994a3df0-e7e1-4806-8f79-01be87326064 req-19704908-7603-4f8a-b152-1f071414944e service nova] [instance: c8482d92-0180-44ee-a4f1-bf84786dad43] Received event network-changed-2f2f1786-934d-478f-be39-7509cfacf86f {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 985.952850] env[62066]: DEBUG nova.compute.manager [req-994a3df0-e7e1-4806-8f79-01be87326064 req-19704908-7603-4f8a-b152-1f071414944e service nova] [instance: c8482d92-0180-44ee-a4f1-bf84786dad43] Refreshing instance network info cache due to event network-changed-2f2f1786-934d-478f-be39-7509cfacf86f. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 985.953020] env[62066]: DEBUG oslo_concurrency.lockutils [req-994a3df0-e7e1-4806-8f79-01be87326064 req-19704908-7603-4f8a-b152-1f071414944e service nova] Acquiring lock "refresh_cache-c8482d92-0180-44ee-a4f1-bf84786dad43" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 985.953164] env[62066]: DEBUG oslo_concurrency.lockutils [req-994a3df0-e7e1-4806-8f79-01be87326064 req-19704908-7603-4f8a-b152-1f071414944e service nova] Acquired lock "refresh_cache-c8482d92-0180-44ee-a4f1-bf84786dad43" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 985.953380] env[62066]: DEBUG nova.network.neutron [req-994a3df0-e7e1-4806-8f79-01be87326064 req-19704908-7603-4f8a-b152-1f071414944e service nova] [instance: c8482d92-0180-44ee-a4f1-bf84786dad43] Refreshing network info cache for port 2f2f1786-934d-478f-be39-7509cfacf86f {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 985.995662] env[62066]: DEBUG nova.network.neutron [None req-6ecf159e-1d26-4f9a-b3e2-f54f67ef8e8e tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: a4ef97ad-bbb5-47fa-b5c1-2867a0d56435] Successfully created port: 4fd312f4-e41d-45eb-8505-fb30e3596b0a {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 986.034455] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a924513d-337e-489e-a67d-fcfcb3abed65 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Acquiring lock "a10a6dc0-7df7-40a1-888a-d1414dbeb1f7" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 986.034603] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a924513d-337e-489e-a67d-fcfcb3abed65 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Lock "a10a6dc0-7df7-40a1-888a-d1414dbeb1f7" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 986.034763] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a924513d-337e-489e-a67d-fcfcb3abed65 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Acquiring lock "a10a6dc0-7df7-40a1-888a-d1414dbeb1f7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 986.034954] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a924513d-337e-489e-a67d-fcfcb3abed65 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Lock "a10a6dc0-7df7-40a1-888a-d1414dbeb1f7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 986.035144] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a924513d-337e-489e-a67d-fcfcb3abed65 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Lock "a10a6dc0-7df7-40a1-888a-d1414dbeb1f7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 986.037368] env[62066]: INFO nova.compute.manager [None req-a924513d-337e-489e-a67d-fcfcb3abed65 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] [instance: a10a6dc0-7df7-40a1-888a-d1414dbeb1f7] Terminating instance [ 986.039588] env[62066]: DEBUG nova.compute.manager [None req-a924513d-337e-489e-a67d-fcfcb3abed65 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] [instance: a10a6dc0-7df7-40a1-888a-d1414dbeb1f7] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 986.039787] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-a924513d-337e-489e-a67d-fcfcb3abed65 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] [instance: a10a6dc0-7df7-40a1-888a-d1414dbeb1f7] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 986.040667] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7744e62-f189-4d94-afa7-1e998efe230c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.048666] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-a924513d-337e-489e-a67d-fcfcb3abed65 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] [instance: a10a6dc0-7df7-40a1-888a-d1414dbeb1f7] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 986.048893] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1cdb5741-f4f4-4377-97cc-feaa1085cd69 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.055079] env[62066]: DEBUG oslo_vmware.api [None req-a924513d-337e-489e-a67d-fcfcb3abed65 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Waiting for the task: (returnval){ [ 986.055079] env[62066]: value = "task-1341272" [ 986.055079] env[62066]: _type = "Task" [ 986.055079] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.062276] env[62066]: DEBUG oslo_vmware.api [None req-a924513d-337e-489e-a67d-fcfcb3abed65 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Task: {'id': task-1341272, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.065941] env[62066]: DEBUG nova.compute.manager [None req-6ecf159e-1d26-4f9a-b3e2-f54f67ef8e8e tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: a4ef97ad-bbb5-47fa-b5c1-2867a0d56435] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 986.178591] env[62066]: DEBUG oslo_concurrency.lockutils [None req-27337d51-e9e8-4b7e-b8c5-38dbed79187f tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Acquiring lock "a2c5277e-af49-4b31-8480-a74d354c7383" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 986.178861] env[62066]: DEBUG oslo_concurrency.lockutils [None req-27337d51-e9e8-4b7e-b8c5-38dbed79187f tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Lock "a2c5277e-af49-4b31-8480-a74d354c7383" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 986.179130] env[62066]: INFO nova.compute.manager [None req-27337d51-e9e8-4b7e-b8c5-38dbed79187f tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: a2c5277e-af49-4b31-8480-a74d354c7383] Shelving [ 986.267006] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1341271, 'name': CreateVM_Task, 'duration_secs': 0.451841} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.267153] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c8482d92-0180-44ee-a4f1-bf84786dad43] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 986.267929] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c39c0227-c481-4e93-a315-7377013f468c tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 986.268134] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c39c0227-c481-4e93-a315-7377013f468c tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Acquired lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 986.268462] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c39c0227-c481-4e93-a315-7377013f468c tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 986.268722] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1ac7add1-9088-4a4d-ade1-b50c15c69893 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.273030] env[62066]: DEBUG oslo_vmware.api [None req-c39c0227-c481-4e93-a315-7377013f468c tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Waiting for the task: (returnval){ [ 986.273030] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]523ffa49-bad9-d987-7d93-6102839b84c8" [ 986.273030] env[62066]: _type = "Task" [ 986.273030] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.281519] env[62066]: DEBUG nova.scheduler.client.report [None req-b907d10e-116c-4b88-b193-0e79f2464198 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 986.284684] env[62066]: DEBUG oslo_vmware.api [None req-c39c0227-c481-4e93-a315-7377013f468c tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]523ffa49-bad9-d987-7d93-6102839b84c8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.565479] env[62066]: DEBUG oslo_vmware.api [None req-a924513d-337e-489e-a67d-fcfcb3abed65 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Task: {'id': task-1341272, 'name': PowerOffVM_Task, 'duration_secs': 0.184738} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.565819] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-a924513d-337e-489e-a67d-fcfcb3abed65 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] [instance: a10a6dc0-7df7-40a1-888a-d1414dbeb1f7] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 986.566035] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-a924513d-337e-489e-a67d-fcfcb3abed65 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] [instance: a10a6dc0-7df7-40a1-888a-d1414dbeb1f7] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 986.566326] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-57294c68-ebe0-4a16-9881-7955e16c1381 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.643638] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-a924513d-337e-489e-a67d-fcfcb3abed65 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] [instance: a10a6dc0-7df7-40a1-888a-d1414dbeb1f7] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 986.644032] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-a924513d-337e-489e-a67d-fcfcb3abed65 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] [instance: a10a6dc0-7df7-40a1-888a-d1414dbeb1f7] Deleting contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 986.644572] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-a924513d-337e-489e-a67d-fcfcb3abed65 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Deleting the datastore file [datastore2] a10a6dc0-7df7-40a1-888a-d1414dbeb1f7 {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 986.644916] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6d036a66-5e24-464e-9b87-49b6e3eb14d7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.652043] env[62066]: DEBUG oslo_vmware.api [None req-a924513d-337e-489e-a67d-fcfcb3abed65 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Waiting for the task: (returnval){ [ 986.652043] env[62066]: value = "task-1341274" [ 986.652043] env[62066]: _type = "Task" [ 986.652043] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.656205] env[62066]: DEBUG nova.network.neutron [req-994a3df0-e7e1-4806-8f79-01be87326064 req-19704908-7603-4f8a-b152-1f071414944e service nova] [instance: c8482d92-0180-44ee-a4f1-bf84786dad43] Updated VIF entry in instance network info cache for port 2f2f1786-934d-478f-be39-7509cfacf86f. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 986.656563] env[62066]: DEBUG nova.network.neutron [req-994a3df0-e7e1-4806-8f79-01be87326064 req-19704908-7603-4f8a-b152-1f071414944e service nova] [instance: c8482d92-0180-44ee-a4f1-bf84786dad43] Updating instance_info_cache with network_info: [{"id": "2f2f1786-934d-478f-be39-7509cfacf86f", "address": "fa:16:3e:24:41:2e", "network": {"id": "449dfe4e-bb49-43b6-9ba7-b57af74ebfb3", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-282029676-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0d4e33a99c7741fb8cdd97f4ec5dbbd8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db68bd64-5b56-49af-a075-13dcf85cb2e0", "external-id": "nsx-vlan-transportzone-590", "segmentation_id": 590, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2f2f1786-93", "ovs_interfaceid": "2f2f1786-934d-478f-be39-7509cfacf86f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 986.662988] env[62066]: DEBUG oslo_vmware.api [None req-a924513d-337e-489e-a67d-fcfcb3abed65 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Task: {'id': task-1341274, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.686749] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-27337d51-e9e8-4b7e-b8c5-38dbed79187f tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: a2c5277e-af49-4b31-8480-a74d354c7383] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 986.687084] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-595a45fe-61a7-4d65-b545-8a750d2b32c0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.693970] env[62066]: DEBUG oslo_vmware.api [None req-27337d51-e9e8-4b7e-b8c5-38dbed79187f tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Waiting for the task: (returnval){ [ 986.693970] env[62066]: value = "task-1341275" [ 986.693970] env[62066]: _type = "Task" [ 986.693970] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.701678] env[62066]: DEBUG oslo_vmware.api [None req-27337d51-e9e8-4b7e-b8c5-38dbed79187f tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341275, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.783263] env[62066]: DEBUG oslo_vmware.api [None req-c39c0227-c481-4e93-a315-7377013f468c tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]523ffa49-bad9-d987-7d93-6102839b84c8, 'name': SearchDatastore_Task, 'duration_secs': 0.00986} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.783555] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c39c0227-c481-4e93-a315-7377013f468c tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Releasing lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 986.783793] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-c39c0227-c481-4e93-a315-7377013f468c tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: c8482d92-0180-44ee-a4f1-bf84786dad43] Processing image 50ff584c-3b50-4395-af07-3e66769bc9f7 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 986.784061] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c39c0227-c481-4e93-a315-7377013f468c tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 986.784444] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c39c0227-c481-4e93-a315-7377013f468c tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Acquired lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 986.784444] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-c39c0227-c481-4e93-a315-7377013f468c tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 986.784676] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c11fd508-9df7-47e9-b7d9-90963f3ea172 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.796182] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-c39c0227-c481-4e93-a315-7377013f468c tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 986.796385] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-c39c0227-c481-4e93-a315-7377013f468c tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 986.797505] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ec4a635e-d820-4832-81db-4fd2f81e2274 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.806251] env[62066]: DEBUG oslo_vmware.api [None req-c39c0227-c481-4e93-a315-7377013f468c tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Waiting for the task: (returnval){ [ 986.806251] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]52b7f7d2-ff05-b18f-2ff4-e5d15bd0b98d" [ 986.806251] env[62066]: _type = "Task" [ 986.806251] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.815978] env[62066]: DEBUG oslo_vmware.api [None req-c39c0227-c481-4e93-a315-7377013f468c tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52b7f7d2-ff05-b18f-2ff4-e5d15bd0b98d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.076087] env[62066]: DEBUG nova.compute.manager [None req-6ecf159e-1d26-4f9a-b3e2-f54f67ef8e8e tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: a4ef97ad-bbb5-47fa-b5c1-2867a0d56435] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 987.101431] env[62066]: DEBUG nova.virt.hardware [None req-6ecf159e-1d26-4f9a-b3e2-f54f67ef8e8e tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-23T13:40:41Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-23T13:40:23Z,direct_url=,disk_format='vmdk',id=50ff584c-3b50-4395-af07-3e66769bc9f7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='eb52f7069a374c61ae946f052007c6d9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-23T13:40:24Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 987.101721] env[62066]: DEBUG nova.virt.hardware [None req-6ecf159e-1d26-4f9a-b3e2-f54f67ef8e8e tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 987.101886] env[62066]: DEBUG nova.virt.hardware [None req-6ecf159e-1d26-4f9a-b3e2-f54f67ef8e8e tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 987.102090] env[62066]: DEBUG nova.virt.hardware [None req-6ecf159e-1d26-4f9a-b3e2-f54f67ef8e8e tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 987.102246] env[62066]: DEBUG nova.virt.hardware [None req-6ecf159e-1d26-4f9a-b3e2-f54f67ef8e8e tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 987.102398] env[62066]: DEBUG nova.virt.hardware [None req-6ecf159e-1d26-4f9a-b3e2-f54f67ef8e8e tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 987.102611] env[62066]: DEBUG nova.virt.hardware [None req-6ecf159e-1d26-4f9a-b3e2-f54f67ef8e8e tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 987.102776] env[62066]: DEBUG nova.virt.hardware [None req-6ecf159e-1d26-4f9a-b3e2-f54f67ef8e8e tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 987.102946] env[62066]: DEBUG nova.virt.hardware [None req-6ecf159e-1d26-4f9a-b3e2-f54f67ef8e8e tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 987.103127] env[62066]: DEBUG nova.virt.hardware [None req-6ecf159e-1d26-4f9a-b3e2-f54f67ef8e8e tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 987.103302] env[62066]: DEBUG nova.virt.hardware [None req-6ecf159e-1d26-4f9a-b3e2-f54f67ef8e8e tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 987.104147] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f2e1bfa-617c-4bdc-9a05-3a39bda6d759 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.111853] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-735e31e6-cf0d-43d0-afdf-9806052dd84d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.161250] env[62066]: DEBUG oslo_vmware.api [None req-a924513d-337e-489e-a67d-fcfcb3abed65 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Task: {'id': task-1341274, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.141373} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.161535] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-a924513d-337e-489e-a67d-fcfcb3abed65 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 987.161756] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-a924513d-337e-489e-a67d-fcfcb3abed65 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] [instance: a10a6dc0-7df7-40a1-888a-d1414dbeb1f7] Deleted contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 987.161941] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-a924513d-337e-489e-a67d-fcfcb3abed65 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] [instance: a10a6dc0-7df7-40a1-888a-d1414dbeb1f7] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 987.162134] env[62066]: INFO nova.compute.manager [None req-a924513d-337e-489e-a67d-fcfcb3abed65 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] [instance: a10a6dc0-7df7-40a1-888a-d1414dbeb1f7] Took 1.12 seconds to destroy the instance on the hypervisor. [ 987.162418] env[62066]: DEBUG oslo.service.loopingcall [None req-a924513d-337e-489e-a67d-fcfcb3abed65 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 987.162605] env[62066]: DEBUG nova.compute.manager [-] [instance: a10a6dc0-7df7-40a1-888a-d1414dbeb1f7] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 987.162701] env[62066]: DEBUG nova.network.neutron [-] [instance: a10a6dc0-7df7-40a1-888a-d1414dbeb1f7] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 987.164474] env[62066]: DEBUG oslo_concurrency.lockutils [req-994a3df0-e7e1-4806-8f79-01be87326064 req-19704908-7603-4f8a-b152-1f071414944e service nova] Releasing lock "refresh_cache-c8482d92-0180-44ee-a4f1-bf84786dad43" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 987.203735] env[62066]: DEBUG oslo_vmware.api [None req-27337d51-e9e8-4b7e-b8c5-38dbed79187f tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341275, 'name': PowerOffVM_Task, 'duration_secs': 0.187713} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.203735] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-27337d51-e9e8-4b7e-b8c5-38dbed79187f tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: a2c5277e-af49-4b31-8480-a74d354c7383] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 987.204479] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34e2437f-b845-4a08-b39b-4679645f19a1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.223354] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cba4cb58-350d-460e-8b98-62e338a2ef96 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.292709] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b907d10e-116c-4b88-b193-0e79f2464198 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.236s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 987.316819] env[62066]: DEBUG oslo_vmware.api [None req-c39c0227-c481-4e93-a315-7377013f468c tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52b7f7d2-ff05-b18f-2ff4-e5d15bd0b98d, 'name': SearchDatastore_Task, 'duration_secs': 0.010589} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.317228] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bbc786e8-c212-4a43-aa01-00090098477b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.322289] env[62066]: DEBUG oslo_vmware.api [None req-c39c0227-c481-4e93-a315-7377013f468c tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Waiting for the task: (returnval){ [ 987.322289] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]52c2b157-e8af-38bb-e6bd-3702e7c9c9c7" [ 987.322289] env[62066]: _type = "Task" [ 987.322289] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.330675] env[62066]: DEBUG oslo_vmware.api [None req-c39c0227-c481-4e93-a315-7377013f468c tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52c2b157-e8af-38bb-e6bd-3702e7c9c9c7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.419925] env[62066]: DEBUG nova.compute.manager [req-1b9d6fe4-c03e-43b9-a2a3-6af20185c451 req-5feae974-c14e-4bf0-bb3a-6edacbffa039 service nova] [instance: a4ef97ad-bbb5-47fa-b5c1-2867a0d56435] Received event network-vif-plugged-4fd312f4-e41d-45eb-8505-fb30e3596b0a {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 987.420137] env[62066]: DEBUG oslo_concurrency.lockutils [req-1b9d6fe4-c03e-43b9-a2a3-6af20185c451 req-5feae974-c14e-4bf0-bb3a-6edacbffa039 service nova] Acquiring lock "a4ef97ad-bbb5-47fa-b5c1-2867a0d56435-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 987.420359] env[62066]: DEBUG oslo_concurrency.lockutils [req-1b9d6fe4-c03e-43b9-a2a3-6af20185c451 req-5feae974-c14e-4bf0-bb3a-6edacbffa039 service nova] Lock "a4ef97ad-bbb5-47fa-b5c1-2867a0d56435-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 987.420557] env[62066]: DEBUG oslo_concurrency.lockutils [req-1b9d6fe4-c03e-43b9-a2a3-6af20185c451 req-5feae974-c14e-4bf0-bb3a-6edacbffa039 service nova] Lock "a4ef97ad-bbb5-47fa-b5c1-2867a0d56435-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 987.420761] env[62066]: DEBUG nova.compute.manager [req-1b9d6fe4-c03e-43b9-a2a3-6af20185c451 req-5feae974-c14e-4bf0-bb3a-6edacbffa039 service nova] [instance: a4ef97ad-bbb5-47fa-b5c1-2867a0d56435] No waiting events found dispatching network-vif-plugged-4fd312f4-e41d-45eb-8505-fb30e3596b0a {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 987.420941] env[62066]: WARNING nova.compute.manager [req-1b9d6fe4-c03e-43b9-a2a3-6af20185c451 req-5feae974-c14e-4bf0-bb3a-6edacbffa039 service nova] [instance: a4ef97ad-bbb5-47fa-b5c1-2867a0d56435] Received unexpected event network-vif-plugged-4fd312f4-e41d-45eb-8505-fb30e3596b0a for instance with vm_state building and task_state spawning. [ 987.627188] env[62066]: DEBUG nova.compute.manager [req-cacec4b3-4250-46a0-a811-11c7ef268c56 req-0b4fe77b-d670-422a-8853-44b97071c2d7 service nova] [instance: a10a6dc0-7df7-40a1-888a-d1414dbeb1f7] Received event network-vif-deleted-26148c05-d41e-4564-a159-de3f90c20550 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 987.627419] env[62066]: INFO nova.compute.manager [req-cacec4b3-4250-46a0-a811-11c7ef268c56 req-0b4fe77b-d670-422a-8853-44b97071c2d7 service nova] [instance: a10a6dc0-7df7-40a1-888a-d1414dbeb1f7] Neutron deleted interface 26148c05-d41e-4564-a159-de3f90c20550; detaching it from the instance and deleting it from the info cache [ 987.627621] env[62066]: DEBUG nova.network.neutron [req-cacec4b3-4250-46a0-a811-11c7ef268c56 req-0b4fe77b-d670-422a-8853-44b97071c2d7 service nova] [instance: a10a6dc0-7df7-40a1-888a-d1414dbeb1f7] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 987.733073] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-27337d51-e9e8-4b7e-b8c5-38dbed79187f tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: a2c5277e-af49-4b31-8480-a74d354c7383] Creating Snapshot of the VM instance {{(pid=62066) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 987.733402] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-dcf2d43a-0a1a-4f3a-9f45-5fa683016043 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.741273] env[62066]: DEBUG oslo_vmware.api [None req-27337d51-e9e8-4b7e-b8c5-38dbed79187f tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Waiting for the task: (returnval){ [ 987.741273] env[62066]: value = "task-1341276" [ 987.741273] env[62066]: _type = "Task" [ 987.741273] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.749888] env[62066]: DEBUG oslo_vmware.api [None req-27337d51-e9e8-4b7e-b8c5-38dbed79187f tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341276, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.838413] env[62066]: DEBUG oslo_vmware.api [None req-c39c0227-c481-4e93-a315-7377013f468c tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52c2b157-e8af-38bb-e6bd-3702e7c9c9c7, 'name': SearchDatastore_Task, 'duration_secs': 0.029505} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.838915] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c39c0227-c481-4e93-a315-7377013f468c tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Releasing lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 987.839468] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-c39c0227-c481-4e93-a315-7377013f468c tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk to [datastore2] c8482d92-0180-44ee-a4f1-bf84786dad43/c8482d92-0180-44ee-a4f1-bf84786dad43.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 987.840967] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3eab7a2b-ed57-4364-a873-1cef94815919 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.849131] env[62066]: DEBUG oslo_vmware.api [None req-c39c0227-c481-4e93-a315-7377013f468c tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Waiting for the task: (returnval){ [ 987.849131] env[62066]: value = "task-1341277" [ 987.849131] env[62066]: _type = "Task" [ 987.849131] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.856767] env[62066]: DEBUG oslo_vmware.api [None req-c39c0227-c481-4e93-a315-7377013f468c tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': task-1341277, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.873654] env[62066]: INFO nova.scheduler.client.report [None req-b907d10e-116c-4b88-b193-0e79f2464198 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Deleted allocation for migration 1253aef0-8d14-4b3f-99eb-ce11746d0cb8 [ 987.998063] env[62066]: DEBUG nova.network.neutron [None req-6ecf159e-1d26-4f9a-b3e2-f54f67ef8e8e tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: a4ef97ad-bbb5-47fa-b5c1-2867a0d56435] Successfully updated port: 4fd312f4-e41d-45eb-8505-fb30e3596b0a {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 988.023052] env[62066]: DEBUG nova.compute.manager [req-59dbcfe6-cc40-40ba-9936-7889214b7ee3 req-f1e6b108-4b89-4c57-8fc6-3806e6300c8f service nova] [instance: a4ef97ad-bbb5-47fa-b5c1-2867a0d56435] Received event network-changed-4fd312f4-e41d-45eb-8505-fb30e3596b0a {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 988.023281] env[62066]: DEBUG nova.compute.manager [req-59dbcfe6-cc40-40ba-9936-7889214b7ee3 req-f1e6b108-4b89-4c57-8fc6-3806e6300c8f service nova] [instance: a4ef97ad-bbb5-47fa-b5c1-2867a0d56435] Refreshing instance network info cache due to event network-changed-4fd312f4-e41d-45eb-8505-fb30e3596b0a. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 988.023508] env[62066]: DEBUG oslo_concurrency.lockutils [req-59dbcfe6-cc40-40ba-9936-7889214b7ee3 req-f1e6b108-4b89-4c57-8fc6-3806e6300c8f service nova] Acquiring lock "refresh_cache-a4ef97ad-bbb5-47fa-b5c1-2867a0d56435" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 988.023656] env[62066]: DEBUG oslo_concurrency.lockutils [req-59dbcfe6-cc40-40ba-9936-7889214b7ee3 req-f1e6b108-4b89-4c57-8fc6-3806e6300c8f service nova] Acquired lock "refresh_cache-a4ef97ad-bbb5-47fa-b5c1-2867a0d56435" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 988.023825] env[62066]: DEBUG nova.network.neutron [req-59dbcfe6-cc40-40ba-9936-7889214b7ee3 req-f1e6b108-4b89-4c57-8fc6-3806e6300c8f service nova] [instance: a4ef97ad-bbb5-47fa-b5c1-2867a0d56435] Refreshing network info cache for port 4fd312f4-e41d-45eb-8505-fb30e3596b0a {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 988.104088] env[62066]: DEBUG nova.network.neutron [-] [instance: a10a6dc0-7df7-40a1-888a-d1414dbeb1f7] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 988.130038] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-673bc158-1f3d-4bce-8b7d-eb3071bf9f14 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.140312] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69f6db3f-5cd8-4416-9252-2d44f07f99c5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.171077] env[62066]: DEBUG nova.compute.manager [req-cacec4b3-4250-46a0-a811-11c7ef268c56 req-0b4fe77b-d670-422a-8853-44b97071c2d7 service nova] [instance: a10a6dc0-7df7-40a1-888a-d1414dbeb1f7] Detach interface failed, port_id=26148c05-d41e-4564-a159-de3f90c20550, reason: Instance a10a6dc0-7df7-40a1-888a-d1414dbeb1f7 could not be found. {{(pid=62066) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 988.251922] env[62066]: DEBUG oslo_vmware.api [None req-27337d51-e9e8-4b7e-b8c5-38dbed79187f tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341276, 'name': CreateSnapshot_Task, 'duration_secs': 0.418566} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.252701] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-27337d51-e9e8-4b7e-b8c5-38dbed79187f tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: a2c5277e-af49-4b31-8480-a74d354c7383] Created Snapshot of the VM instance {{(pid=62066) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 988.253529] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6afee0ca-ff13-4aa0-881d-c97ca00cf4e8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.358449] env[62066]: DEBUG oslo_vmware.api [None req-c39c0227-c481-4e93-a315-7377013f468c tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': task-1341277, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.458067} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.358748] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-c39c0227-c481-4e93-a315-7377013f468c tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk to [datastore2] c8482d92-0180-44ee-a4f1-bf84786dad43/c8482d92-0180-44ee-a4f1-bf84786dad43.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 988.358997] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-c39c0227-c481-4e93-a315-7377013f468c tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: c8482d92-0180-44ee-a4f1-bf84786dad43] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 988.359276] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2c970892-54ce-4502-9e85-a4f1b14c1d3d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.366760] env[62066]: DEBUG oslo_vmware.api [None req-c39c0227-c481-4e93-a315-7377013f468c tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Waiting for the task: (returnval){ [ 988.366760] env[62066]: value = "task-1341278" [ 988.366760] env[62066]: _type = "Task" [ 988.366760] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.374895] env[62066]: DEBUG oslo_vmware.api [None req-c39c0227-c481-4e93-a315-7377013f468c tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': task-1341278, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.379090] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b907d10e-116c-4b88-b193-0e79f2464198 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Lock "ca504be5-209c-456c-af00-d403bfa2d634" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 7.105s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 988.426195] env[62066]: DEBUG oslo_concurrency.lockutils [None req-70deee26-4fa9-4fac-9f76-d8cda1c1fab9 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Acquiring lock "ca504be5-209c-456c-af00-d403bfa2d634" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 988.426455] env[62066]: DEBUG oslo_concurrency.lockutils [None req-70deee26-4fa9-4fac-9f76-d8cda1c1fab9 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Lock "ca504be5-209c-456c-af00-d403bfa2d634" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 988.426669] env[62066]: DEBUG oslo_concurrency.lockutils [None req-70deee26-4fa9-4fac-9f76-d8cda1c1fab9 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Acquiring lock "ca504be5-209c-456c-af00-d403bfa2d634-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 988.426861] env[62066]: DEBUG oslo_concurrency.lockutils [None req-70deee26-4fa9-4fac-9f76-d8cda1c1fab9 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Lock "ca504be5-209c-456c-af00-d403bfa2d634-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 988.427051] env[62066]: DEBUG oslo_concurrency.lockutils [None req-70deee26-4fa9-4fac-9f76-d8cda1c1fab9 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Lock "ca504be5-209c-456c-af00-d403bfa2d634-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 988.429248] env[62066]: INFO nova.compute.manager [None req-70deee26-4fa9-4fac-9f76-d8cda1c1fab9 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: ca504be5-209c-456c-af00-d403bfa2d634] Terminating instance [ 988.431055] env[62066]: DEBUG nova.compute.manager [None req-70deee26-4fa9-4fac-9f76-d8cda1c1fab9 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: ca504be5-209c-456c-af00-d403bfa2d634] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 988.431247] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-70deee26-4fa9-4fac-9f76-d8cda1c1fab9 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: ca504be5-209c-456c-af00-d403bfa2d634] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 988.432128] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd67cdd9-4997-4e58-860b-997a18793518 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.439552] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-70deee26-4fa9-4fac-9f76-d8cda1c1fab9 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: ca504be5-209c-456c-af00-d403bfa2d634] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 988.439791] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f7556184-66f8-4eda-bccd-6e0b61a3ed46 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.445012] env[62066]: DEBUG oslo_vmware.api [None req-70deee26-4fa9-4fac-9f76-d8cda1c1fab9 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Waiting for the task: (returnval){ [ 988.445012] env[62066]: value = "task-1341279" [ 988.445012] env[62066]: _type = "Task" [ 988.445012] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.452304] env[62066]: DEBUG oslo_vmware.api [None req-70deee26-4fa9-4fac-9f76-d8cda1c1fab9 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1341279, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.500507] env[62066]: DEBUG oslo_concurrency.lockutils [None req-6ecf159e-1d26-4f9a-b3e2-f54f67ef8e8e tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Acquiring lock "refresh_cache-a4ef97ad-bbb5-47fa-b5c1-2867a0d56435" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 988.568356] env[62066]: DEBUG nova.network.neutron [req-59dbcfe6-cc40-40ba-9936-7889214b7ee3 req-f1e6b108-4b89-4c57-8fc6-3806e6300c8f service nova] [instance: a4ef97ad-bbb5-47fa-b5c1-2867a0d56435] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 988.607014] env[62066]: INFO nova.compute.manager [-] [instance: a10a6dc0-7df7-40a1-888a-d1414dbeb1f7] Took 1.44 seconds to deallocate network for instance. [ 988.680596] env[62066]: DEBUG nova.network.neutron [req-59dbcfe6-cc40-40ba-9936-7889214b7ee3 req-f1e6b108-4b89-4c57-8fc6-3806e6300c8f service nova] [instance: a4ef97ad-bbb5-47fa-b5c1-2867a0d56435] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 988.773907] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-27337d51-e9e8-4b7e-b8c5-38dbed79187f tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: a2c5277e-af49-4b31-8480-a74d354c7383] Creating linked-clone VM from snapshot {{(pid=62066) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 988.774271] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-aaec2d09-4739-4ea6-b954-46b9dd502ab1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.782751] env[62066]: DEBUG oslo_vmware.api [None req-27337d51-e9e8-4b7e-b8c5-38dbed79187f tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Waiting for the task: (returnval){ [ 988.782751] env[62066]: value = "task-1341280" [ 988.782751] env[62066]: _type = "Task" [ 988.782751] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.790442] env[62066]: DEBUG oslo_vmware.api [None req-27337d51-e9e8-4b7e-b8c5-38dbed79187f tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341280, 'name': CloneVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.877326] env[62066]: DEBUG oslo_vmware.api [None req-c39c0227-c481-4e93-a315-7377013f468c tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': task-1341278, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.057518} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.877834] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-c39c0227-c481-4e93-a315-7377013f468c tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: c8482d92-0180-44ee-a4f1-bf84786dad43] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 988.878275] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e851facc-eafd-4608-b6d6-e4e63643a6c4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.902399] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-c39c0227-c481-4e93-a315-7377013f468c tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: c8482d92-0180-44ee-a4f1-bf84786dad43] Reconfiguring VM instance instance-00000062 to attach disk [datastore2] c8482d92-0180-44ee-a4f1-bf84786dad43/c8482d92-0180-44ee-a4f1-bf84786dad43.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 988.902752] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d61d45f1-695e-43ce-9420-3e0b06ac3acc {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.924562] env[62066]: DEBUG oslo_vmware.api [None req-c39c0227-c481-4e93-a315-7377013f468c tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Waiting for the task: (returnval){ [ 988.924562] env[62066]: value = "task-1341281" [ 988.924562] env[62066]: _type = "Task" [ 988.924562] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.933798] env[62066]: DEBUG oslo_vmware.api [None req-c39c0227-c481-4e93-a315-7377013f468c tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': task-1341281, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.955100] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b3f57e08-3a8e-4c04-9171-e3846694852d tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Acquiring lock "9dc7c179-4360-425d-915f-c2d7bc591b1e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 988.955100] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b3f57e08-3a8e-4c04-9171-e3846694852d tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Lock "9dc7c179-4360-425d-915f-c2d7bc591b1e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 988.963100] env[62066]: DEBUG oslo_vmware.api [None req-70deee26-4fa9-4fac-9f76-d8cda1c1fab9 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1341279, 'name': PowerOffVM_Task, 'duration_secs': 0.212105} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.963100] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-70deee26-4fa9-4fac-9f76-d8cda1c1fab9 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: ca504be5-209c-456c-af00-d403bfa2d634] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 988.963271] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-70deee26-4fa9-4fac-9f76-d8cda1c1fab9 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: ca504be5-209c-456c-af00-d403bfa2d634] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 988.964044] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-110f4eb6-3991-4fe7-a4c4-946f7a751aef {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.036742] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-70deee26-4fa9-4fac-9f76-d8cda1c1fab9 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: ca504be5-209c-456c-af00-d403bfa2d634] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 989.036960] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-70deee26-4fa9-4fac-9f76-d8cda1c1fab9 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: ca504be5-209c-456c-af00-d403bfa2d634] Deleting contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 989.037171] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-70deee26-4fa9-4fac-9f76-d8cda1c1fab9 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Deleting the datastore file [datastore2] ca504be5-209c-456c-af00-d403bfa2d634 {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 989.037449] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-07183566-7755-4d32-b75b-6e8e35738f1f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.043210] env[62066]: DEBUG oslo_vmware.api [None req-70deee26-4fa9-4fac-9f76-d8cda1c1fab9 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Waiting for the task: (returnval){ [ 989.043210] env[62066]: value = "task-1341283" [ 989.043210] env[62066]: _type = "Task" [ 989.043210] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.050589] env[62066]: DEBUG oslo_vmware.api [None req-70deee26-4fa9-4fac-9f76-d8cda1c1fab9 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1341283, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.113111] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a924513d-337e-489e-a67d-fcfcb3abed65 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 989.113375] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a924513d-337e-489e-a67d-fcfcb3abed65 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 989.113639] env[62066]: DEBUG nova.objects.instance [None req-a924513d-337e-489e-a67d-fcfcb3abed65 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Lazy-loading 'resources' on Instance uuid a10a6dc0-7df7-40a1-888a-d1414dbeb1f7 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 989.183485] env[62066]: DEBUG oslo_concurrency.lockutils [req-59dbcfe6-cc40-40ba-9936-7889214b7ee3 req-f1e6b108-4b89-4c57-8fc6-3806e6300c8f service nova] Releasing lock "refresh_cache-a4ef97ad-bbb5-47fa-b5c1-2867a0d56435" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 989.183879] env[62066]: DEBUG oslo_concurrency.lockutils [None req-6ecf159e-1d26-4f9a-b3e2-f54f67ef8e8e tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Acquired lock "refresh_cache-a4ef97ad-bbb5-47fa-b5c1-2867a0d56435" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 989.184054] env[62066]: DEBUG nova.network.neutron [None req-6ecf159e-1d26-4f9a-b3e2-f54f67ef8e8e tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: a4ef97ad-bbb5-47fa-b5c1-2867a0d56435] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 989.292919] env[62066]: DEBUG oslo_vmware.api [None req-27337d51-e9e8-4b7e-b8c5-38dbed79187f tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341280, 'name': CloneVM_Task} progress is 94%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.436519] env[62066]: DEBUG oslo_vmware.api [None req-c39c0227-c481-4e93-a315-7377013f468c tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': task-1341281, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.456496] env[62066]: DEBUG nova.compute.manager [None req-b3f57e08-3a8e-4c04-9171-e3846694852d tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 9dc7c179-4360-425d-915f-c2d7bc591b1e] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 989.553275] env[62066]: DEBUG oslo_vmware.api [None req-70deee26-4fa9-4fac-9f76-d8cda1c1fab9 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1341283, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.137149} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.553571] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-70deee26-4fa9-4fac-9f76-d8cda1c1fab9 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 989.553737] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-70deee26-4fa9-4fac-9f76-d8cda1c1fab9 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: ca504be5-209c-456c-af00-d403bfa2d634] Deleted contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 989.553917] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-70deee26-4fa9-4fac-9f76-d8cda1c1fab9 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: ca504be5-209c-456c-af00-d403bfa2d634] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 989.554118] env[62066]: INFO nova.compute.manager [None req-70deee26-4fa9-4fac-9f76-d8cda1c1fab9 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: ca504be5-209c-456c-af00-d403bfa2d634] Took 1.12 seconds to destroy the instance on the hypervisor. [ 989.554374] env[62066]: DEBUG oslo.service.loopingcall [None req-70deee26-4fa9-4fac-9f76-d8cda1c1fab9 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 989.554572] env[62066]: DEBUG nova.compute.manager [-] [instance: ca504be5-209c-456c-af00-d403bfa2d634] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 989.554671] env[62066]: DEBUG nova.network.neutron [-] [instance: ca504be5-209c-456c-af00-d403bfa2d634] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 989.720951] env[62066]: DEBUG nova.network.neutron [None req-6ecf159e-1d26-4f9a-b3e2-f54f67ef8e8e tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: a4ef97ad-bbb5-47fa-b5c1-2867a0d56435] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 989.735330] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91cac0d2-255d-4b98-b4d8-e0a1b6f3bf88 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.745160] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c570ca3-48e6-4a98-8258-b0396218154b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.780580] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-555324c5-6d74-428c-883e-672b5c2e92fb {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.790907] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58fdc83b-e2b9-49be-9866-804c1b10ca65 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.797906] env[62066]: DEBUG oslo_vmware.api [None req-27337d51-e9e8-4b7e-b8c5-38dbed79187f tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341280, 'name': CloneVM_Task} progress is 94%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.807868] env[62066]: DEBUG nova.compute.provider_tree [None req-a924513d-337e-489e-a67d-fcfcb3abed65 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 989.891796] env[62066]: DEBUG nova.network.neutron [None req-6ecf159e-1d26-4f9a-b3e2-f54f67ef8e8e tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: a4ef97ad-bbb5-47fa-b5c1-2867a0d56435] Updating instance_info_cache with network_info: [{"id": "4fd312f4-e41d-45eb-8505-fb30e3596b0a", "address": "fa:16:3e:c2:1a:05", "network": {"id": "9f3396ce-ee48-4eaf-8cc7-3e1711f37537", "bridge": "br-int", "label": "tempest-ServersTestJSON-2129775093-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6497ab02f327476d8ff81c2ecc0371e5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7ab8d568-adb0-4f3b-b6cc-68413e6546ae", "external-id": "nsx-vlan-transportzone-86", "segmentation_id": 86, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4fd312f4-e4", "ovs_interfaceid": "4fd312f4-e41d-45eb-8505-fb30e3596b0a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 989.937256] env[62066]: DEBUG oslo_vmware.api [None req-c39c0227-c481-4e93-a315-7377013f468c tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': task-1341281, 'name': ReconfigVM_Task, 'duration_secs': 0.782586} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.937566] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-c39c0227-c481-4e93-a315-7377013f468c tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: c8482d92-0180-44ee-a4f1-bf84786dad43] Reconfigured VM instance instance-00000062 to attach disk [datastore2] c8482d92-0180-44ee-a4f1-bf84786dad43/c8482d92-0180-44ee-a4f1-bf84786dad43.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 989.938528] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c6805327-61ad-403b-bf3a-c2905a454ee4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.944321] env[62066]: DEBUG oslo_vmware.api [None req-c39c0227-c481-4e93-a315-7377013f468c tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Waiting for the task: (returnval){ [ 989.944321] env[62066]: value = "task-1341284" [ 989.944321] env[62066]: _type = "Task" [ 989.944321] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.952312] env[62066]: DEBUG oslo_vmware.api [None req-c39c0227-c481-4e93-a315-7377013f468c tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': task-1341284, 'name': Rename_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.976208] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b3f57e08-3a8e-4c04-9171-e3846694852d tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 990.053232] env[62066]: DEBUG nova.compute.manager [req-0bdfefd6-5bc0-48b6-8925-d557a54fa6d9 req-c734b267-4d55-402a-83d5-e62bfa6afce3 service nova] [instance: ca504be5-209c-456c-af00-d403bfa2d634] Received event network-vif-deleted-11deacd3-d3a5-49a4-98f3-3226eb532e76 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 990.053449] env[62066]: INFO nova.compute.manager [req-0bdfefd6-5bc0-48b6-8925-d557a54fa6d9 req-c734b267-4d55-402a-83d5-e62bfa6afce3 service nova] [instance: ca504be5-209c-456c-af00-d403bfa2d634] Neutron deleted interface 11deacd3-d3a5-49a4-98f3-3226eb532e76; detaching it from the instance and deleting it from the info cache [ 990.053687] env[62066]: DEBUG nova.network.neutron [req-0bdfefd6-5bc0-48b6-8925-d557a54fa6d9 req-c734b267-4d55-402a-83d5-e62bfa6afce3 service nova] [instance: ca504be5-209c-456c-af00-d403bfa2d634] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 990.295015] env[62066]: DEBUG oslo_vmware.api [None req-27337d51-e9e8-4b7e-b8c5-38dbed79187f tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341280, 'name': CloneVM_Task, 'duration_secs': 1.379986} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.295335] env[62066]: INFO nova.virt.vmwareapi.vmops [None req-27337d51-e9e8-4b7e-b8c5-38dbed79187f tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: a2c5277e-af49-4b31-8480-a74d354c7383] Created linked-clone VM from snapshot [ 990.296112] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74c9f7e6-e2e0-444a-917d-d08ef80ce694 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.303625] env[62066]: DEBUG nova.virt.vmwareapi.images [None req-27337d51-e9e8-4b7e-b8c5-38dbed79187f tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: a2c5277e-af49-4b31-8480-a74d354c7383] Uploading image e4ed31bf-18c9-4cc2-9bff-65cfb25f04fc {{(pid=62066) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 990.310221] env[62066]: DEBUG nova.scheduler.client.report [None req-a924513d-337e-489e-a67d-fcfcb3abed65 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 990.324525] env[62066]: DEBUG nova.network.neutron [-] [instance: ca504be5-209c-456c-af00-d403bfa2d634] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 990.329071] env[62066]: DEBUG oslo_vmware.rw_handles [None req-27337d51-e9e8-4b7e-b8c5-38dbed79187f tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 990.329071] env[62066]: value = "vm-286121" [ 990.329071] env[62066]: _type = "VirtualMachine" [ 990.329071] env[62066]: }. {{(pid=62066) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 990.329335] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-cb030fd5-bafc-47c4-b738-2dcfb56e67e6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.335837] env[62066]: DEBUG oslo_vmware.rw_handles [None req-27337d51-e9e8-4b7e-b8c5-38dbed79187f tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Lease: (returnval){ [ 990.335837] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]52e07b1d-338a-5cc5-1a5a-46a25d470b55" [ 990.335837] env[62066]: _type = "HttpNfcLease" [ 990.335837] env[62066]: } obtained for exporting VM: (result){ [ 990.335837] env[62066]: value = "vm-286121" [ 990.335837] env[62066]: _type = "VirtualMachine" [ 990.335837] env[62066]: }. {{(pid=62066) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 990.336094] env[62066]: DEBUG oslo_vmware.api [None req-27337d51-e9e8-4b7e-b8c5-38dbed79187f tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Waiting for the lease: (returnval){ [ 990.336094] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]52e07b1d-338a-5cc5-1a5a-46a25d470b55" [ 990.336094] env[62066]: _type = "HttpNfcLease" [ 990.336094] env[62066]: } to be ready. {{(pid=62066) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 990.342151] env[62066]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 990.342151] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]52e07b1d-338a-5cc5-1a5a-46a25d470b55" [ 990.342151] env[62066]: _type = "HttpNfcLease" [ 990.342151] env[62066]: } is initializing. {{(pid=62066) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 990.394866] env[62066]: DEBUG oslo_concurrency.lockutils [None req-6ecf159e-1d26-4f9a-b3e2-f54f67ef8e8e tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Releasing lock "refresh_cache-a4ef97ad-bbb5-47fa-b5c1-2867a0d56435" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 990.395243] env[62066]: DEBUG nova.compute.manager [None req-6ecf159e-1d26-4f9a-b3e2-f54f67ef8e8e tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: a4ef97ad-bbb5-47fa-b5c1-2867a0d56435] Instance network_info: |[{"id": "4fd312f4-e41d-45eb-8505-fb30e3596b0a", "address": "fa:16:3e:c2:1a:05", "network": {"id": "9f3396ce-ee48-4eaf-8cc7-3e1711f37537", "bridge": "br-int", "label": "tempest-ServersTestJSON-2129775093-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6497ab02f327476d8ff81c2ecc0371e5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7ab8d568-adb0-4f3b-b6cc-68413e6546ae", "external-id": "nsx-vlan-transportzone-86", "segmentation_id": 86, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4fd312f4-e4", "ovs_interfaceid": "4fd312f4-e41d-45eb-8505-fb30e3596b0a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 990.395671] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-6ecf159e-1d26-4f9a-b3e2-f54f67ef8e8e tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: a4ef97ad-bbb5-47fa-b5c1-2867a0d56435] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c2:1a:05', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7ab8d568-adb0-4f3b-b6cc-68413e6546ae', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4fd312f4-e41d-45eb-8505-fb30e3596b0a', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 990.403306] env[62066]: DEBUG oslo.service.loopingcall [None req-6ecf159e-1d26-4f9a-b3e2-f54f67ef8e8e tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 990.403787] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a4ef97ad-bbb5-47fa-b5c1-2867a0d56435] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 990.404031] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-39415205-c868-4710-862e-c3d50322d534 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.422617] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 990.422617] env[62066]: value = "task-1341286" [ 990.422617] env[62066]: _type = "Task" [ 990.422617] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.430968] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1341286, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.452823] env[62066]: DEBUG oslo_vmware.api [None req-c39c0227-c481-4e93-a315-7377013f468c tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': task-1341284, 'name': Rename_Task, 'duration_secs': 0.137319} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.453138] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-c39c0227-c481-4e93-a315-7377013f468c tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: c8482d92-0180-44ee-a4f1-bf84786dad43] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 990.453398] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7cfe4df9-9fd5-43b4-8226-112d596b3ea7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.459324] env[62066]: DEBUG oslo_vmware.api [None req-c39c0227-c481-4e93-a315-7377013f468c tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Waiting for the task: (returnval){ [ 990.459324] env[62066]: value = "task-1341287" [ 990.459324] env[62066]: _type = "Task" [ 990.459324] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.466811] env[62066]: DEBUG oslo_vmware.api [None req-c39c0227-c481-4e93-a315-7377013f468c tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': task-1341287, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.556503] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b24215e1-c854-4059-a34b-d7689b8ee58e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.566385] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0797ad00-1265-40d5-9343-e5cf144e8c24 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.593608] env[62066]: DEBUG nova.compute.manager [req-0bdfefd6-5bc0-48b6-8925-d557a54fa6d9 req-c734b267-4d55-402a-83d5-e62bfa6afce3 service nova] [instance: ca504be5-209c-456c-af00-d403bfa2d634] Detach interface failed, port_id=11deacd3-d3a5-49a4-98f3-3226eb532e76, reason: Instance ca504be5-209c-456c-af00-d403bfa2d634 could not be found. {{(pid=62066) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 990.815569] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a924513d-337e-489e-a67d-fcfcb3abed65 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.702s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 990.819196] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b3f57e08-3a8e-4c04-9171-e3846694852d tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.842s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 990.820803] env[62066]: INFO nova.compute.claims [None req-b3f57e08-3a8e-4c04-9171-e3846694852d tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 9dc7c179-4360-425d-915f-c2d7bc591b1e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 990.826120] env[62066]: INFO nova.compute.manager [-] [instance: ca504be5-209c-456c-af00-d403bfa2d634] Took 1.27 seconds to deallocate network for instance. [ 990.836700] env[62066]: INFO nova.scheduler.client.report [None req-a924513d-337e-489e-a67d-fcfcb3abed65 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Deleted allocations for instance a10a6dc0-7df7-40a1-888a-d1414dbeb1f7 [ 990.848694] env[62066]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 990.848694] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]52e07b1d-338a-5cc5-1a5a-46a25d470b55" [ 990.848694] env[62066]: _type = "HttpNfcLease" [ 990.848694] env[62066]: } is ready. {{(pid=62066) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 990.849049] env[62066]: DEBUG oslo_vmware.rw_handles [None req-27337d51-e9e8-4b7e-b8c5-38dbed79187f tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 990.849049] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]52e07b1d-338a-5cc5-1a5a-46a25d470b55" [ 990.849049] env[62066]: _type = "HttpNfcLease" [ 990.849049] env[62066]: }. {{(pid=62066) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 990.849792] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3849045-f80a-41da-abe5-483341adee90 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.858196] env[62066]: DEBUG oslo_vmware.rw_handles [None req-27337d51-e9e8-4b7e-b8c5-38dbed79187f tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/525c4626-1f31-8d3e-bb46-aa44cac71c21/disk-0.vmdk from lease info. {{(pid=62066) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 990.858607] env[62066]: DEBUG oslo_vmware.rw_handles [None req-27337d51-e9e8-4b7e-b8c5-38dbed79187f tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/525c4626-1f31-8d3e-bb46-aa44cac71c21/disk-0.vmdk for reading. {{(pid=62066) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 990.932762] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1341286, 'name': CreateVM_Task, 'duration_secs': 0.331614} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.932997] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a4ef97ad-bbb5-47fa-b5c1-2867a0d56435] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 990.933759] env[62066]: DEBUG oslo_concurrency.lockutils [None req-6ecf159e-1d26-4f9a-b3e2-f54f67ef8e8e tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 990.933961] env[62066]: DEBUG oslo_concurrency.lockutils [None req-6ecf159e-1d26-4f9a-b3e2-f54f67ef8e8e tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Acquired lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 990.934346] env[62066]: DEBUG oslo_concurrency.lockutils [None req-6ecf159e-1d26-4f9a-b3e2-f54f67ef8e8e tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 990.934623] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ee0b40b3-5324-4c06-b3b2-7f33dff6d95e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.939331] env[62066]: DEBUG oslo_vmware.api [None req-6ecf159e-1d26-4f9a-b3e2-f54f67ef8e8e tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Waiting for the task: (returnval){ [ 990.939331] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]5215ed81-789d-9ba0-76e5-f806480f7cbf" [ 990.939331] env[62066]: _type = "Task" [ 990.939331] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.947254] env[62066]: DEBUG oslo_vmware.api [None req-6ecf159e-1d26-4f9a-b3e2-f54f67ef8e8e tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]5215ed81-789d-9ba0-76e5-f806480f7cbf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.968677] env[62066]: DEBUG oslo_vmware.api [None req-c39c0227-c481-4e93-a315-7377013f468c tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': task-1341287, 'name': PowerOnVM_Task, 'duration_secs': 0.477264} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.969029] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-c39c0227-c481-4e93-a315-7377013f468c tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: c8482d92-0180-44ee-a4f1-bf84786dad43] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 990.969285] env[62066]: INFO nova.compute.manager [None req-c39c0227-c481-4e93-a315-7377013f468c tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: c8482d92-0180-44ee-a4f1-bf84786dad43] Took 7.23 seconds to spawn the instance on the hypervisor. [ 990.969515] env[62066]: DEBUG nova.compute.manager [None req-c39c0227-c481-4e93-a315-7377013f468c tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: c8482d92-0180-44ee-a4f1-bf84786dad43] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 990.970310] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac3503a2-a511-4110-9940-c9bfebbb1433 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.993476] env[62066]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-35aed056-8444-4af5-b40a-906f1a78747b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.332883] env[62066]: DEBUG oslo_concurrency.lockutils [None req-70deee26-4fa9-4fac-9f76-d8cda1c1fab9 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 991.348039] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a924513d-337e-489e-a67d-fcfcb3abed65 tempest-AttachVolumeNegativeTest-1223289584 tempest-AttachVolumeNegativeTest-1223289584-project-member] Lock "a10a6dc0-7df7-40a1-888a-d1414dbeb1f7" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.313s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 991.450252] env[62066]: DEBUG oslo_vmware.api [None req-6ecf159e-1d26-4f9a-b3e2-f54f67ef8e8e tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]5215ed81-789d-9ba0-76e5-f806480f7cbf, 'name': SearchDatastore_Task, 'duration_secs': 0.011377} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.450728] env[62066]: DEBUG oslo_concurrency.lockutils [None req-6ecf159e-1d26-4f9a-b3e2-f54f67ef8e8e tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Releasing lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 991.450991] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-6ecf159e-1d26-4f9a-b3e2-f54f67ef8e8e tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: a4ef97ad-bbb5-47fa-b5c1-2867a0d56435] Processing image 50ff584c-3b50-4395-af07-3e66769bc9f7 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 991.451260] env[62066]: DEBUG oslo_concurrency.lockutils [None req-6ecf159e-1d26-4f9a-b3e2-f54f67ef8e8e tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 991.451431] env[62066]: DEBUG oslo_concurrency.lockutils [None req-6ecf159e-1d26-4f9a-b3e2-f54f67ef8e8e tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Acquired lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 991.451981] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-6ecf159e-1d26-4f9a-b3e2-f54f67ef8e8e tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 991.452107] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0594bf2e-6522-45bb-abc0-203c6b7f552f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.460412] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-6ecf159e-1d26-4f9a-b3e2-f54f67ef8e8e tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 991.460650] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-6ecf159e-1d26-4f9a-b3e2-f54f67ef8e8e tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 991.461430] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1be6cd32-5ec2-49df-a693-8224ea683f5f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.466703] env[62066]: DEBUG oslo_vmware.api [None req-6ecf159e-1d26-4f9a-b3e2-f54f67ef8e8e tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Waiting for the task: (returnval){ [ 991.466703] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]52ffb65a-dbdd-c6bd-980c-58f5c3b4ab76" [ 991.466703] env[62066]: _type = "Task" [ 991.466703] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.474318] env[62066]: DEBUG oslo_vmware.api [None req-6ecf159e-1d26-4f9a-b3e2-f54f67ef8e8e tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52ffb65a-dbdd-c6bd-980c-58f5c3b4ab76, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.490293] env[62066]: INFO nova.compute.manager [None req-c39c0227-c481-4e93-a315-7377013f468c tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: c8482d92-0180-44ee-a4f1-bf84786dad43] Took 13.30 seconds to build instance. [ 991.942396] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-924c6ed2-c021-4578-87f9-89f2ac4f1693 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.949987] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09c76748-819c-4506-ae78-a0534d93555c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.982896] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4304be63-0879-4069-875f-0b413fe76299 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.990833] env[62066]: DEBUG oslo_vmware.api [None req-6ecf159e-1d26-4f9a-b3e2-f54f67ef8e8e tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52ffb65a-dbdd-c6bd-980c-58f5c3b4ab76, 'name': SearchDatastore_Task, 'duration_secs': 0.008436} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.993559] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c39c0227-c481-4e93-a315-7377013f468c tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Lock "c8482d92-0180-44ee-a4f1-bf84786dad43" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.814s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 991.993934] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-76aefc1f-8931-4209-9ca0-1d2c2613557b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.997092] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a355957a-6682-4940-af66-2f300892ceb1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.003622] env[62066]: DEBUG oslo_vmware.api [None req-6ecf159e-1d26-4f9a-b3e2-f54f67ef8e8e tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Waiting for the task: (returnval){ [ 992.003622] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]52a26a9d-6276-d456-449f-ad161f37568e" [ 992.003622] env[62066]: _type = "Task" [ 992.003622] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.011486] env[62066]: DEBUG nova.compute.provider_tree [None req-b3f57e08-3a8e-4c04-9171-e3846694852d tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 992.020762] env[62066]: DEBUG oslo_vmware.api [None req-6ecf159e-1d26-4f9a-b3e2-f54f67ef8e8e tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52a26a9d-6276-d456-449f-ad161f37568e, 'name': SearchDatastore_Task, 'duration_secs': 0.008768} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.022184] env[62066]: DEBUG oslo_concurrency.lockutils [None req-6ecf159e-1d26-4f9a-b3e2-f54f67ef8e8e tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Releasing lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 992.022446] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-6ecf159e-1d26-4f9a-b3e2-f54f67ef8e8e tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk to [datastore2] a4ef97ad-bbb5-47fa-b5c1-2867a0d56435/a4ef97ad-bbb5-47fa-b5c1-2867a0d56435.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 992.022701] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d3c36743-73c4-4698-86a0-749c7e3cc1a9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.031898] env[62066]: DEBUG oslo_vmware.api [None req-6ecf159e-1d26-4f9a-b3e2-f54f67ef8e8e tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Waiting for the task: (returnval){ [ 992.031898] env[62066]: value = "task-1341289" [ 992.031898] env[62066]: _type = "Task" [ 992.031898] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.040736] env[62066]: DEBUG oslo_vmware.api [None req-6ecf159e-1d26-4f9a-b3e2-f54f67ef8e8e tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': task-1341289, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.517695] env[62066]: DEBUG nova.scheduler.client.report [None req-b3f57e08-3a8e-4c04-9171-e3846694852d tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 992.546821] env[62066]: DEBUG oslo_vmware.api [None req-6ecf159e-1d26-4f9a-b3e2-f54f67ef8e8e tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': task-1341289, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.446802} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.551606] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-6ecf159e-1d26-4f9a-b3e2-f54f67ef8e8e tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk to [datastore2] a4ef97ad-bbb5-47fa-b5c1-2867a0d56435/a4ef97ad-bbb5-47fa-b5c1-2867a0d56435.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 992.551917] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-6ecf159e-1d26-4f9a-b3e2-f54f67ef8e8e tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: a4ef97ad-bbb5-47fa-b5c1-2867a0d56435] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 992.552502] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d248e1ea-c6f4-4b46-81b6-68904bce3ce1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.558954] env[62066]: DEBUG oslo_vmware.api [None req-6ecf159e-1d26-4f9a-b3e2-f54f67ef8e8e tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Waiting for the task: (returnval){ [ 992.558954] env[62066]: value = "task-1341290" [ 992.558954] env[62066]: _type = "Task" [ 992.558954] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.568635] env[62066]: DEBUG oslo_vmware.api [None req-6ecf159e-1d26-4f9a-b3e2-f54f67ef8e8e tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': task-1341290, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.888232] env[62066]: DEBUG nova.compute.manager [req-525dd1bd-5120-421c-8c05-c3d226ab42c3 req-eea1cb57-f610-4a26-9143-676c911adfcc service nova] [instance: c8482d92-0180-44ee-a4f1-bf84786dad43] Received event network-changed-2f2f1786-934d-478f-be39-7509cfacf86f {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 992.888232] env[62066]: DEBUG nova.compute.manager [req-525dd1bd-5120-421c-8c05-c3d226ab42c3 req-eea1cb57-f610-4a26-9143-676c911adfcc service nova] [instance: c8482d92-0180-44ee-a4f1-bf84786dad43] Refreshing instance network info cache due to event network-changed-2f2f1786-934d-478f-be39-7509cfacf86f. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 992.888426] env[62066]: DEBUG oslo_concurrency.lockutils [req-525dd1bd-5120-421c-8c05-c3d226ab42c3 req-eea1cb57-f610-4a26-9143-676c911adfcc service nova] Acquiring lock "refresh_cache-c8482d92-0180-44ee-a4f1-bf84786dad43" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 992.888654] env[62066]: DEBUG oslo_concurrency.lockutils [req-525dd1bd-5120-421c-8c05-c3d226ab42c3 req-eea1cb57-f610-4a26-9143-676c911adfcc service nova] Acquired lock "refresh_cache-c8482d92-0180-44ee-a4f1-bf84786dad43" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 992.888878] env[62066]: DEBUG nova.network.neutron [req-525dd1bd-5120-421c-8c05-c3d226ab42c3 req-eea1cb57-f610-4a26-9143-676c911adfcc service nova] [instance: c8482d92-0180-44ee-a4f1-bf84786dad43] Refreshing network info cache for port 2f2f1786-934d-478f-be39-7509cfacf86f {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 993.024339] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b3f57e08-3a8e-4c04-9171-e3846694852d tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.206s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 993.024886] env[62066]: DEBUG nova.compute.manager [None req-b3f57e08-3a8e-4c04-9171-e3846694852d tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 9dc7c179-4360-425d-915f-c2d7bc591b1e] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 993.027517] env[62066]: DEBUG oslo_concurrency.lockutils [None req-70deee26-4fa9-4fac-9f76-d8cda1c1fab9 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.695s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 993.027722] env[62066]: DEBUG oslo_concurrency.lockutils [None req-70deee26-4fa9-4fac-9f76-d8cda1c1fab9 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 993.054471] env[62066]: INFO nova.scheduler.client.report [None req-70deee26-4fa9-4fac-9f76-d8cda1c1fab9 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Deleted allocations for instance ca504be5-209c-456c-af00-d403bfa2d634 [ 993.070432] env[62066]: DEBUG oslo_vmware.api [None req-6ecf159e-1d26-4f9a-b3e2-f54f67ef8e8e tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': task-1341290, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.097505} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.070792] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-6ecf159e-1d26-4f9a-b3e2-f54f67ef8e8e tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: a4ef97ad-bbb5-47fa-b5c1-2867a0d56435] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 993.072099] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea6b6efa-8bbe-4534-b9f0-8b8651ed49ef {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.096435] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-6ecf159e-1d26-4f9a-b3e2-f54f67ef8e8e tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: a4ef97ad-bbb5-47fa-b5c1-2867a0d56435] Reconfiguring VM instance instance-00000063 to attach disk [datastore2] a4ef97ad-bbb5-47fa-b5c1-2867a0d56435/a4ef97ad-bbb5-47fa-b5c1-2867a0d56435.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 993.097955] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4a92dcf9-4b23-42f3-82ec-eac7dec4a35b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.117303] env[62066]: DEBUG oslo_vmware.api [None req-6ecf159e-1d26-4f9a-b3e2-f54f67ef8e8e tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Waiting for the task: (returnval){ [ 993.117303] env[62066]: value = "task-1341291" [ 993.117303] env[62066]: _type = "Task" [ 993.117303] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.125919] env[62066]: DEBUG oslo_vmware.api [None req-6ecf159e-1d26-4f9a-b3e2-f54f67ef8e8e tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': task-1341291, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.535764] env[62066]: DEBUG nova.compute.utils [None req-b3f57e08-3a8e-4c04-9171-e3846694852d tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 993.540280] env[62066]: DEBUG nova.compute.manager [None req-b3f57e08-3a8e-4c04-9171-e3846694852d tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 9dc7c179-4360-425d-915f-c2d7bc591b1e] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 993.540606] env[62066]: DEBUG nova.network.neutron [None req-b3f57e08-3a8e-4c04-9171-e3846694852d tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 9dc7c179-4360-425d-915f-c2d7bc591b1e] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 993.565083] env[62066]: DEBUG oslo_concurrency.lockutils [None req-70deee26-4fa9-4fac-9f76-d8cda1c1fab9 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Lock "ca504be5-209c-456c-af00-d403bfa2d634" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.138s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 993.611682] env[62066]: DEBUG nova.policy [None req-b3f57e08-3a8e-4c04-9171-e3846694852d tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2ab20cbd0bab4ae7ba46d9135605a509', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd609babc987a42e2a8ddb4bfb9c3b103', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 993.627361] env[62066]: DEBUG oslo_vmware.api [None req-6ecf159e-1d26-4f9a-b3e2-f54f67ef8e8e tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': task-1341291, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.642913] env[62066]: DEBUG nova.network.neutron [req-525dd1bd-5120-421c-8c05-c3d226ab42c3 req-eea1cb57-f610-4a26-9143-676c911adfcc service nova] [instance: c8482d92-0180-44ee-a4f1-bf84786dad43] Updated VIF entry in instance network info cache for port 2f2f1786-934d-478f-be39-7509cfacf86f. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 993.646988] env[62066]: DEBUG nova.network.neutron [req-525dd1bd-5120-421c-8c05-c3d226ab42c3 req-eea1cb57-f610-4a26-9143-676c911adfcc service nova] [instance: c8482d92-0180-44ee-a4f1-bf84786dad43] Updating instance_info_cache with network_info: [{"id": "2f2f1786-934d-478f-be39-7509cfacf86f", "address": "fa:16:3e:24:41:2e", "network": {"id": "449dfe4e-bb49-43b6-9ba7-b57af74ebfb3", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-282029676-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.231", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0d4e33a99c7741fb8cdd97f4ec5dbbd8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db68bd64-5b56-49af-a075-13dcf85cb2e0", "external-id": "nsx-vlan-transportzone-590", "segmentation_id": 590, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2f2f1786-93", "ovs_interfaceid": "2f2f1786-934d-478f-be39-7509cfacf86f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 993.809547] env[62066]: DEBUG oslo_concurrency.lockutils [None req-2cce9c2e-a55d-474b-be2b-37d32fa31b9b tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Acquiring lock "7178eb83-41e7-4516-809d-a0972d7e6e23" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 993.809793] env[62066]: DEBUG oslo_concurrency.lockutils [None req-2cce9c2e-a55d-474b-be2b-37d32fa31b9b tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Lock "7178eb83-41e7-4516-809d-a0972d7e6e23" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 993.982222] env[62066]: DEBUG nova.network.neutron [None req-b3f57e08-3a8e-4c04-9171-e3846694852d tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 9dc7c179-4360-425d-915f-c2d7bc591b1e] Successfully created port: 8499c7af-d27f-4b10-b63d-a50b2a34d620 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 994.040157] env[62066]: DEBUG nova.compute.manager [None req-b3f57e08-3a8e-4c04-9171-e3846694852d tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 9dc7c179-4360-425d-915f-c2d7bc591b1e] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 994.128970] env[62066]: DEBUG oslo_vmware.api [None req-6ecf159e-1d26-4f9a-b3e2-f54f67ef8e8e tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': task-1341291, 'name': ReconfigVM_Task, 'duration_secs': 0.631625} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.129287] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-6ecf159e-1d26-4f9a-b3e2-f54f67ef8e8e tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: a4ef97ad-bbb5-47fa-b5c1-2867a0d56435] Reconfigured VM instance instance-00000063 to attach disk [datastore2] a4ef97ad-bbb5-47fa-b5c1-2867a0d56435/a4ef97ad-bbb5-47fa-b5c1-2867a0d56435.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 994.129912] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-fcfbcadf-3838-4524-8cf3-ede6f6675eaa {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.136353] env[62066]: DEBUG oslo_vmware.api [None req-6ecf159e-1d26-4f9a-b3e2-f54f67ef8e8e tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Waiting for the task: (returnval){ [ 994.136353] env[62066]: value = "task-1341293" [ 994.136353] env[62066]: _type = "Task" [ 994.136353] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.145671] env[62066]: DEBUG oslo_vmware.api [None req-6ecf159e-1d26-4f9a-b3e2-f54f67ef8e8e tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': task-1341293, 'name': Rename_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.147342] env[62066]: DEBUG oslo_concurrency.lockutils [req-525dd1bd-5120-421c-8c05-c3d226ab42c3 req-eea1cb57-f610-4a26-9143-676c911adfcc service nova] Releasing lock "refresh_cache-c8482d92-0180-44ee-a4f1-bf84786dad43" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 994.313035] env[62066]: DEBUG nova.compute.manager [None req-2cce9c2e-a55d-474b-be2b-37d32fa31b9b tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: 7178eb83-41e7-4516-809d-a0972d7e6e23] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 994.646441] env[62066]: DEBUG oslo_vmware.api [None req-6ecf159e-1d26-4f9a-b3e2-f54f67ef8e8e tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': task-1341293, 'name': Rename_Task, 'duration_secs': 0.40681} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.646730] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-6ecf159e-1d26-4f9a-b3e2-f54f67ef8e8e tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: a4ef97ad-bbb5-47fa-b5c1-2867a0d56435] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 994.646978] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-19a9206e-496d-4de7-9789-1329a0a71c8a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.652946] env[62066]: DEBUG oslo_vmware.api [None req-6ecf159e-1d26-4f9a-b3e2-f54f67ef8e8e tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Waiting for the task: (returnval){ [ 994.652946] env[62066]: value = "task-1341294" [ 994.652946] env[62066]: _type = "Task" [ 994.652946] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.660385] env[62066]: DEBUG oslo_vmware.api [None req-6ecf159e-1d26-4f9a-b3e2-f54f67ef8e8e tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': task-1341294, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.835608] env[62066]: DEBUG oslo_concurrency.lockutils [None req-2cce9c2e-a55d-474b-be2b-37d32fa31b9b tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 994.835889] env[62066]: DEBUG oslo_concurrency.lockutils [None req-2cce9c2e-a55d-474b-be2b-37d32fa31b9b tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 994.837486] env[62066]: INFO nova.compute.claims [None req-2cce9c2e-a55d-474b-be2b-37d32fa31b9b tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: 7178eb83-41e7-4516-809d-a0972d7e6e23] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 995.049945] env[62066]: DEBUG nova.compute.manager [None req-b3f57e08-3a8e-4c04-9171-e3846694852d tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 9dc7c179-4360-425d-915f-c2d7bc591b1e] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 995.076982] env[62066]: DEBUG nova.virt.hardware [None req-b3f57e08-3a8e-4c04-9171-e3846694852d tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-23T13:40:41Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-23T13:40:23Z,direct_url=,disk_format='vmdk',id=50ff584c-3b50-4395-af07-3e66769bc9f7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='eb52f7069a374c61ae946f052007c6d9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-23T13:40:24Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 995.077312] env[62066]: DEBUG nova.virt.hardware [None req-b3f57e08-3a8e-4c04-9171-e3846694852d tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 995.077533] env[62066]: DEBUG nova.virt.hardware [None req-b3f57e08-3a8e-4c04-9171-e3846694852d tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 995.077760] env[62066]: DEBUG nova.virt.hardware [None req-b3f57e08-3a8e-4c04-9171-e3846694852d tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 995.077952] env[62066]: DEBUG nova.virt.hardware [None req-b3f57e08-3a8e-4c04-9171-e3846694852d tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 995.078170] env[62066]: DEBUG nova.virt.hardware [None req-b3f57e08-3a8e-4c04-9171-e3846694852d tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 995.078428] env[62066]: DEBUG nova.virt.hardware [None req-b3f57e08-3a8e-4c04-9171-e3846694852d tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 995.078610] env[62066]: DEBUG nova.virt.hardware [None req-b3f57e08-3a8e-4c04-9171-e3846694852d tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 995.078840] env[62066]: DEBUG nova.virt.hardware [None req-b3f57e08-3a8e-4c04-9171-e3846694852d tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 995.079054] env[62066]: DEBUG nova.virt.hardware [None req-b3f57e08-3a8e-4c04-9171-e3846694852d tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 995.079296] env[62066]: DEBUG nova.virt.hardware [None req-b3f57e08-3a8e-4c04-9171-e3846694852d tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 995.080292] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59fbc9d1-b5fc-46ab-be57-5fe0d8438a70 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.088537] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db5e4df8-a2ad-42f9-8903-71b7eac60a23 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.163109] env[62066]: DEBUG oslo_vmware.api [None req-6ecf159e-1d26-4f9a-b3e2-f54f67ef8e8e tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': task-1341294, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.395609] env[62066]: DEBUG nova.compute.manager [req-400d5a7f-79b9-4d8e-a30e-aad6a2798224 req-c05dcc56-b2e2-4f73-978a-9f14bcd3b758 service nova] [instance: 9dc7c179-4360-425d-915f-c2d7bc591b1e] Received event network-vif-plugged-8499c7af-d27f-4b10-b63d-a50b2a34d620 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 995.395887] env[62066]: DEBUG oslo_concurrency.lockutils [req-400d5a7f-79b9-4d8e-a30e-aad6a2798224 req-c05dcc56-b2e2-4f73-978a-9f14bcd3b758 service nova] Acquiring lock "9dc7c179-4360-425d-915f-c2d7bc591b1e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 995.396342] env[62066]: DEBUG oslo_concurrency.lockutils [req-400d5a7f-79b9-4d8e-a30e-aad6a2798224 req-c05dcc56-b2e2-4f73-978a-9f14bcd3b758 service nova] Lock "9dc7c179-4360-425d-915f-c2d7bc591b1e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 995.396575] env[62066]: DEBUG oslo_concurrency.lockutils [req-400d5a7f-79b9-4d8e-a30e-aad6a2798224 req-c05dcc56-b2e2-4f73-978a-9f14bcd3b758 service nova] Lock "9dc7c179-4360-425d-915f-c2d7bc591b1e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 995.396762] env[62066]: DEBUG nova.compute.manager [req-400d5a7f-79b9-4d8e-a30e-aad6a2798224 req-c05dcc56-b2e2-4f73-978a-9f14bcd3b758 service nova] [instance: 9dc7c179-4360-425d-915f-c2d7bc591b1e] No waiting events found dispatching network-vif-plugged-8499c7af-d27f-4b10-b63d-a50b2a34d620 {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 995.396934] env[62066]: WARNING nova.compute.manager [req-400d5a7f-79b9-4d8e-a30e-aad6a2798224 req-c05dcc56-b2e2-4f73-978a-9f14bcd3b758 service nova] [instance: 9dc7c179-4360-425d-915f-c2d7bc591b1e] Received unexpected event network-vif-plugged-8499c7af-d27f-4b10-b63d-a50b2a34d620 for instance with vm_state building and task_state spawning. [ 995.485933] env[62066]: DEBUG nova.network.neutron [None req-b3f57e08-3a8e-4c04-9171-e3846694852d tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 9dc7c179-4360-425d-915f-c2d7bc591b1e] Successfully updated port: 8499c7af-d27f-4b10-b63d-a50b2a34d620 {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 995.665124] env[62066]: DEBUG oslo_vmware.api [None req-6ecf159e-1d26-4f9a-b3e2-f54f67ef8e8e tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': task-1341294, 'name': PowerOnVM_Task, 'duration_secs': 0.715237} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.665460] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-6ecf159e-1d26-4f9a-b3e2-f54f67ef8e8e tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: a4ef97ad-bbb5-47fa-b5c1-2867a0d56435] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 995.665715] env[62066]: INFO nova.compute.manager [None req-6ecf159e-1d26-4f9a-b3e2-f54f67ef8e8e tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: a4ef97ad-bbb5-47fa-b5c1-2867a0d56435] Took 8.59 seconds to spawn the instance on the hypervisor. [ 995.665939] env[62066]: DEBUG nova.compute.manager [None req-6ecf159e-1d26-4f9a-b3e2-f54f67ef8e8e tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: a4ef97ad-bbb5-47fa-b5c1-2867a0d56435] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 995.666792] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b99a0ed0-d71f-4b36-8702-c07a05c6776a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.953613] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a12576be-0ad0-4c3c-b564-207f9059dcee {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.962120] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b71d259-44be-4880-b01b-6d6866b2def4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.992395] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b3f57e08-3a8e-4c04-9171-e3846694852d tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Acquiring lock "refresh_cache-9dc7c179-4360-425d-915f-c2d7bc591b1e" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 995.992584] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b3f57e08-3a8e-4c04-9171-e3846694852d tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Acquired lock "refresh_cache-9dc7c179-4360-425d-915f-c2d7bc591b1e" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 995.992744] env[62066]: DEBUG nova.network.neutron [None req-b3f57e08-3a8e-4c04-9171-e3846694852d tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 9dc7c179-4360-425d-915f-c2d7bc591b1e] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 995.994648] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4eae98bf-4a71-4a19-80b3-607ee2eb5efa {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.002385] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8f180f3-daa5-4b5c-bfd9-32666f6f379f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.018616] env[62066]: DEBUG nova.compute.provider_tree [None req-2cce9c2e-a55d-474b-be2b-37d32fa31b9b tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 996.185905] env[62066]: INFO nova.compute.manager [None req-6ecf159e-1d26-4f9a-b3e2-f54f67ef8e8e tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: a4ef97ad-bbb5-47fa-b5c1-2867a0d56435] Took 13.35 seconds to build instance. [ 996.521761] env[62066]: DEBUG nova.scheduler.client.report [None req-2cce9c2e-a55d-474b-be2b-37d32fa31b9b tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 996.553305] env[62066]: DEBUG nova.network.neutron [None req-b3f57e08-3a8e-4c04-9171-e3846694852d tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 9dc7c179-4360-425d-915f-c2d7bc591b1e] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 996.688229] env[62066]: DEBUG oslo_concurrency.lockutils [None req-6ecf159e-1d26-4f9a-b3e2-f54f67ef8e8e tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Lock "a4ef97ad-bbb5-47fa-b5c1-2867a0d56435" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.869s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 996.778603] env[62066]: DEBUG nova.network.neutron [None req-b3f57e08-3a8e-4c04-9171-e3846694852d tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 9dc7c179-4360-425d-915f-c2d7bc591b1e] Updating instance_info_cache with network_info: [{"id": "8499c7af-d27f-4b10-b63d-a50b2a34d620", "address": "fa:16:3e:94:5a:ef", "network": {"id": "1c6c51e4-96f8-4175-b7f5-96b8a42fcb13", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-2006089809-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d609babc987a42e2a8ddb4bfb9c3b103", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a965790c-2d2f-4c2a-9ee7-745f4d53039b", "external-id": "nsx-vlan-transportzone-708", "segmentation_id": 708, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8499c7af-d2", "ovs_interfaceid": "8499c7af-d27f-4b10-b63d-a50b2a34d620", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 996.923739] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0c802fd1-f1cf-4d9e-9a93-9a2ca93e6d84 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Acquiring lock "a4ef97ad-bbb5-47fa-b5c1-2867a0d56435" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 996.924021] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0c802fd1-f1cf-4d9e-9a93-9a2ca93e6d84 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Lock "a4ef97ad-bbb5-47fa-b5c1-2867a0d56435" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 996.924242] env[62066]: DEBUG nova.compute.manager [None req-0c802fd1-f1cf-4d9e-9a93-9a2ca93e6d84 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: a4ef97ad-bbb5-47fa-b5c1-2867a0d56435] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 996.925242] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-148650bc-d34a-40fc-bf23-97a5d3359b88 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.932462] env[62066]: DEBUG nova.compute.manager [None req-0c802fd1-f1cf-4d9e-9a93-9a2ca93e6d84 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: a4ef97ad-bbb5-47fa-b5c1-2867a0d56435] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62066) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 996.933115] env[62066]: DEBUG nova.objects.instance [None req-0c802fd1-f1cf-4d9e-9a93-9a2ca93e6d84 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Lazy-loading 'flavor' on Instance uuid a4ef97ad-bbb5-47fa-b5c1-2867a0d56435 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 997.031875] env[62066]: DEBUG oslo_concurrency.lockutils [None req-2cce9c2e-a55d-474b-be2b-37d32fa31b9b tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.194s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 997.031875] env[62066]: DEBUG nova.compute.manager [None req-2cce9c2e-a55d-474b-be2b-37d32fa31b9b tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: 7178eb83-41e7-4516-809d-a0972d7e6e23] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 997.281288] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b3f57e08-3a8e-4c04-9171-e3846694852d tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Releasing lock "refresh_cache-9dc7c179-4360-425d-915f-c2d7bc591b1e" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 997.281697] env[62066]: DEBUG nova.compute.manager [None req-b3f57e08-3a8e-4c04-9171-e3846694852d tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 9dc7c179-4360-425d-915f-c2d7bc591b1e] Instance network_info: |[{"id": "8499c7af-d27f-4b10-b63d-a50b2a34d620", "address": "fa:16:3e:94:5a:ef", "network": {"id": "1c6c51e4-96f8-4175-b7f5-96b8a42fcb13", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-2006089809-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d609babc987a42e2a8ddb4bfb9c3b103", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a965790c-2d2f-4c2a-9ee7-745f4d53039b", "external-id": "nsx-vlan-transportzone-708", "segmentation_id": 708, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8499c7af-d2", "ovs_interfaceid": "8499c7af-d27f-4b10-b63d-a50b2a34d620", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 997.282135] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-b3f57e08-3a8e-4c04-9171-e3846694852d tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 9dc7c179-4360-425d-915f-c2d7bc591b1e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:94:5a:ef', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a965790c-2d2f-4c2a-9ee7-745f4d53039b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8499c7af-d27f-4b10-b63d-a50b2a34d620', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 997.291848] env[62066]: DEBUG oslo.service.loopingcall [None req-b3f57e08-3a8e-4c04-9171-e3846694852d tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 997.292179] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9dc7c179-4360-425d-915f-c2d7bc591b1e] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 997.292432] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-bb21ab2c-6d37-460c-b0f9-4cede37c1cde {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.315579] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 997.315579] env[62066]: value = "task-1341295" [ 997.315579] env[62066]: _type = "Task" [ 997.315579] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.324566] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1341295, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.436517] env[62066]: DEBUG nova.compute.manager [req-cc50cdee-5e4e-4dc5-9e07-0e0da0f8ed59 req-c62aa5d3-4d41-4b36-8554-34541be70499 service nova] [instance: 9dc7c179-4360-425d-915f-c2d7bc591b1e] Received event network-changed-8499c7af-d27f-4b10-b63d-a50b2a34d620 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 997.436517] env[62066]: DEBUG nova.compute.manager [req-cc50cdee-5e4e-4dc5-9e07-0e0da0f8ed59 req-c62aa5d3-4d41-4b36-8554-34541be70499 service nova] [instance: 9dc7c179-4360-425d-915f-c2d7bc591b1e] Refreshing instance network info cache due to event network-changed-8499c7af-d27f-4b10-b63d-a50b2a34d620. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 997.436517] env[62066]: DEBUG oslo_concurrency.lockutils [req-cc50cdee-5e4e-4dc5-9e07-0e0da0f8ed59 req-c62aa5d3-4d41-4b36-8554-34541be70499 service nova] Acquiring lock "refresh_cache-9dc7c179-4360-425d-915f-c2d7bc591b1e" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 997.436517] env[62066]: DEBUG oslo_concurrency.lockutils [req-cc50cdee-5e4e-4dc5-9e07-0e0da0f8ed59 req-c62aa5d3-4d41-4b36-8554-34541be70499 service nova] Acquired lock "refresh_cache-9dc7c179-4360-425d-915f-c2d7bc591b1e" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 997.436517] env[62066]: DEBUG nova.network.neutron [req-cc50cdee-5e4e-4dc5-9e07-0e0da0f8ed59 req-c62aa5d3-4d41-4b36-8554-34541be70499 service nova] [instance: 9dc7c179-4360-425d-915f-c2d7bc591b1e] Refreshing network info cache for port 8499c7af-d27f-4b10-b63d-a50b2a34d620 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 997.440472] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-0c802fd1-f1cf-4d9e-9a93-9a2ca93e6d84 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: a4ef97ad-bbb5-47fa-b5c1-2867a0d56435] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 997.441940] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-97743c97-9d91-45b2-b77c-c39e3555ac1e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.450087] env[62066]: DEBUG oslo_vmware.api [None req-0c802fd1-f1cf-4d9e-9a93-9a2ca93e6d84 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Waiting for the task: (returnval){ [ 997.450087] env[62066]: value = "task-1341296" [ 997.450087] env[62066]: _type = "Task" [ 997.450087] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.462341] env[62066]: DEBUG oslo_vmware.api [None req-0c802fd1-f1cf-4d9e-9a93-9a2ca93e6d84 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': task-1341296, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.538244] env[62066]: DEBUG nova.compute.utils [None req-2cce9c2e-a55d-474b-be2b-37d32fa31b9b tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 997.539763] env[62066]: DEBUG nova.compute.manager [None req-2cce9c2e-a55d-474b-be2b-37d32fa31b9b tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: 7178eb83-41e7-4516-809d-a0972d7e6e23] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 997.539763] env[62066]: DEBUG nova.network.neutron [None req-2cce9c2e-a55d-474b-be2b-37d32fa31b9b tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: 7178eb83-41e7-4516-809d-a0972d7e6e23] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 997.594792] env[62066]: DEBUG nova.policy [None req-2cce9c2e-a55d-474b-be2b-37d32fa31b9b tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '747026d272d4467bbf14a0c30589c3b7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0d4e33a99c7741fb8cdd97f4ec5dbbd8', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 997.826283] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1341295, 'name': CreateVM_Task, 'duration_secs': 0.431905} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.826510] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9dc7c179-4360-425d-915f-c2d7bc591b1e] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 997.827426] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b3f57e08-3a8e-4c04-9171-e3846694852d tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 997.827553] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b3f57e08-3a8e-4c04-9171-e3846694852d tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Acquired lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 997.828316] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b3f57e08-3a8e-4c04-9171-e3846694852d tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 997.828316] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-71ff3f3e-c546-4050-8ac7-875b396c27df {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.834027] env[62066]: DEBUG oslo_vmware.api [None req-b3f57e08-3a8e-4c04-9171-e3846694852d tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Waiting for the task: (returnval){ [ 997.834027] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]52711361-8757-2d0e-f8b3-5c2b7fd23c67" [ 997.834027] env[62066]: _type = "Task" [ 997.834027] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.843083] env[62066]: DEBUG oslo_vmware.api [None req-b3f57e08-3a8e-4c04-9171-e3846694852d tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52711361-8757-2d0e-f8b3-5c2b7fd23c67, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.961195] env[62066]: DEBUG oslo_vmware.api [None req-0c802fd1-f1cf-4d9e-9a93-9a2ca93e6d84 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': task-1341296, 'name': PowerOffVM_Task, 'duration_secs': 0.288464} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.961518] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-0c802fd1-f1cf-4d9e-9a93-9a2ca93e6d84 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: a4ef97ad-bbb5-47fa-b5c1-2867a0d56435] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 997.961710] env[62066]: DEBUG nova.compute.manager [None req-0c802fd1-f1cf-4d9e-9a93-9a2ca93e6d84 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: a4ef97ad-bbb5-47fa-b5c1-2867a0d56435] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 997.962599] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89f095a5-3dcd-4845-a51a-e599ee477d5b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.044226] env[62066]: DEBUG nova.compute.manager [None req-2cce9c2e-a55d-474b-be2b-37d32fa31b9b tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: 7178eb83-41e7-4516-809d-a0972d7e6e23] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 998.055053] env[62066]: DEBUG nova.network.neutron [None req-2cce9c2e-a55d-474b-be2b-37d32fa31b9b tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: 7178eb83-41e7-4516-809d-a0972d7e6e23] Successfully created port: 63f16c96-7de0-40e7-9c0c-782122865437 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 998.172017] env[62066]: DEBUG nova.network.neutron [req-cc50cdee-5e4e-4dc5-9e07-0e0da0f8ed59 req-c62aa5d3-4d41-4b36-8554-34541be70499 service nova] [instance: 9dc7c179-4360-425d-915f-c2d7bc591b1e] Updated VIF entry in instance network info cache for port 8499c7af-d27f-4b10-b63d-a50b2a34d620. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 998.172456] env[62066]: DEBUG nova.network.neutron [req-cc50cdee-5e4e-4dc5-9e07-0e0da0f8ed59 req-c62aa5d3-4d41-4b36-8554-34541be70499 service nova] [instance: 9dc7c179-4360-425d-915f-c2d7bc591b1e] Updating instance_info_cache with network_info: [{"id": "8499c7af-d27f-4b10-b63d-a50b2a34d620", "address": "fa:16:3e:94:5a:ef", "network": {"id": "1c6c51e4-96f8-4175-b7f5-96b8a42fcb13", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-2006089809-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d609babc987a42e2a8ddb4bfb9c3b103", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a965790c-2d2f-4c2a-9ee7-745f4d53039b", "external-id": "nsx-vlan-transportzone-708", "segmentation_id": 708, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8499c7af-d2", "ovs_interfaceid": "8499c7af-d27f-4b10-b63d-a50b2a34d620", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 998.349900] env[62066]: DEBUG oslo_vmware.api [None req-b3f57e08-3a8e-4c04-9171-e3846694852d tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52711361-8757-2d0e-f8b3-5c2b7fd23c67, 'name': SearchDatastore_Task, 'duration_secs': 0.015469} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.350298] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b3f57e08-3a8e-4c04-9171-e3846694852d tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Releasing lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 998.350551] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-b3f57e08-3a8e-4c04-9171-e3846694852d tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 9dc7c179-4360-425d-915f-c2d7bc591b1e] Processing image 50ff584c-3b50-4395-af07-3e66769bc9f7 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 998.350860] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b3f57e08-3a8e-4c04-9171-e3846694852d tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 998.351053] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b3f57e08-3a8e-4c04-9171-e3846694852d tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Acquired lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 998.351263] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-b3f57e08-3a8e-4c04-9171-e3846694852d tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 998.351557] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0a07391f-0c7e-43c7-94ad-291b9a5cfc32 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.361111] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-b3f57e08-3a8e-4c04-9171-e3846694852d tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 998.361319] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-b3f57e08-3a8e-4c04-9171-e3846694852d tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 998.362120] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bc1453fc-c9d8-4b10-b29a-86d5ac57ae92 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.368305] env[62066]: DEBUG oslo_vmware.api [None req-b3f57e08-3a8e-4c04-9171-e3846694852d tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Waiting for the task: (returnval){ [ 998.368305] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]52cd401d-5a97-2029-5a43-e6ce9e75a428" [ 998.368305] env[62066]: _type = "Task" [ 998.368305] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.376539] env[62066]: DEBUG oslo_vmware.api [None req-b3f57e08-3a8e-4c04-9171-e3846694852d tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52cd401d-5a97-2029-5a43-e6ce9e75a428, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.478045] env[62066]: DEBUG oslo_concurrency.lockutils [None req-0c802fd1-f1cf-4d9e-9a93-9a2ca93e6d84 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Lock "a4ef97ad-bbb5-47fa-b5c1-2867a0d56435" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.554s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 998.675396] env[62066]: DEBUG oslo_concurrency.lockutils [req-cc50cdee-5e4e-4dc5-9e07-0e0da0f8ed59 req-c62aa5d3-4d41-4b36-8554-34541be70499 service nova] Releasing lock "refresh_cache-9dc7c179-4360-425d-915f-c2d7bc591b1e" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 998.881454] env[62066]: DEBUG oslo_vmware.api [None req-b3f57e08-3a8e-4c04-9171-e3846694852d tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52cd401d-5a97-2029-5a43-e6ce9e75a428, 'name': SearchDatastore_Task, 'duration_secs': 0.024123} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.882485] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6b411dea-ae2b-486e-94e8-91f10220c867 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.888987] env[62066]: DEBUG oslo_vmware.api [None req-b3f57e08-3a8e-4c04-9171-e3846694852d tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Waiting for the task: (returnval){ [ 998.888987] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]5282e198-b34a-70e0-67cc-59171f970692" [ 998.888987] env[62066]: _type = "Task" [ 998.888987] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.899757] env[62066]: DEBUG oslo_vmware.api [None req-b3f57e08-3a8e-4c04-9171-e3846694852d tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]5282e198-b34a-70e0-67cc-59171f970692, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.055983] env[62066]: DEBUG nova.compute.manager [None req-2cce9c2e-a55d-474b-be2b-37d32fa31b9b tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: 7178eb83-41e7-4516-809d-a0972d7e6e23] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 999.083255] env[62066]: DEBUG nova.virt.hardware [None req-2cce9c2e-a55d-474b-be2b-37d32fa31b9b tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-23T13:40:41Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-23T13:40:23Z,direct_url=,disk_format='vmdk',id=50ff584c-3b50-4395-af07-3e66769bc9f7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='eb52f7069a374c61ae946f052007c6d9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-23T13:40:24Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 999.083536] env[62066]: DEBUG nova.virt.hardware [None req-2cce9c2e-a55d-474b-be2b-37d32fa31b9b tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 999.083715] env[62066]: DEBUG nova.virt.hardware [None req-2cce9c2e-a55d-474b-be2b-37d32fa31b9b tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 999.083912] env[62066]: DEBUG nova.virt.hardware [None req-2cce9c2e-a55d-474b-be2b-37d32fa31b9b tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 999.084135] env[62066]: DEBUG nova.virt.hardware [None req-2cce9c2e-a55d-474b-be2b-37d32fa31b9b tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 999.084330] env[62066]: DEBUG nova.virt.hardware [None req-2cce9c2e-a55d-474b-be2b-37d32fa31b9b tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 999.084579] env[62066]: DEBUG nova.virt.hardware [None req-2cce9c2e-a55d-474b-be2b-37d32fa31b9b tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 999.084864] env[62066]: DEBUG nova.virt.hardware [None req-2cce9c2e-a55d-474b-be2b-37d32fa31b9b tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 999.085093] env[62066]: DEBUG nova.virt.hardware [None req-2cce9c2e-a55d-474b-be2b-37d32fa31b9b tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 999.085293] env[62066]: DEBUG nova.virt.hardware [None req-2cce9c2e-a55d-474b-be2b-37d32fa31b9b tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 999.085536] env[62066]: DEBUG nova.virt.hardware [None req-2cce9c2e-a55d-474b-be2b-37d32fa31b9b tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 999.086483] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a19c4c9-d0a3-4978-9fc1-1781f20ad78f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.095642] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72833de7-8396-43b7-8b00-443548070f18 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.403349] env[62066]: DEBUG oslo_vmware.api [None req-b3f57e08-3a8e-4c04-9171-e3846694852d tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]5282e198-b34a-70e0-67cc-59171f970692, 'name': SearchDatastore_Task, 'duration_secs': 0.009858} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.403816] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b3f57e08-3a8e-4c04-9171-e3846694852d tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Releasing lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 999.404255] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-b3f57e08-3a8e-4c04-9171-e3846694852d tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk to [datastore2] 9dc7c179-4360-425d-915f-c2d7bc591b1e/9dc7c179-4360-425d-915f-c2d7bc591b1e.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 999.404567] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-538d5ae6-6d27-41e2-83fb-25a8a0338748 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.413225] env[62066]: DEBUG oslo_vmware.api [None req-b3f57e08-3a8e-4c04-9171-e3846694852d tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Waiting for the task: (returnval){ [ 999.413225] env[62066]: value = "task-1341297" [ 999.413225] env[62066]: _type = "Task" [ 999.413225] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.422070] env[62066]: DEBUG oslo_vmware.api [None req-b3f57e08-3a8e-4c04-9171-e3846694852d tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1341297, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.605875] env[62066]: DEBUG nova.compute.manager [req-c614e4df-783a-418a-9a4e-e6039c715d54 req-bed4eacd-ea0c-4d74-99ff-92d9cdc2175e service nova] [instance: 7178eb83-41e7-4516-809d-a0972d7e6e23] Received event network-vif-plugged-63f16c96-7de0-40e7-9c0c-782122865437 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 999.606129] env[62066]: DEBUG oslo_concurrency.lockutils [req-c614e4df-783a-418a-9a4e-e6039c715d54 req-bed4eacd-ea0c-4d74-99ff-92d9cdc2175e service nova] Acquiring lock "7178eb83-41e7-4516-809d-a0972d7e6e23-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 999.606355] env[62066]: DEBUG oslo_concurrency.lockutils [req-c614e4df-783a-418a-9a4e-e6039c715d54 req-bed4eacd-ea0c-4d74-99ff-92d9cdc2175e service nova] Lock "7178eb83-41e7-4516-809d-a0972d7e6e23-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 999.606532] env[62066]: DEBUG oslo_concurrency.lockutils [req-c614e4df-783a-418a-9a4e-e6039c715d54 req-bed4eacd-ea0c-4d74-99ff-92d9cdc2175e service nova] Lock "7178eb83-41e7-4516-809d-a0972d7e6e23-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 999.606713] env[62066]: DEBUG nova.compute.manager [req-c614e4df-783a-418a-9a4e-e6039c715d54 req-bed4eacd-ea0c-4d74-99ff-92d9cdc2175e service nova] [instance: 7178eb83-41e7-4516-809d-a0972d7e6e23] No waiting events found dispatching network-vif-plugged-63f16c96-7de0-40e7-9c0c-782122865437 {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 999.606884] env[62066]: WARNING nova.compute.manager [req-c614e4df-783a-418a-9a4e-e6039c715d54 req-bed4eacd-ea0c-4d74-99ff-92d9cdc2175e service nova] [instance: 7178eb83-41e7-4516-809d-a0972d7e6e23] Received unexpected event network-vif-plugged-63f16c96-7de0-40e7-9c0c-782122865437 for instance with vm_state building and task_state spawning. [ 999.610200] env[62066]: DEBUG oslo_vmware.rw_handles [None req-27337d51-e9e8-4b7e-b8c5-38dbed79187f tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/525c4626-1f31-8d3e-bb46-aa44cac71c21/disk-0.vmdk. {{(pid=62066) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 999.611697] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8600e99-5000-4f29-b7b0-a41c1089e158 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.619159] env[62066]: DEBUG oslo_vmware.rw_handles [None req-27337d51-e9e8-4b7e-b8c5-38dbed79187f tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/525c4626-1f31-8d3e-bb46-aa44cac71c21/disk-0.vmdk is in state: ready. {{(pid=62066) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 999.619355] env[62066]: ERROR oslo_vmware.rw_handles [None req-27337d51-e9e8-4b7e-b8c5-38dbed79187f tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/525c4626-1f31-8d3e-bb46-aa44cac71c21/disk-0.vmdk due to incomplete transfer. [ 999.619608] env[62066]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-4b35e679-79d5-4434-a003-2eaca29a3047 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.628191] env[62066]: DEBUG oslo_vmware.rw_handles [None req-27337d51-e9e8-4b7e-b8c5-38dbed79187f tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/525c4626-1f31-8d3e-bb46-aa44cac71c21/disk-0.vmdk. {{(pid=62066) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 999.628529] env[62066]: DEBUG nova.virt.vmwareapi.images [None req-27337d51-e9e8-4b7e-b8c5-38dbed79187f tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: a2c5277e-af49-4b31-8480-a74d354c7383] Uploaded image e4ed31bf-18c9-4cc2-9bff-65cfb25f04fc to the Glance image server {{(pid=62066) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 999.631812] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-27337d51-e9e8-4b7e-b8c5-38dbed79187f tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: a2c5277e-af49-4b31-8480-a74d354c7383] Destroying the VM {{(pid=62066) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 999.631812] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-d04359c3-b694-43af-a385-67dc9e0a08a9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.640735] env[62066]: DEBUG oslo_vmware.api [None req-27337d51-e9e8-4b7e-b8c5-38dbed79187f tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Waiting for the task: (returnval){ [ 999.640735] env[62066]: value = "task-1341298" [ 999.640735] env[62066]: _type = "Task" [ 999.640735] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.648769] env[62066]: DEBUG oslo_vmware.api [None req-27337d51-e9e8-4b7e-b8c5-38dbed79187f tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341298, 'name': Destroy_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.730575] env[62066]: DEBUG nova.network.neutron [None req-2cce9c2e-a55d-474b-be2b-37d32fa31b9b tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: 7178eb83-41e7-4516-809d-a0972d7e6e23] Successfully updated port: 63f16c96-7de0-40e7-9c0c-782122865437 {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 999.763297] env[62066]: DEBUG oslo_concurrency.lockutils [None req-711de73e-4168-430a-b07e-e18f814cac80 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Acquiring lock "a4ef97ad-bbb5-47fa-b5c1-2867a0d56435" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 999.763464] env[62066]: DEBUG oslo_concurrency.lockutils [None req-711de73e-4168-430a-b07e-e18f814cac80 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Lock "a4ef97ad-bbb5-47fa-b5c1-2867a0d56435" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 999.763706] env[62066]: DEBUG oslo_concurrency.lockutils [None req-711de73e-4168-430a-b07e-e18f814cac80 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Acquiring lock "a4ef97ad-bbb5-47fa-b5c1-2867a0d56435-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 999.763955] env[62066]: DEBUG oslo_concurrency.lockutils [None req-711de73e-4168-430a-b07e-e18f814cac80 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Lock "a4ef97ad-bbb5-47fa-b5c1-2867a0d56435-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 999.764200] env[62066]: DEBUG oslo_concurrency.lockutils [None req-711de73e-4168-430a-b07e-e18f814cac80 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Lock "a4ef97ad-bbb5-47fa-b5c1-2867a0d56435-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 999.766569] env[62066]: INFO nova.compute.manager [None req-711de73e-4168-430a-b07e-e18f814cac80 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: a4ef97ad-bbb5-47fa-b5c1-2867a0d56435] Terminating instance [ 999.768930] env[62066]: DEBUG nova.compute.manager [None req-711de73e-4168-430a-b07e-e18f814cac80 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: a4ef97ad-bbb5-47fa-b5c1-2867a0d56435] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 999.769194] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-711de73e-4168-430a-b07e-e18f814cac80 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: a4ef97ad-bbb5-47fa-b5c1-2867a0d56435] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 999.770140] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-992844bb-de27-4201-9841-305b5347fecb {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.780263] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-711de73e-4168-430a-b07e-e18f814cac80 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: a4ef97ad-bbb5-47fa-b5c1-2867a0d56435] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 999.780674] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b655b030-5ba3-4896-b711-9a0826d38613 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.852132] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-711de73e-4168-430a-b07e-e18f814cac80 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: a4ef97ad-bbb5-47fa-b5c1-2867a0d56435] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 999.852397] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-711de73e-4168-430a-b07e-e18f814cac80 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: a4ef97ad-bbb5-47fa-b5c1-2867a0d56435] Deleting contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 999.852739] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-711de73e-4168-430a-b07e-e18f814cac80 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Deleting the datastore file [datastore2] a4ef97ad-bbb5-47fa-b5c1-2867a0d56435 {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 999.853186] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4e1fec2e-a305-4c9c-a6d3-0caf44388b28 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.860963] env[62066]: DEBUG oslo_vmware.api [None req-711de73e-4168-430a-b07e-e18f814cac80 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Waiting for the task: (returnval){ [ 999.860963] env[62066]: value = "task-1341300" [ 999.860963] env[62066]: _type = "Task" [ 999.860963] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.869808] env[62066]: DEBUG oslo_vmware.api [None req-711de73e-4168-430a-b07e-e18f814cac80 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': task-1341300, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.923679] env[62066]: DEBUG oslo_vmware.api [None req-b3f57e08-3a8e-4c04-9171-e3846694852d tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1341297, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.484843} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.924025] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-b3f57e08-3a8e-4c04-9171-e3846694852d tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk to [datastore2] 9dc7c179-4360-425d-915f-c2d7bc591b1e/9dc7c179-4360-425d-915f-c2d7bc591b1e.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 999.924261] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-b3f57e08-3a8e-4c04-9171-e3846694852d tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 9dc7c179-4360-425d-915f-c2d7bc591b1e] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 999.924526] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c247a549-e19b-49a5-ab21-355034f65eca {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.929906] env[62066]: DEBUG oslo_vmware.api [None req-b3f57e08-3a8e-4c04-9171-e3846694852d tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Waiting for the task: (returnval){ [ 999.929906] env[62066]: value = "task-1341301" [ 999.929906] env[62066]: _type = "Task" [ 999.929906] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.936907] env[62066]: DEBUG oslo_vmware.api [None req-b3f57e08-3a8e-4c04-9171-e3846694852d tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1341301, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.150555] env[62066]: DEBUG oslo_vmware.api [None req-27337d51-e9e8-4b7e-b8c5-38dbed79187f tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341298, 'name': Destroy_Task, 'duration_secs': 0.313964} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.150910] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-27337d51-e9e8-4b7e-b8c5-38dbed79187f tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: a2c5277e-af49-4b31-8480-a74d354c7383] Destroyed the VM [ 1000.151268] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-27337d51-e9e8-4b7e-b8c5-38dbed79187f tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: a2c5277e-af49-4b31-8480-a74d354c7383] Deleting Snapshot of the VM instance {{(pid=62066) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1000.151435] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-e1da2634-b07d-480b-b5ed-f0091f765704 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.157500] env[62066]: DEBUG oslo_vmware.api [None req-27337d51-e9e8-4b7e-b8c5-38dbed79187f tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Waiting for the task: (returnval){ [ 1000.157500] env[62066]: value = "task-1341302" [ 1000.157500] env[62066]: _type = "Task" [ 1000.157500] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.164853] env[62066]: DEBUG oslo_vmware.api [None req-27337d51-e9e8-4b7e-b8c5-38dbed79187f tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341302, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.230999] env[62066]: DEBUG oslo_concurrency.lockutils [None req-2cce9c2e-a55d-474b-be2b-37d32fa31b9b tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Acquiring lock "refresh_cache-7178eb83-41e7-4516-809d-a0972d7e6e23" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1000.231217] env[62066]: DEBUG oslo_concurrency.lockutils [None req-2cce9c2e-a55d-474b-be2b-37d32fa31b9b tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Acquired lock "refresh_cache-7178eb83-41e7-4516-809d-a0972d7e6e23" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1000.231379] env[62066]: DEBUG nova.network.neutron [None req-2cce9c2e-a55d-474b-be2b-37d32fa31b9b tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: 7178eb83-41e7-4516-809d-a0972d7e6e23] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1000.372730] env[62066]: DEBUG oslo_vmware.api [None req-711de73e-4168-430a-b07e-e18f814cac80 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': task-1341300, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.138502} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.373135] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-711de73e-4168-430a-b07e-e18f814cac80 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1000.373399] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-711de73e-4168-430a-b07e-e18f814cac80 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: a4ef97ad-bbb5-47fa-b5c1-2867a0d56435] Deleted contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1000.373619] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-711de73e-4168-430a-b07e-e18f814cac80 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: a4ef97ad-bbb5-47fa-b5c1-2867a0d56435] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1000.373806] env[62066]: INFO nova.compute.manager [None req-711de73e-4168-430a-b07e-e18f814cac80 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: a4ef97ad-bbb5-47fa-b5c1-2867a0d56435] Took 0.60 seconds to destroy the instance on the hypervisor. [ 1000.374107] env[62066]: DEBUG oslo.service.loopingcall [None req-711de73e-4168-430a-b07e-e18f814cac80 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1000.374324] env[62066]: DEBUG nova.compute.manager [-] [instance: a4ef97ad-bbb5-47fa-b5c1-2867a0d56435] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1000.374418] env[62066]: DEBUG nova.network.neutron [-] [instance: a4ef97ad-bbb5-47fa-b5c1-2867a0d56435] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1000.439591] env[62066]: DEBUG oslo_vmware.api [None req-b3f57e08-3a8e-4c04-9171-e3846694852d tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1341301, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.07021} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.439923] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-b3f57e08-3a8e-4c04-9171-e3846694852d tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 9dc7c179-4360-425d-915f-c2d7bc591b1e] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1000.440711] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4550fd3-9683-4035-9d8b-c73bf4568779 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.461858] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-b3f57e08-3a8e-4c04-9171-e3846694852d tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 9dc7c179-4360-425d-915f-c2d7bc591b1e] Reconfiguring VM instance instance-00000064 to attach disk [datastore2] 9dc7c179-4360-425d-915f-c2d7bc591b1e/9dc7c179-4360-425d-915f-c2d7bc591b1e.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1000.462145] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0500931f-b01a-403e-af9e-3a5e52eeef46 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.480270] env[62066]: DEBUG oslo_vmware.api [None req-b3f57e08-3a8e-4c04-9171-e3846694852d tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Waiting for the task: (returnval){ [ 1000.480270] env[62066]: value = "task-1341303" [ 1000.480270] env[62066]: _type = "Task" [ 1000.480270] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.487840] env[62066]: DEBUG oslo_vmware.api [None req-b3f57e08-3a8e-4c04-9171-e3846694852d tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1341303, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.670099] env[62066]: DEBUG oslo_vmware.api [None req-27337d51-e9e8-4b7e-b8c5-38dbed79187f tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341302, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.765985] env[62066]: DEBUG nova.network.neutron [None req-2cce9c2e-a55d-474b-be2b-37d32fa31b9b tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: 7178eb83-41e7-4516-809d-a0972d7e6e23] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1000.932164] env[62066]: DEBUG nova.network.neutron [None req-2cce9c2e-a55d-474b-be2b-37d32fa31b9b tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: 7178eb83-41e7-4516-809d-a0972d7e6e23] Updating instance_info_cache with network_info: [{"id": "63f16c96-7de0-40e7-9c0c-782122865437", "address": "fa:16:3e:c6:a7:3e", "network": {"id": "449dfe4e-bb49-43b6-9ba7-b57af74ebfb3", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-282029676-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0d4e33a99c7741fb8cdd97f4ec5dbbd8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db68bd64-5b56-49af-a075-13dcf85cb2e0", "external-id": "nsx-vlan-transportzone-590", "segmentation_id": 590, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap63f16c96-7d", "ovs_interfaceid": "63f16c96-7de0-40e7-9c0c-782122865437", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1000.991078] env[62066]: DEBUG oslo_vmware.api [None req-b3f57e08-3a8e-4c04-9171-e3846694852d tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1341303, 'name': ReconfigVM_Task, 'duration_secs': 0.50479} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.991449] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-b3f57e08-3a8e-4c04-9171-e3846694852d tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 9dc7c179-4360-425d-915f-c2d7bc591b1e] Reconfigured VM instance instance-00000064 to attach disk [datastore2] 9dc7c179-4360-425d-915f-c2d7bc591b1e/9dc7c179-4360-425d-915f-c2d7bc591b1e.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1000.992155] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-abdb4129-bbec-4bff-9f13-a494c8298dac {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.997807] env[62066]: DEBUG oslo_vmware.api [None req-b3f57e08-3a8e-4c04-9171-e3846694852d tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Waiting for the task: (returnval){ [ 1000.997807] env[62066]: value = "task-1341304" [ 1000.997807] env[62066]: _type = "Task" [ 1000.997807] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.005600] env[62066]: DEBUG oslo_vmware.api [None req-b3f57e08-3a8e-4c04-9171-e3846694852d tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1341304, 'name': Rename_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.167267] env[62066]: DEBUG oslo_vmware.api [None req-27337d51-e9e8-4b7e-b8c5-38dbed79187f tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341302, 'name': RemoveSnapshot_Task, 'duration_secs': 0.869875} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.167490] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-27337d51-e9e8-4b7e-b8c5-38dbed79187f tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: a2c5277e-af49-4b31-8480-a74d354c7383] Deleted Snapshot of the VM instance {{(pid=62066) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1001.167782] env[62066]: DEBUG nova.compute.manager [None req-27337d51-e9e8-4b7e-b8c5-38dbed79187f tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: a2c5277e-af49-4b31-8480-a74d354c7383] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1001.168581] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f187f1e9-b0a2-4163-bcb8-209b07f697b7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.436519] env[62066]: DEBUG oslo_concurrency.lockutils [None req-2cce9c2e-a55d-474b-be2b-37d32fa31b9b tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Releasing lock "refresh_cache-7178eb83-41e7-4516-809d-a0972d7e6e23" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1001.436827] env[62066]: DEBUG nova.compute.manager [None req-2cce9c2e-a55d-474b-be2b-37d32fa31b9b tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: 7178eb83-41e7-4516-809d-a0972d7e6e23] Instance network_info: |[{"id": "63f16c96-7de0-40e7-9c0c-782122865437", "address": "fa:16:3e:c6:a7:3e", "network": {"id": "449dfe4e-bb49-43b6-9ba7-b57af74ebfb3", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-282029676-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0d4e33a99c7741fb8cdd97f4ec5dbbd8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db68bd64-5b56-49af-a075-13dcf85cb2e0", "external-id": "nsx-vlan-transportzone-590", "segmentation_id": 590, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap63f16c96-7d", "ovs_interfaceid": "63f16c96-7de0-40e7-9c0c-782122865437", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1001.437326] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-2cce9c2e-a55d-474b-be2b-37d32fa31b9b tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: 7178eb83-41e7-4516-809d-a0972d7e6e23] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c6:a7:3e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'db68bd64-5b56-49af-a075-13dcf85cb2e0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '63f16c96-7de0-40e7-9c0c-782122865437', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1001.445111] env[62066]: DEBUG oslo.service.loopingcall [None req-2cce9c2e-a55d-474b-be2b-37d32fa31b9b tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1001.445345] env[62066]: DEBUG nova.network.neutron [-] [instance: a4ef97ad-bbb5-47fa-b5c1-2867a0d56435] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1001.446550] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7178eb83-41e7-4516-809d-a0972d7e6e23] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1001.447114] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-750bc23d-3e01-49fb-87da-b96d167e1241 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.467577] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1001.467577] env[62066]: value = "task-1341305" [ 1001.467577] env[62066]: _type = "Task" [ 1001.467577] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.475394] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1341305, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.507026] env[62066]: DEBUG oslo_vmware.api [None req-b3f57e08-3a8e-4c04-9171-e3846694852d tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1341304, 'name': Rename_Task, 'duration_secs': 0.136668} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.508949] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-b3f57e08-3a8e-4c04-9171-e3846694852d tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 9dc7c179-4360-425d-915f-c2d7bc591b1e] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1001.509278] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b4f68706-625b-409a-b920-2b1da78c647d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.514444] env[62066]: DEBUG oslo_vmware.api [None req-b3f57e08-3a8e-4c04-9171-e3846694852d tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Waiting for the task: (returnval){ [ 1001.514444] env[62066]: value = "task-1341306" [ 1001.514444] env[62066]: _type = "Task" [ 1001.514444] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.526351] env[62066]: DEBUG oslo_vmware.api [None req-b3f57e08-3a8e-4c04-9171-e3846694852d tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1341306, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.634327] env[62066]: DEBUG nova.compute.manager [req-2a6009eb-8b0a-4004-8a50-26722eabaadf req-d068368d-a9ae-4ac5-873d-1585166024b9 service nova] [instance: 7178eb83-41e7-4516-809d-a0972d7e6e23] Received event network-changed-63f16c96-7de0-40e7-9c0c-782122865437 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1001.634534] env[62066]: DEBUG nova.compute.manager [req-2a6009eb-8b0a-4004-8a50-26722eabaadf req-d068368d-a9ae-4ac5-873d-1585166024b9 service nova] [instance: 7178eb83-41e7-4516-809d-a0972d7e6e23] Refreshing instance network info cache due to event network-changed-63f16c96-7de0-40e7-9c0c-782122865437. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1001.634762] env[62066]: DEBUG oslo_concurrency.lockutils [req-2a6009eb-8b0a-4004-8a50-26722eabaadf req-d068368d-a9ae-4ac5-873d-1585166024b9 service nova] Acquiring lock "refresh_cache-7178eb83-41e7-4516-809d-a0972d7e6e23" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1001.634908] env[62066]: DEBUG oslo_concurrency.lockutils [req-2a6009eb-8b0a-4004-8a50-26722eabaadf req-d068368d-a9ae-4ac5-873d-1585166024b9 service nova] Acquired lock "refresh_cache-7178eb83-41e7-4516-809d-a0972d7e6e23" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1001.635112] env[62066]: DEBUG nova.network.neutron [req-2a6009eb-8b0a-4004-8a50-26722eabaadf req-d068368d-a9ae-4ac5-873d-1585166024b9 service nova] [instance: 7178eb83-41e7-4516-809d-a0972d7e6e23] Refreshing network info cache for port 63f16c96-7de0-40e7-9c0c-782122865437 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1001.679673] env[62066]: INFO nova.compute.manager [None req-27337d51-e9e8-4b7e-b8c5-38dbed79187f tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: a2c5277e-af49-4b31-8480-a74d354c7383] Shelve offloading [ 1001.681499] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-27337d51-e9e8-4b7e-b8c5-38dbed79187f tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: a2c5277e-af49-4b31-8480-a74d354c7383] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1001.681775] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d8b8ce2c-ff87-4b6d-a3b0-29d69d5373fb {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.689346] env[62066]: DEBUG oslo_vmware.api [None req-27337d51-e9e8-4b7e-b8c5-38dbed79187f tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Waiting for the task: (returnval){ [ 1001.689346] env[62066]: value = "task-1341307" [ 1001.689346] env[62066]: _type = "Task" [ 1001.689346] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.696987] env[62066]: DEBUG oslo_vmware.api [None req-27337d51-e9e8-4b7e-b8c5-38dbed79187f tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341307, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.948633] env[62066]: INFO nova.compute.manager [-] [instance: a4ef97ad-bbb5-47fa-b5c1-2867a0d56435] Took 1.57 seconds to deallocate network for instance. [ 1001.977370] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1341305, 'name': CreateVM_Task, 'duration_secs': 0.470655} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.977880] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7178eb83-41e7-4516-809d-a0972d7e6e23] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1001.978341] env[62066]: DEBUG oslo_concurrency.lockutils [None req-2cce9c2e-a55d-474b-be2b-37d32fa31b9b tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1001.978576] env[62066]: DEBUG oslo_concurrency.lockutils [None req-2cce9c2e-a55d-474b-be2b-37d32fa31b9b tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Acquired lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1001.978907] env[62066]: DEBUG oslo_concurrency.lockutils [None req-2cce9c2e-a55d-474b-be2b-37d32fa31b9b tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1001.979317] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3bf6b382-71bb-4e8f-be0d-9c28cad3bf9b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.983746] env[62066]: DEBUG oslo_vmware.api [None req-2cce9c2e-a55d-474b-be2b-37d32fa31b9b tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Waiting for the task: (returnval){ [ 1001.983746] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]52b46955-71cb-1365-d39b-3467dd3cbf2d" [ 1001.983746] env[62066]: _type = "Task" [ 1001.983746] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.993304] env[62066]: DEBUG oslo_vmware.api [None req-2cce9c2e-a55d-474b-be2b-37d32fa31b9b tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52b46955-71cb-1365-d39b-3467dd3cbf2d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.026388] env[62066]: DEBUG oslo_vmware.api [None req-b3f57e08-3a8e-4c04-9171-e3846694852d tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1341306, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.104076] env[62066]: DEBUG oslo_concurrency.lockutils [None req-77cb8d13-ef45-4b61-b6f4-a82b4c16e2d2 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Acquiring lock "c8e0d47c-4421-4e00-9183-206fceeabc40" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1002.104361] env[62066]: DEBUG oslo_concurrency.lockutils [None req-77cb8d13-ef45-4b61-b6f4-a82b4c16e2d2 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Lock "c8e0d47c-4421-4e00-9183-206fceeabc40" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1002.104579] env[62066]: DEBUG oslo_concurrency.lockutils [None req-77cb8d13-ef45-4b61-b6f4-a82b4c16e2d2 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Acquiring lock "c8e0d47c-4421-4e00-9183-206fceeabc40-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1002.104770] env[62066]: DEBUG oslo_concurrency.lockutils [None req-77cb8d13-ef45-4b61-b6f4-a82b4c16e2d2 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Lock "c8e0d47c-4421-4e00-9183-206fceeabc40-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1002.104950] env[62066]: DEBUG oslo_concurrency.lockutils [None req-77cb8d13-ef45-4b61-b6f4-a82b4c16e2d2 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Lock "c8e0d47c-4421-4e00-9183-206fceeabc40-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1002.107067] env[62066]: INFO nova.compute.manager [None req-77cb8d13-ef45-4b61-b6f4-a82b4c16e2d2 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: c8e0d47c-4421-4e00-9183-206fceeabc40] Terminating instance [ 1002.108820] env[62066]: DEBUG nova.compute.manager [None req-77cb8d13-ef45-4b61-b6f4-a82b4c16e2d2 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: c8e0d47c-4421-4e00-9183-206fceeabc40] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1002.109022] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-77cb8d13-ef45-4b61-b6f4-a82b4c16e2d2 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: c8e0d47c-4421-4e00-9183-206fceeabc40] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1002.109836] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8416e368-f300-4e56-b745-b72535a7da0c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.117336] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-77cb8d13-ef45-4b61-b6f4-a82b4c16e2d2 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: c8e0d47c-4421-4e00-9183-206fceeabc40] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1002.117573] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-cb7e4f06-76ef-4cad-adac-4b29426cfcd2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.124222] env[62066]: DEBUG oslo_vmware.api [None req-77cb8d13-ef45-4b61-b6f4-a82b4c16e2d2 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Waiting for the task: (returnval){ [ 1002.124222] env[62066]: value = "task-1341308" [ 1002.124222] env[62066]: _type = "Task" [ 1002.124222] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.131571] env[62066]: DEBUG oslo_vmware.api [None req-77cb8d13-ef45-4b61-b6f4-a82b4c16e2d2 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Task: {'id': task-1341308, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.201414] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-27337d51-e9e8-4b7e-b8c5-38dbed79187f tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: a2c5277e-af49-4b31-8480-a74d354c7383] VM already powered off {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1002.201632] env[62066]: DEBUG nova.compute.manager [None req-27337d51-e9e8-4b7e-b8c5-38dbed79187f tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: a2c5277e-af49-4b31-8480-a74d354c7383] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1002.202442] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21893577-3fef-460a-9eb0-59b170651862 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.208191] env[62066]: DEBUG oslo_concurrency.lockutils [None req-27337d51-e9e8-4b7e-b8c5-38dbed79187f tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Acquiring lock "refresh_cache-a2c5277e-af49-4b31-8480-a74d354c7383" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1002.208369] env[62066]: DEBUG oslo_concurrency.lockutils [None req-27337d51-e9e8-4b7e-b8c5-38dbed79187f tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Acquired lock "refresh_cache-a2c5277e-af49-4b31-8480-a74d354c7383" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1002.208549] env[62066]: DEBUG nova.network.neutron [None req-27337d51-e9e8-4b7e-b8c5-38dbed79187f tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: a2c5277e-af49-4b31-8480-a74d354c7383] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1002.332051] env[62066]: DEBUG nova.network.neutron [req-2a6009eb-8b0a-4004-8a50-26722eabaadf req-d068368d-a9ae-4ac5-873d-1585166024b9 service nova] [instance: 7178eb83-41e7-4516-809d-a0972d7e6e23] Updated VIF entry in instance network info cache for port 63f16c96-7de0-40e7-9c0c-782122865437. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1002.332439] env[62066]: DEBUG nova.network.neutron [req-2a6009eb-8b0a-4004-8a50-26722eabaadf req-d068368d-a9ae-4ac5-873d-1585166024b9 service nova] [instance: 7178eb83-41e7-4516-809d-a0972d7e6e23] Updating instance_info_cache with network_info: [{"id": "63f16c96-7de0-40e7-9c0c-782122865437", "address": "fa:16:3e:c6:a7:3e", "network": {"id": "449dfe4e-bb49-43b6-9ba7-b57af74ebfb3", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-282029676-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0d4e33a99c7741fb8cdd97f4ec5dbbd8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db68bd64-5b56-49af-a075-13dcf85cb2e0", "external-id": "nsx-vlan-transportzone-590", "segmentation_id": 590, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap63f16c96-7d", "ovs_interfaceid": "63f16c96-7de0-40e7-9c0c-782122865437", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1002.456061] env[62066]: DEBUG oslo_concurrency.lockutils [None req-711de73e-4168-430a-b07e-e18f814cac80 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1002.456414] env[62066]: DEBUG oslo_concurrency.lockutils [None req-711de73e-4168-430a-b07e-e18f814cac80 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1002.456555] env[62066]: DEBUG nova.objects.instance [None req-711de73e-4168-430a-b07e-e18f814cac80 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Lazy-loading 'resources' on Instance uuid a4ef97ad-bbb5-47fa-b5c1-2867a0d56435 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1002.495654] env[62066]: DEBUG oslo_vmware.api [None req-2cce9c2e-a55d-474b-be2b-37d32fa31b9b tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52b46955-71cb-1365-d39b-3467dd3cbf2d, 'name': SearchDatastore_Task, 'duration_secs': 0.010408} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.496322] env[62066]: DEBUG oslo_concurrency.lockutils [None req-2cce9c2e-a55d-474b-be2b-37d32fa31b9b tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Releasing lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1002.496552] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-2cce9c2e-a55d-474b-be2b-37d32fa31b9b tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: 7178eb83-41e7-4516-809d-a0972d7e6e23] Processing image 50ff584c-3b50-4395-af07-3e66769bc9f7 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1002.496780] env[62066]: DEBUG oslo_concurrency.lockutils [None req-2cce9c2e-a55d-474b-be2b-37d32fa31b9b tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1002.496931] env[62066]: DEBUG oslo_concurrency.lockutils [None req-2cce9c2e-a55d-474b-be2b-37d32fa31b9b tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Acquired lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1002.497123] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-2cce9c2e-a55d-474b-be2b-37d32fa31b9b tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1002.497377] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-095d441b-14a6-4a2a-ae8e-d30b47e70545 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.505112] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-2cce9c2e-a55d-474b-be2b-37d32fa31b9b tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1002.505294] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-2cce9c2e-a55d-474b-be2b-37d32fa31b9b tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1002.505965] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-499ca28f-713b-449d-87e5-c896b6e2f7ce {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.510589] env[62066]: DEBUG oslo_vmware.api [None req-2cce9c2e-a55d-474b-be2b-37d32fa31b9b tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Waiting for the task: (returnval){ [ 1002.510589] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]52ac0fce-a834-dab6-a059-e3bb12cfb48d" [ 1002.510589] env[62066]: _type = "Task" [ 1002.510589] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.517986] env[62066]: DEBUG oslo_vmware.api [None req-2cce9c2e-a55d-474b-be2b-37d32fa31b9b tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52ac0fce-a834-dab6-a059-e3bb12cfb48d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.524875] env[62066]: DEBUG oslo_vmware.api [None req-b3f57e08-3a8e-4c04-9171-e3846694852d tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1341306, 'name': PowerOnVM_Task, 'duration_secs': 0.57846} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.525122] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-b3f57e08-3a8e-4c04-9171-e3846694852d tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 9dc7c179-4360-425d-915f-c2d7bc591b1e] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1002.525327] env[62066]: INFO nova.compute.manager [None req-b3f57e08-3a8e-4c04-9171-e3846694852d tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 9dc7c179-4360-425d-915f-c2d7bc591b1e] Took 7.48 seconds to spawn the instance on the hypervisor. [ 1002.525509] env[62066]: DEBUG nova.compute.manager [None req-b3f57e08-3a8e-4c04-9171-e3846694852d tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 9dc7c179-4360-425d-915f-c2d7bc591b1e] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1002.526208] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-871a4c57-6388-4af8-8672-37215f1f3ba0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.634358] env[62066]: DEBUG oslo_vmware.api [None req-77cb8d13-ef45-4b61-b6f4-a82b4c16e2d2 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Task: {'id': task-1341308, 'name': PowerOffVM_Task, 'duration_secs': 0.195235} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.634640] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-77cb8d13-ef45-4b61-b6f4-a82b4c16e2d2 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: c8e0d47c-4421-4e00-9183-206fceeabc40] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1002.634931] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-77cb8d13-ef45-4b61-b6f4-a82b4c16e2d2 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: c8e0d47c-4421-4e00-9183-206fceeabc40] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1002.635073] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-86b16502-3aba-4975-8d8f-abd8adae5f2c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.704857] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-77cb8d13-ef45-4b61-b6f4-a82b4c16e2d2 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: c8e0d47c-4421-4e00-9183-206fceeabc40] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1002.705101] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-77cb8d13-ef45-4b61-b6f4-a82b4c16e2d2 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: c8e0d47c-4421-4e00-9183-206fceeabc40] Deleting contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1002.705291] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-77cb8d13-ef45-4b61-b6f4-a82b4c16e2d2 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Deleting the datastore file [datastore2] c8e0d47c-4421-4e00-9183-206fceeabc40 {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1002.705557] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d2ee6f06-5ac9-4ac3-8b7c-90418ed263dc {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.713359] env[62066]: DEBUG oslo_vmware.api [None req-77cb8d13-ef45-4b61-b6f4-a82b4c16e2d2 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Waiting for the task: (returnval){ [ 1002.713359] env[62066]: value = "task-1341310" [ 1002.713359] env[62066]: _type = "Task" [ 1002.713359] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.721198] env[62066]: DEBUG oslo_vmware.api [None req-77cb8d13-ef45-4b61-b6f4-a82b4c16e2d2 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Task: {'id': task-1341310, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.835528] env[62066]: DEBUG oslo_concurrency.lockutils [req-2a6009eb-8b0a-4004-8a50-26722eabaadf req-d068368d-a9ae-4ac5-873d-1585166024b9 service nova] Releasing lock "refresh_cache-7178eb83-41e7-4516-809d-a0972d7e6e23" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1002.835812] env[62066]: DEBUG nova.compute.manager [req-2a6009eb-8b0a-4004-8a50-26722eabaadf req-d068368d-a9ae-4ac5-873d-1585166024b9 service nova] [instance: a4ef97ad-bbb5-47fa-b5c1-2867a0d56435] Received event network-vif-deleted-4fd312f4-e41d-45eb-8505-fb30e3596b0a {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1002.930219] env[62066]: DEBUG nova.network.neutron [None req-27337d51-e9e8-4b7e-b8c5-38dbed79187f tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: a2c5277e-af49-4b31-8480-a74d354c7383] Updating instance_info_cache with network_info: [{"id": "370027d1-b4b7-4b8c-8258-8a5350637a2c", "address": "fa:16:3e:42:e1:91", "network": {"id": "7617f887-bdd0-400a-8b5f-606d7785df78", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-558766873-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8475ad5a900548cba568360999c846ea", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2b6a4065-12af-4fb9-ac47-ec9143f7297e", "external-id": "nsx-vlan-transportzone-95", "segmentation_id": 95, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap370027d1-b4", "ovs_interfaceid": "370027d1-b4b7-4b8c-8258-8a5350637a2c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1003.023126] env[62066]: DEBUG oslo_vmware.api [None req-2cce9c2e-a55d-474b-be2b-37d32fa31b9b tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52ac0fce-a834-dab6-a059-e3bb12cfb48d, 'name': SearchDatastore_Task, 'duration_secs': 0.007595} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.023920] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8187b754-583a-41b6-979c-83b5f10a1bfb {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.031547] env[62066]: DEBUG oslo_vmware.api [None req-2cce9c2e-a55d-474b-be2b-37d32fa31b9b tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Waiting for the task: (returnval){ [ 1003.031547] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]520e287d-9d50-0295-ae19-aafeeec8fbef" [ 1003.031547] env[62066]: _type = "Task" [ 1003.031547] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.042990] env[62066]: DEBUG oslo_vmware.api [None req-2cce9c2e-a55d-474b-be2b-37d32fa31b9b tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]520e287d-9d50-0295-ae19-aafeeec8fbef, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.047985] env[62066]: INFO nova.compute.manager [None req-b3f57e08-3a8e-4c04-9171-e3846694852d tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 9dc7c179-4360-425d-915f-c2d7bc591b1e] Took 13.09 seconds to build instance. [ 1003.066075] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fcc30d3-058d-440e-8cb9-80c16ca63774 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.074053] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7d5c6b1-158c-453c-8e36-92fdd769228f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.105503] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81d0f7a3-9a7c-4526-a68d-7ded0f3559fd {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.113558] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e16f6f11-aaed-42db-b9d0-f7e35b59ed59 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.126898] env[62066]: DEBUG nova.compute.provider_tree [None req-711de73e-4168-430a-b07e-e18f814cac80 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1003.223411] env[62066]: DEBUG oslo_vmware.api [None req-77cb8d13-ef45-4b61-b6f4-a82b4c16e2d2 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Task: {'id': task-1341310, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.140778} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.223596] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-77cb8d13-ef45-4b61-b6f4-a82b4c16e2d2 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1003.225061] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-77cb8d13-ef45-4b61-b6f4-a82b4c16e2d2 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: c8e0d47c-4421-4e00-9183-206fceeabc40] Deleted contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1003.225061] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-77cb8d13-ef45-4b61-b6f4-a82b4c16e2d2 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: c8e0d47c-4421-4e00-9183-206fceeabc40] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1003.225061] env[62066]: INFO nova.compute.manager [None req-77cb8d13-ef45-4b61-b6f4-a82b4c16e2d2 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: c8e0d47c-4421-4e00-9183-206fceeabc40] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1003.225061] env[62066]: DEBUG oslo.service.loopingcall [None req-77cb8d13-ef45-4b61-b6f4-a82b4c16e2d2 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1003.225061] env[62066]: DEBUG nova.compute.manager [-] [instance: c8e0d47c-4421-4e00-9183-206fceeabc40] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1003.225061] env[62066]: DEBUG nova.network.neutron [-] [instance: c8e0d47c-4421-4e00-9183-206fceeabc40] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1003.432782] env[62066]: DEBUG oslo_concurrency.lockutils [None req-27337d51-e9e8-4b7e-b8c5-38dbed79187f tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Releasing lock "refresh_cache-a2c5277e-af49-4b31-8480-a74d354c7383" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1003.541939] env[62066]: DEBUG oslo_vmware.api [None req-2cce9c2e-a55d-474b-be2b-37d32fa31b9b tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]520e287d-9d50-0295-ae19-aafeeec8fbef, 'name': SearchDatastore_Task, 'duration_secs': 0.009048} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.542316] env[62066]: DEBUG oslo_concurrency.lockutils [None req-2cce9c2e-a55d-474b-be2b-37d32fa31b9b tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Releasing lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1003.542510] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-2cce9c2e-a55d-474b-be2b-37d32fa31b9b tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk to [datastore2] 7178eb83-41e7-4516-809d-a0972d7e6e23/7178eb83-41e7-4516-809d-a0972d7e6e23.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1003.542824] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-fd8ddf5b-525a-464d-b74e-a3ef58a95ddf {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.549721] env[62066]: DEBUG oslo_concurrency.lockutils [None req-b3f57e08-3a8e-4c04-9171-e3846694852d tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Lock "9dc7c179-4360-425d-915f-c2d7bc591b1e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.595s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1003.550115] env[62066]: DEBUG oslo_vmware.api [None req-2cce9c2e-a55d-474b-be2b-37d32fa31b9b tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Waiting for the task: (returnval){ [ 1003.550115] env[62066]: value = "task-1341311" [ 1003.550115] env[62066]: _type = "Task" [ 1003.550115] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.558282] env[62066]: DEBUG oslo_vmware.api [None req-2cce9c2e-a55d-474b-be2b-37d32fa31b9b tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': task-1341311, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.630652] env[62066]: DEBUG nova.scheduler.client.report [None req-711de73e-4168-430a-b07e-e18f814cac80 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1003.711565] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-27337d51-e9e8-4b7e-b8c5-38dbed79187f tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: a2c5277e-af49-4b31-8480-a74d354c7383] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1003.712504] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f9c3849-bc7d-4b2d-a2a5-64b2cb3f8701 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.721753] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-27337d51-e9e8-4b7e-b8c5-38dbed79187f tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: a2c5277e-af49-4b31-8480-a74d354c7383] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1003.722075] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0bd0c3b0-5c51-45af-9ddb-ca53471b2f0b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.736991] env[62066]: DEBUG nova.compute.manager [req-9c0ebf1a-3990-4ca5-aa8e-6de07a8b03a1 req-53e0c9e0-85b1-427a-9c86-cd7e6776134d service nova] [instance: a2c5277e-af49-4b31-8480-a74d354c7383] Received event network-vif-unplugged-370027d1-b4b7-4b8c-8258-8a5350637a2c {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1003.737242] env[62066]: DEBUG oslo_concurrency.lockutils [req-9c0ebf1a-3990-4ca5-aa8e-6de07a8b03a1 req-53e0c9e0-85b1-427a-9c86-cd7e6776134d service nova] Acquiring lock "a2c5277e-af49-4b31-8480-a74d354c7383-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1003.737467] env[62066]: DEBUG oslo_concurrency.lockutils [req-9c0ebf1a-3990-4ca5-aa8e-6de07a8b03a1 req-53e0c9e0-85b1-427a-9c86-cd7e6776134d service nova] Lock "a2c5277e-af49-4b31-8480-a74d354c7383-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1003.737642] env[62066]: DEBUG oslo_concurrency.lockutils [req-9c0ebf1a-3990-4ca5-aa8e-6de07a8b03a1 req-53e0c9e0-85b1-427a-9c86-cd7e6776134d service nova] Lock "a2c5277e-af49-4b31-8480-a74d354c7383-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1003.737834] env[62066]: DEBUG nova.compute.manager [req-9c0ebf1a-3990-4ca5-aa8e-6de07a8b03a1 req-53e0c9e0-85b1-427a-9c86-cd7e6776134d service nova] [instance: a2c5277e-af49-4b31-8480-a74d354c7383] No waiting events found dispatching network-vif-unplugged-370027d1-b4b7-4b8c-8258-8a5350637a2c {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1003.738216] env[62066]: WARNING nova.compute.manager [req-9c0ebf1a-3990-4ca5-aa8e-6de07a8b03a1 req-53e0c9e0-85b1-427a-9c86-cd7e6776134d service nova] [instance: a2c5277e-af49-4b31-8480-a74d354c7383] Received unexpected event network-vif-unplugged-370027d1-b4b7-4b8c-8258-8a5350637a2c for instance with vm_state shelved and task_state shelving_offloading. [ 1003.788436] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-27337d51-e9e8-4b7e-b8c5-38dbed79187f tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: a2c5277e-af49-4b31-8480-a74d354c7383] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1003.788731] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-27337d51-e9e8-4b7e-b8c5-38dbed79187f tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: a2c5277e-af49-4b31-8480-a74d354c7383] Deleting contents of the VM from datastore datastore1 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1003.788927] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-27337d51-e9e8-4b7e-b8c5-38dbed79187f tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Deleting the datastore file [datastore1] a2c5277e-af49-4b31-8480-a74d354c7383 {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1003.789210] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c20b2a51-d28f-4900-9052-381adbf818fe {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.793326] env[62066]: DEBUG nova.compute.manager [req-d9a5d7af-9a7c-48db-b5b3-0889d37e4e16 req-eb97a1b2-4451-4e1d-ae9b-643fa38ae2b6 service nova] [instance: c8e0d47c-4421-4e00-9183-206fceeabc40] Received event network-vif-deleted-bb9924e0-8b18-43b2-bda0-3e409fa16718 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1003.793543] env[62066]: INFO nova.compute.manager [req-d9a5d7af-9a7c-48db-b5b3-0889d37e4e16 req-eb97a1b2-4451-4e1d-ae9b-643fa38ae2b6 service nova] [instance: c8e0d47c-4421-4e00-9183-206fceeabc40] Neutron deleted interface bb9924e0-8b18-43b2-bda0-3e409fa16718; detaching it from the instance and deleting it from the info cache [ 1003.793749] env[62066]: DEBUG nova.network.neutron [req-d9a5d7af-9a7c-48db-b5b3-0889d37e4e16 req-eb97a1b2-4451-4e1d-ae9b-643fa38ae2b6 service nova] [instance: c8e0d47c-4421-4e00-9183-206fceeabc40] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1003.800071] env[62066]: DEBUG oslo_vmware.api [None req-27337d51-e9e8-4b7e-b8c5-38dbed79187f tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Waiting for the task: (returnval){ [ 1003.800071] env[62066]: value = "task-1341313" [ 1003.800071] env[62066]: _type = "Task" [ 1003.800071] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.813810] env[62066]: DEBUG oslo_vmware.api [None req-27337d51-e9e8-4b7e-b8c5-38dbed79187f tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341313, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.060957] env[62066]: DEBUG oslo_vmware.api [None req-2cce9c2e-a55d-474b-be2b-37d32fa31b9b tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': task-1341311, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.492451} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.061261] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-2cce9c2e-a55d-474b-be2b-37d32fa31b9b tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk to [datastore2] 7178eb83-41e7-4516-809d-a0972d7e6e23/7178eb83-41e7-4516-809d-a0972d7e6e23.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1004.061482] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-2cce9c2e-a55d-474b-be2b-37d32fa31b9b tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: 7178eb83-41e7-4516-809d-a0972d7e6e23] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1004.061765] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-343858ac-9c5c-4c46-aca4-871c45b21855 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.067919] env[62066]: DEBUG oslo_vmware.api [None req-2cce9c2e-a55d-474b-be2b-37d32fa31b9b tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Waiting for the task: (returnval){ [ 1004.067919] env[62066]: value = "task-1341314" [ 1004.067919] env[62066]: _type = "Task" [ 1004.067919] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.076806] env[62066]: DEBUG oslo_vmware.api [None req-2cce9c2e-a55d-474b-be2b-37d32fa31b9b tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': task-1341314, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.135952] env[62066]: DEBUG oslo_concurrency.lockutils [None req-711de73e-4168-430a-b07e-e18f814cac80 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.679s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1004.160035] env[62066]: INFO nova.scheduler.client.report [None req-711de73e-4168-430a-b07e-e18f814cac80 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Deleted allocations for instance a4ef97ad-bbb5-47fa-b5c1-2867a0d56435 [ 1004.266756] env[62066]: DEBUG nova.network.neutron [-] [instance: c8e0d47c-4421-4e00-9183-206fceeabc40] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1004.297441] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d2f871a1-1e10-4298-aa3b-621253d197e0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.308555] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4636fdc-6455-4ff6-8b1c-b8a6cce1740c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.321996] env[62066]: DEBUG oslo_vmware.api [None req-27337d51-e9e8-4b7e-b8c5-38dbed79187f tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341313, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.277631} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.322657] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-27337d51-e9e8-4b7e-b8c5-38dbed79187f tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1004.322888] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-27337d51-e9e8-4b7e-b8c5-38dbed79187f tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: a2c5277e-af49-4b31-8480-a74d354c7383] Deleted contents of the VM from datastore datastore1 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1004.323131] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-27337d51-e9e8-4b7e-b8c5-38dbed79187f tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: a2c5277e-af49-4b31-8480-a74d354c7383] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1004.336457] env[62066]: DEBUG nova.compute.manager [req-d9a5d7af-9a7c-48db-b5b3-0889d37e4e16 req-eb97a1b2-4451-4e1d-ae9b-643fa38ae2b6 service nova] [instance: c8e0d47c-4421-4e00-9183-206fceeabc40] Detach interface failed, port_id=bb9924e0-8b18-43b2-bda0-3e409fa16718, reason: Instance c8e0d47c-4421-4e00-9183-206fceeabc40 could not be found. {{(pid=62066) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1004.343039] env[62066]: INFO nova.scheduler.client.report [None req-27337d51-e9e8-4b7e-b8c5-38dbed79187f tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Deleted allocations for instance a2c5277e-af49-4b31-8480-a74d354c7383 [ 1004.482800] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c56f704c-7907-46df-ae47-acf7d6ca81d1 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Acquiring lock "9dc7c179-4360-425d-915f-c2d7bc591b1e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1004.483264] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c56f704c-7907-46df-ae47-acf7d6ca81d1 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Lock "9dc7c179-4360-425d-915f-c2d7bc591b1e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1004.483535] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c56f704c-7907-46df-ae47-acf7d6ca81d1 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Acquiring lock "9dc7c179-4360-425d-915f-c2d7bc591b1e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1004.483798] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c56f704c-7907-46df-ae47-acf7d6ca81d1 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Lock "9dc7c179-4360-425d-915f-c2d7bc591b1e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1004.484060] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c56f704c-7907-46df-ae47-acf7d6ca81d1 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Lock "9dc7c179-4360-425d-915f-c2d7bc591b1e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1004.487681] env[62066]: INFO nova.compute.manager [None req-c56f704c-7907-46df-ae47-acf7d6ca81d1 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 9dc7c179-4360-425d-915f-c2d7bc591b1e] Terminating instance [ 1004.489726] env[62066]: DEBUG nova.compute.manager [None req-c56f704c-7907-46df-ae47-acf7d6ca81d1 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 9dc7c179-4360-425d-915f-c2d7bc591b1e] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1004.489932] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-c56f704c-7907-46df-ae47-acf7d6ca81d1 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 9dc7c179-4360-425d-915f-c2d7bc591b1e] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1004.490858] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae382f63-77e1-4084-8746-74db66425d69 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.498290] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-c56f704c-7907-46df-ae47-acf7d6ca81d1 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 9dc7c179-4360-425d-915f-c2d7bc591b1e] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1004.498554] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4c2a4a12-e606-4883-b6bd-4290ea953f16 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.505107] env[62066]: DEBUG oslo_vmware.api [None req-c56f704c-7907-46df-ae47-acf7d6ca81d1 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Waiting for the task: (returnval){ [ 1004.505107] env[62066]: value = "task-1341315" [ 1004.505107] env[62066]: _type = "Task" [ 1004.505107] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.512925] env[62066]: DEBUG oslo_vmware.api [None req-c56f704c-7907-46df-ae47-acf7d6ca81d1 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1341315, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.578335] env[62066]: DEBUG oslo_vmware.api [None req-2cce9c2e-a55d-474b-be2b-37d32fa31b9b tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': task-1341314, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.060289} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.578749] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-2cce9c2e-a55d-474b-be2b-37d32fa31b9b tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: 7178eb83-41e7-4516-809d-a0972d7e6e23] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1004.579970] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc847694-c4fb-4ab9-b5d3-2e35c5f88fb1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.604278] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-2cce9c2e-a55d-474b-be2b-37d32fa31b9b tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: 7178eb83-41e7-4516-809d-a0972d7e6e23] Reconfiguring VM instance instance-00000065 to attach disk [datastore2] 7178eb83-41e7-4516-809d-a0972d7e6e23/7178eb83-41e7-4516-809d-a0972d7e6e23.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1004.604526] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1a71243a-e9d5-4bad-923f-4f1a05f79ef9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.625016] env[62066]: DEBUG oslo_vmware.api [None req-2cce9c2e-a55d-474b-be2b-37d32fa31b9b tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Waiting for the task: (returnval){ [ 1004.625016] env[62066]: value = "task-1341316" [ 1004.625016] env[62066]: _type = "Task" [ 1004.625016] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.633166] env[62066]: DEBUG oslo_vmware.api [None req-2cce9c2e-a55d-474b-be2b-37d32fa31b9b tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': task-1341316, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.667707] env[62066]: DEBUG oslo_concurrency.lockutils [None req-711de73e-4168-430a-b07e-e18f814cac80 tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Lock "a4ef97ad-bbb5-47fa-b5c1-2867a0d56435" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.904s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1004.769514] env[62066]: INFO nova.compute.manager [-] [instance: c8e0d47c-4421-4e00-9183-206fceeabc40] Took 1.54 seconds to deallocate network for instance. [ 1004.847545] env[62066]: DEBUG oslo_concurrency.lockutils [None req-27337d51-e9e8-4b7e-b8c5-38dbed79187f tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1004.847858] env[62066]: DEBUG oslo_concurrency.lockutils [None req-27337d51-e9e8-4b7e-b8c5-38dbed79187f tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1004.848132] env[62066]: DEBUG nova.objects.instance [None req-27337d51-e9e8-4b7e-b8c5-38dbed79187f tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Lazy-loading 'resources' on Instance uuid a2c5277e-af49-4b31-8480-a74d354c7383 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1005.015773] env[62066]: DEBUG oslo_vmware.api [None req-c56f704c-7907-46df-ae47-acf7d6ca81d1 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1341315, 'name': PowerOffVM_Task, 'duration_secs': 0.197902} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.016116] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-c56f704c-7907-46df-ae47-acf7d6ca81d1 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 9dc7c179-4360-425d-915f-c2d7bc591b1e] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1005.016295] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-c56f704c-7907-46df-ae47-acf7d6ca81d1 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 9dc7c179-4360-425d-915f-c2d7bc591b1e] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1005.016553] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8b14d849-0589-48d5-8f7f-c2432e852ae1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.077795] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-c56f704c-7907-46df-ae47-acf7d6ca81d1 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 9dc7c179-4360-425d-915f-c2d7bc591b1e] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1005.078195] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-c56f704c-7907-46df-ae47-acf7d6ca81d1 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 9dc7c179-4360-425d-915f-c2d7bc591b1e] Deleting contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1005.078534] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-c56f704c-7907-46df-ae47-acf7d6ca81d1 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Deleting the datastore file [datastore2] 9dc7c179-4360-425d-915f-c2d7bc591b1e {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1005.079182] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0c23e019-2cc5-42a5-a20b-2e453e54ba33 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.086633] env[62066]: DEBUG oslo_vmware.api [None req-c56f704c-7907-46df-ae47-acf7d6ca81d1 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Waiting for the task: (returnval){ [ 1005.086633] env[62066]: value = "task-1341318" [ 1005.086633] env[62066]: _type = "Task" [ 1005.086633] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.094843] env[62066]: DEBUG oslo_vmware.api [None req-c56f704c-7907-46df-ae47-acf7d6ca81d1 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1341318, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.135872] env[62066]: DEBUG oslo_vmware.api [None req-2cce9c2e-a55d-474b-be2b-37d32fa31b9b tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': task-1341316, 'name': ReconfigVM_Task, 'duration_secs': 0.289019} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.136194] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-2cce9c2e-a55d-474b-be2b-37d32fa31b9b tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: 7178eb83-41e7-4516-809d-a0972d7e6e23] Reconfigured VM instance instance-00000065 to attach disk [datastore2] 7178eb83-41e7-4516-809d-a0972d7e6e23/7178eb83-41e7-4516-809d-a0972d7e6e23.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1005.136903] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-953fa95e-fa55-4422-a131-0df2e07c9889 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.143781] env[62066]: DEBUG oslo_vmware.api [None req-2cce9c2e-a55d-474b-be2b-37d32fa31b9b tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Waiting for the task: (returnval){ [ 1005.143781] env[62066]: value = "task-1341319" [ 1005.143781] env[62066]: _type = "Task" [ 1005.143781] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.152804] env[62066]: DEBUG oslo_vmware.api [None req-2cce9c2e-a55d-474b-be2b-37d32fa31b9b tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': task-1341319, 'name': Rename_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.277155] env[62066]: DEBUG oslo_concurrency.lockutils [None req-77cb8d13-ef45-4b61-b6f4-a82b4c16e2d2 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1005.350461] env[62066]: DEBUG nova.objects.instance [None req-27337d51-e9e8-4b7e-b8c5-38dbed79187f tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Lazy-loading 'numa_topology' on Instance uuid a2c5277e-af49-4b31-8480-a74d354c7383 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1005.573614] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c80a9ac9-d682-4646-84b4-272b36063bda tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Acquiring lock "02fb3e92-5dd6-4b1e-a6e4-d60d3fc85b07" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1005.573900] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c80a9ac9-d682-4646-84b4-272b36063bda tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Lock "02fb3e92-5dd6-4b1e-a6e4-d60d3fc85b07" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1005.574132] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c80a9ac9-d682-4646-84b4-272b36063bda tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Acquiring lock "02fb3e92-5dd6-4b1e-a6e4-d60d3fc85b07-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1005.574325] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c80a9ac9-d682-4646-84b4-272b36063bda tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Lock "02fb3e92-5dd6-4b1e-a6e4-d60d3fc85b07-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1005.574545] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c80a9ac9-d682-4646-84b4-272b36063bda tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Lock "02fb3e92-5dd6-4b1e-a6e4-d60d3fc85b07-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1005.576802] env[62066]: INFO nova.compute.manager [None req-c80a9ac9-d682-4646-84b4-272b36063bda tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 02fb3e92-5dd6-4b1e-a6e4-d60d3fc85b07] Terminating instance [ 1005.579805] env[62066]: DEBUG nova.compute.manager [None req-c80a9ac9-d682-4646-84b4-272b36063bda tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 02fb3e92-5dd6-4b1e-a6e4-d60d3fc85b07] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1005.580130] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-c80a9ac9-d682-4646-84b4-272b36063bda tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 02fb3e92-5dd6-4b1e-a6e4-d60d3fc85b07] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1005.580886] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd32efcd-eea0-4a8f-a916-4828f724d6c3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.588499] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-c80a9ac9-d682-4646-84b4-272b36063bda tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 02fb3e92-5dd6-4b1e-a6e4-d60d3fc85b07] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1005.591389] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2393b587-439e-42e2-84d2-96117f4f92dd {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.597231] env[62066]: DEBUG oslo_vmware.api [None req-c56f704c-7907-46df-ae47-acf7d6ca81d1 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Task: {'id': task-1341318, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.175681} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.598299] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-c56f704c-7907-46df-ae47-acf7d6ca81d1 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1005.598494] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-c56f704c-7907-46df-ae47-acf7d6ca81d1 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 9dc7c179-4360-425d-915f-c2d7bc591b1e] Deleted contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1005.598688] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-c56f704c-7907-46df-ae47-acf7d6ca81d1 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 9dc7c179-4360-425d-915f-c2d7bc591b1e] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1005.598868] env[62066]: INFO nova.compute.manager [None req-c56f704c-7907-46df-ae47-acf7d6ca81d1 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] [instance: 9dc7c179-4360-425d-915f-c2d7bc591b1e] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1005.599203] env[62066]: DEBUG oslo.service.loopingcall [None req-c56f704c-7907-46df-ae47-acf7d6ca81d1 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1005.599429] env[62066]: DEBUG oslo_vmware.api [None req-c80a9ac9-d682-4646-84b4-272b36063bda tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Waiting for the task: (returnval){ [ 1005.599429] env[62066]: value = "task-1341320" [ 1005.599429] env[62066]: _type = "Task" [ 1005.599429] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.599620] env[62066]: DEBUG nova.compute.manager [-] [instance: 9dc7c179-4360-425d-915f-c2d7bc591b1e] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1005.599716] env[62066]: DEBUG nova.network.neutron [-] [instance: 9dc7c179-4360-425d-915f-c2d7bc591b1e] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1005.610465] env[62066]: DEBUG oslo_vmware.api [None req-c80a9ac9-d682-4646-84b4-272b36063bda tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': task-1341320, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.652775] env[62066]: DEBUG oslo_vmware.api [None req-2cce9c2e-a55d-474b-be2b-37d32fa31b9b tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': task-1341319, 'name': Rename_Task, 'duration_secs': 0.157186} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.653203] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-2cce9c2e-a55d-474b-be2b-37d32fa31b9b tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: 7178eb83-41e7-4516-809d-a0972d7e6e23] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1005.653429] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e298b080-4832-4ba2-8b7b-38eac5fe5be2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.659467] env[62066]: DEBUG oslo_vmware.api [None req-2cce9c2e-a55d-474b-be2b-37d32fa31b9b tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Waiting for the task: (returnval){ [ 1005.659467] env[62066]: value = "task-1341321" [ 1005.659467] env[62066]: _type = "Task" [ 1005.659467] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.666913] env[62066]: DEBUG oslo_vmware.api [None req-2cce9c2e-a55d-474b-be2b-37d32fa31b9b tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': task-1341321, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.762533] env[62066]: DEBUG nova.compute.manager [req-2242e646-be9a-4002-8e39-3744d6fd989c req-c8476b2f-9162-40e8-91b6-44bd6b53f062 service nova] [instance: a2c5277e-af49-4b31-8480-a74d354c7383] Received event network-changed-370027d1-b4b7-4b8c-8258-8a5350637a2c {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1005.762755] env[62066]: DEBUG nova.compute.manager [req-2242e646-be9a-4002-8e39-3744d6fd989c req-c8476b2f-9162-40e8-91b6-44bd6b53f062 service nova] [instance: a2c5277e-af49-4b31-8480-a74d354c7383] Refreshing instance network info cache due to event network-changed-370027d1-b4b7-4b8c-8258-8a5350637a2c. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1005.762979] env[62066]: DEBUG oslo_concurrency.lockutils [req-2242e646-be9a-4002-8e39-3744d6fd989c req-c8476b2f-9162-40e8-91b6-44bd6b53f062 service nova] Acquiring lock "refresh_cache-a2c5277e-af49-4b31-8480-a74d354c7383" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1005.763150] env[62066]: DEBUG oslo_concurrency.lockutils [req-2242e646-be9a-4002-8e39-3744d6fd989c req-c8476b2f-9162-40e8-91b6-44bd6b53f062 service nova] Acquired lock "refresh_cache-a2c5277e-af49-4b31-8480-a74d354c7383" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1005.763315] env[62066]: DEBUG nova.network.neutron [req-2242e646-be9a-4002-8e39-3744d6fd989c req-c8476b2f-9162-40e8-91b6-44bd6b53f062 service nova] [instance: a2c5277e-af49-4b31-8480-a74d354c7383] Refreshing network info cache for port 370027d1-b4b7-4b8c-8258-8a5350637a2c {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1005.837970] env[62066]: DEBUG nova.compute.manager [req-cf78d94b-19bd-4151-807a-17bd01cb1409 req-f23e2ded-8170-4421-9fe5-579924e71ada service nova] [instance: 9dc7c179-4360-425d-915f-c2d7bc591b1e] Received event network-vif-deleted-8499c7af-d27f-4b10-b63d-a50b2a34d620 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1005.837970] env[62066]: INFO nova.compute.manager [req-cf78d94b-19bd-4151-807a-17bd01cb1409 req-f23e2ded-8170-4421-9fe5-579924e71ada service nova] [instance: 9dc7c179-4360-425d-915f-c2d7bc591b1e] Neutron deleted interface 8499c7af-d27f-4b10-b63d-a50b2a34d620; detaching it from the instance and deleting it from the info cache [ 1005.837970] env[62066]: DEBUG nova.network.neutron [req-cf78d94b-19bd-4151-807a-17bd01cb1409 req-f23e2ded-8170-4421-9fe5-579924e71ada service nova] [instance: 9dc7c179-4360-425d-915f-c2d7bc591b1e] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1005.852909] env[62066]: DEBUG nova.objects.base [None req-27337d51-e9e8-4b7e-b8c5-38dbed79187f tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Object Instance lazy-loaded attributes: resources,numa_topology {{(pid=62066) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1005.928214] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-395ba5ec-df8e-4e69-98a5-eea9d7dfebbf {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.935862] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4802c4e0-efef-4ae0-82d8-4b543fee6eb2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.963927] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12e586f4-6bbf-434b-9e41-de46c91aefc7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.970717] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69b1ad60-91a3-456a-ab92-36ba60faf15f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.985084] env[62066]: DEBUG nova.compute.provider_tree [None req-27337d51-e9e8-4b7e-b8c5-38dbed79187f tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1005.989031] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4d5077a1-0ba8-4e51-9357-618c8fde121e tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Acquiring lock "a2c5277e-af49-4b31-8480-a74d354c7383" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1006.111658] env[62066]: DEBUG oslo_vmware.api [None req-c80a9ac9-d682-4646-84b4-272b36063bda tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': task-1341320, 'name': PowerOffVM_Task, 'duration_secs': 0.219809} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.113071] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-c80a9ac9-d682-4646-84b4-272b36063bda tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 02fb3e92-5dd6-4b1e-a6e4-d60d3fc85b07] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1006.113071] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-c80a9ac9-d682-4646-84b4-272b36063bda tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 02fb3e92-5dd6-4b1e-a6e4-d60d3fc85b07] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1006.113071] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-22485b90-95f1-4957-a7db-a3bae35a2f40 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.169015] env[62066]: DEBUG oslo_vmware.api [None req-2cce9c2e-a55d-474b-be2b-37d32fa31b9b tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': task-1341321, 'name': PowerOnVM_Task, 'duration_secs': 0.442922} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.169307] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-2cce9c2e-a55d-474b-be2b-37d32fa31b9b tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: 7178eb83-41e7-4516-809d-a0972d7e6e23] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1006.169518] env[62066]: INFO nova.compute.manager [None req-2cce9c2e-a55d-474b-be2b-37d32fa31b9b tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: 7178eb83-41e7-4516-809d-a0972d7e6e23] Took 7.11 seconds to spawn the instance on the hypervisor. [ 1006.169704] env[62066]: DEBUG nova.compute.manager [None req-2cce9c2e-a55d-474b-be2b-37d32fa31b9b tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: 7178eb83-41e7-4516-809d-a0972d7e6e23] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1006.170489] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a517d045-2fe6-46be-9607-6ee57e518669 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.173813] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-c80a9ac9-d682-4646-84b4-272b36063bda tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 02fb3e92-5dd6-4b1e-a6e4-d60d3fc85b07] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1006.174026] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-c80a9ac9-d682-4646-84b4-272b36063bda tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 02fb3e92-5dd6-4b1e-a6e4-d60d3fc85b07] Deleting contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1006.174214] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-c80a9ac9-d682-4646-84b4-272b36063bda tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Deleting the datastore file [datastore2] 02fb3e92-5dd6-4b1e-a6e4-d60d3fc85b07 {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1006.174783] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b72112af-9bc5-4277-b317-d88582951de3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.184502] env[62066]: DEBUG oslo_vmware.api [None req-c80a9ac9-d682-4646-84b4-272b36063bda tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Waiting for the task: (returnval){ [ 1006.184502] env[62066]: value = "task-1341323" [ 1006.184502] env[62066]: _type = "Task" [ 1006.184502] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1006.191829] env[62066]: DEBUG oslo_vmware.api [None req-c80a9ac9-d682-4646-84b4-272b36063bda tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': task-1341323, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.314882] env[62066]: DEBUG nova.network.neutron [-] [instance: 9dc7c179-4360-425d-915f-c2d7bc591b1e] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1006.340818] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e0353150-de3d-40c4-80da-2904526b25d7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.350394] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-970e14b7-1c75-43e3-8135-78d88b33495f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.373895] env[62066]: DEBUG nova.compute.manager [req-cf78d94b-19bd-4151-807a-17bd01cb1409 req-f23e2ded-8170-4421-9fe5-579924e71ada service nova] [instance: 9dc7c179-4360-425d-915f-c2d7bc591b1e] Detach interface failed, port_id=8499c7af-d27f-4b10-b63d-a50b2a34d620, reason: Instance 9dc7c179-4360-425d-915f-c2d7bc591b1e could not be found. {{(pid=62066) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1006.488157] env[62066]: DEBUG nova.scheduler.client.report [None req-27337d51-e9e8-4b7e-b8c5-38dbed79187f tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1006.586131] env[62066]: DEBUG nova.network.neutron [req-2242e646-be9a-4002-8e39-3744d6fd989c req-c8476b2f-9162-40e8-91b6-44bd6b53f062 service nova] [instance: a2c5277e-af49-4b31-8480-a74d354c7383] Updated VIF entry in instance network info cache for port 370027d1-b4b7-4b8c-8258-8a5350637a2c. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1006.586514] env[62066]: DEBUG nova.network.neutron [req-2242e646-be9a-4002-8e39-3744d6fd989c req-c8476b2f-9162-40e8-91b6-44bd6b53f062 service nova] [instance: a2c5277e-af49-4b31-8480-a74d354c7383] Updating instance_info_cache with network_info: [{"id": "370027d1-b4b7-4b8c-8258-8a5350637a2c", "address": "fa:16:3e:42:e1:91", "network": {"id": "7617f887-bdd0-400a-8b5f-606d7785df78", "bridge": null, "label": "tempest-DeleteServersTestJSON-558766873-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8475ad5a900548cba568360999c846ea", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap370027d1-b4", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1006.692167] env[62066]: INFO nova.compute.manager [None req-2cce9c2e-a55d-474b-be2b-37d32fa31b9b tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: 7178eb83-41e7-4516-809d-a0972d7e6e23] Took 11.87 seconds to build instance. [ 1006.696846] env[62066]: DEBUG oslo_vmware.api [None req-c80a9ac9-d682-4646-84b4-272b36063bda tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Task: {'id': task-1341323, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.138972} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.697101] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-c80a9ac9-d682-4646-84b4-272b36063bda tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1006.697292] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-c80a9ac9-d682-4646-84b4-272b36063bda tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 02fb3e92-5dd6-4b1e-a6e4-d60d3fc85b07] Deleted contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1006.697476] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-c80a9ac9-d682-4646-84b4-272b36063bda tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 02fb3e92-5dd6-4b1e-a6e4-d60d3fc85b07] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1006.697650] env[62066]: INFO nova.compute.manager [None req-c80a9ac9-d682-4646-84b4-272b36063bda tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] [instance: 02fb3e92-5dd6-4b1e-a6e4-d60d3fc85b07] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1006.697984] env[62066]: DEBUG oslo.service.loopingcall [None req-c80a9ac9-d682-4646-84b4-272b36063bda tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1006.698239] env[62066]: DEBUG nova.compute.manager [-] [instance: 02fb3e92-5dd6-4b1e-a6e4-d60d3fc85b07] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1006.698374] env[62066]: DEBUG nova.network.neutron [-] [instance: 02fb3e92-5dd6-4b1e-a6e4-d60d3fc85b07] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1006.819273] env[62066]: INFO nova.compute.manager [-] [instance: 9dc7c179-4360-425d-915f-c2d7bc591b1e] Took 1.22 seconds to deallocate network for instance. [ 1006.995029] env[62066]: DEBUG oslo_concurrency.lockutils [None req-27337d51-e9e8-4b7e-b8c5-38dbed79187f tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.145s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1006.995791] env[62066]: DEBUG oslo_concurrency.lockutils [None req-77cb8d13-ef45-4b61-b6f4-a82b4c16e2d2 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.719s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1006.996190] env[62066]: DEBUG nova.objects.instance [None req-77cb8d13-ef45-4b61-b6f4-a82b4c16e2d2 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Lazy-loading 'resources' on Instance uuid c8e0d47c-4421-4e00-9183-206fceeabc40 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1007.089604] env[62066]: DEBUG oslo_concurrency.lockutils [req-2242e646-be9a-4002-8e39-3744d6fd989c req-c8476b2f-9162-40e8-91b6-44bd6b53f062 service nova] Releasing lock "refresh_cache-a2c5277e-af49-4b31-8480-a74d354c7383" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1007.194591] env[62066]: DEBUG oslo_concurrency.lockutils [None req-2cce9c2e-a55d-474b-be2b-37d32fa31b9b tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Lock "7178eb83-41e7-4516-809d-a0972d7e6e23" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.385s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1007.324748] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c56f704c-7907-46df-ae47-acf7d6ca81d1 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1007.424362] env[62066]: DEBUG nova.network.neutron [-] [instance: 02fb3e92-5dd6-4b1e-a6e4-d60d3fc85b07] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1007.503187] env[62066]: DEBUG oslo_concurrency.lockutils [None req-27337d51-e9e8-4b7e-b8c5-38dbed79187f tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Lock "a2c5277e-af49-4b31-8480-a74d354c7383" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 21.324s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1007.504041] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4d5077a1-0ba8-4e51-9357-618c8fde121e tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Lock "a2c5277e-af49-4b31-8480-a74d354c7383" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 1.515s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1007.504271] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4d5077a1-0ba8-4e51-9357-618c8fde121e tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Acquiring lock "a2c5277e-af49-4b31-8480-a74d354c7383-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1007.504481] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4d5077a1-0ba8-4e51-9357-618c8fde121e tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Lock "a2c5277e-af49-4b31-8480-a74d354c7383-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1007.504650] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4d5077a1-0ba8-4e51-9357-618c8fde121e tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Lock "a2c5277e-af49-4b31-8480-a74d354c7383-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1007.507763] env[62066]: INFO nova.compute.manager [None req-4d5077a1-0ba8-4e51-9357-618c8fde121e tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: a2c5277e-af49-4b31-8480-a74d354c7383] Terminating instance [ 1007.509814] env[62066]: DEBUG nova.compute.manager [None req-4d5077a1-0ba8-4e51-9357-618c8fde121e tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: a2c5277e-af49-4b31-8480-a74d354c7383] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1007.509980] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-4d5077a1-0ba8-4e51-9357-618c8fde121e tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: a2c5277e-af49-4b31-8480-a74d354c7383] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1007.510915] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3b377bbc-079c-4cda-94e8-247e01f50826 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.522644] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-baab2aca-9f51-4823-815f-4e38294977b6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.546323] env[62066]: WARNING nova.virt.vmwareapi.vmops [None req-4d5077a1-0ba8-4e51-9357-618c8fde121e tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: a2c5277e-af49-4b31-8480-a74d354c7383] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance a2c5277e-af49-4b31-8480-a74d354c7383 could not be found. [ 1007.546547] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-4d5077a1-0ba8-4e51-9357-618c8fde121e tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: a2c5277e-af49-4b31-8480-a74d354c7383] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1007.546762] env[62066]: INFO nova.compute.manager [None req-4d5077a1-0ba8-4e51-9357-618c8fde121e tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: a2c5277e-af49-4b31-8480-a74d354c7383] Took 0.04 seconds to destroy the instance on the hypervisor. [ 1007.547113] env[62066]: DEBUG oslo.service.loopingcall [None req-4d5077a1-0ba8-4e51-9357-618c8fde121e tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1007.549500] env[62066]: DEBUG nova.compute.manager [-] [instance: a2c5277e-af49-4b31-8480-a74d354c7383] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1007.549605] env[62066]: DEBUG nova.network.neutron [-] [instance: a2c5277e-af49-4b31-8480-a74d354c7383] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1007.595199] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f56e9eb-d548-45a7-9704-e4c281bb3553 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.603157] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9cfe3fd-9a05-4e87-a461-bed9ca4e6e95 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.633191] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc7b5409-4b7a-4599-8e66-dde4d305f39b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.641058] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d10a0819-144f-479e-995f-22a25cbef14c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.656215] env[62066]: DEBUG nova.compute.provider_tree [None req-77cb8d13-ef45-4b61-b6f4-a82b4c16e2d2 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1007.800895] env[62066]: DEBUG nova.compute.manager [req-ca4a485a-ec0b-4758-bd4c-212af00683cb req-deb2052c-948a-49ff-96ef-df9e830dbc53 service nova] [instance: c8482d92-0180-44ee-a4f1-bf84786dad43] Received event network-changed-2f2f1786-934d-478f-be39-7509cfacf86f {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1007.801910] env[62066]: DEBUG nova.compute.manager [req-ca4a485a-ec0b-4758-bd4c-212af00683cb req-deb2052c-948a-49ff-96ef-df9e830dbc53 service nova] [instance: c8482d92-0180-44ee-a4f1-bf84786dad43] Refreshing instance network info cache due to event network-changed-2f2f1786-934d-478f-be39-7509cfacf86f. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1007.802593] env[62066]: DEBUG oslo_concurrency.lockutils [req-ca4a485a-ec0b-4758-bd4c-212af00683cb req-deb2052c-948a-49ff-96ef-df9e830dbc53 service nova] Acquiring lock "refresh_cache-c8482d92-0180-44ee-a4f1-bf84786dad43" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1007.802692] env[62066]: DEBUG oslo_concurrency.lockutils [req-ca4a485a-ec0b-4758-bd4c-212af00683cb req-deb2052c-948a-49ff-96ef-df9e830dbc53 service nova] Acquired lock "refresh_cache-c8482d92-0180-44ee-a4f1-bf84786dad43" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1007.802875] env[62066]: DEBUG nova.network.neutron [req-ca4a485a-ec0b-4758-bd4c-212af00683cb req-deb2052c-948a-49ff-96ef-df9e830dbc53 service nova] [instance: c8482d92-0180-44ee-a4f1-bf84786dad43] Refreshing network info cache for port 2f2f1786-934d-478f-be39-7509cfacf86f {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1007.863995] env[62066]: DEBUG nova.compute.manager [req-f7df429c-1e6d-4258-a9e6-9295fb103a8b req-444f028c-dfd3-4a7b-9a68-9d1aa3fdb072 service nova] [instance: 02fb3e92-5dd6-4b1e-a6e4-d60d3fc85b07] Received event network-vif-deleted-16bb647a-353c-4bf8-832d-8f6e49a6ce4b {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1007.926396] env[62066]: INFO nova.compute.manager [-] [instance: 02fb3e92-5dd6-4b1e-a6e4-d60d3fc85b07] Took 1.23 seconds to deallocate network for instance. [ 1008.164023] env[62066]: DEBUG nova.scheduler.client.report [None req-77cb8d13-ef45-4b61-b6f4-a82b4c16e2d2 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1008.305089] env[62066]: DEBUG nova.network.neutron [-] [instance: a2c5277e-af49-4b31-8480-a74d354c7383] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1008.434504] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c80a9ac9-d682-4646-84b4-272b36063bda tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1008.541464] env[62066]: DEBUG nova.network.neutron [req-ca4a485a-ec0b-4758-bd4c-212af00683cb req-deb2052c-948a-49ff-96ef-df9e830dbc53 service nova] [instance: c8482d92-0180-44ee-a4f1-bf84786dad43] Updated VIF entry in instance network info cache for port 2f2f1786-934d-478f-be39-7509cfacf86f. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1008.541464] env[62066]: DEBUG nova.network.neutron [req-ca4a485a-ec0b-4758-bd4c-212af00683cb req-deb2052c-948a-49ff-96ef-df9e830dbc53 service nova] [instance: c8482d92-0180-44ee-a4f1-bf84786dad43] Updating instance_info_cache with network_info: [{"id": "2f2f1786-934d-478f-be39-7509cfacf86f", "address": "fa:16:3e:24:41:2e", "network": {"id": "449dfe4e-bb49-43b6-9ba7-b57af74ebfb3", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-282029676-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.231", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0d4e33a99c7741fb8cdd97f4ec5dbbd8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db68bd64-5b56-49af-a075-13dcf85cb2e0", "external-id": "nsx-vlan-transportzone-590", "segmentation_id": 590, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2f2f1786-93", "ovs_interfaceid": "2f2f1786-934d-478f-be39-7509cfacf86f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1008.667410] env[62066]: DEBUG oslo_concurrency.lockutils [None req-77cb8d13-ef45-4b61-b6f4-a82b4c16e2d2 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.671s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1008.669756] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c56f704c-7907-46df-ae47-acf7d6ca81d1 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.345s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1008.670075] env[62066]: DEBUG nova.objects.instance [None req-c56f704c-7907-46df-ae47-acf7d6ca81d1 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Lazy-loading 'resources' on Instance uuid 9dc7c179-4360-425d-915f-c2d7bc591b1e {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1008.690077] env[62066]: INFO nova.scheduler.client.report [None req-77cb8d13-ef45-4b61-b6f4-a82b4c16e2d2 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Deleted allocations for instance c8e0d47c-4421-4e00-9183-206fceeabc40 [ 1008.810357] env[62066]: INFO nova.compute.manager [-] [instance: a2c5277e-af49-4b31-8480-a74d354c7383] Took 1.26 seconds to deallocate network for instance. [ 1009.045121] env[62066]: DEBUG oslo_concurrency.lockutils [req-ca4a485a-ec0b-4758-bd4c-212af00683cb req-deb2052c-948a-49ff-96ef-df9e830dbc53 service nova] Releasing lock "refresh_cache-c8482d92-0180-44ee-a4f1-bf84786dad43" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1009.045394] env[62066]: DEBUG nova.compute.manager [req-ca4a485a-ec0b-4758-bd4c-212af00683cb req-deb2052c-948a-49ff-96ef-df9e830dbc53 service nova] [instance: 7178eb83-41e7-4516-809d-a0972d7e6e23] Received event network-changed-63f16c96-7de0-40e7-9c0c-782122865437 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1009.045575] env[62066]: DEBUG nova.compute.manager [req-ca4a485a-ec0b-4758-bd4c-212af00683cb req-deb2052c-948a-49ff-96ef-df9e830dbc53 service nova] [instance: 7178eb83-41e7-4516-809d-a0972d7e6e23] Refreshing instance network info cache due to event network-changed-63f16c96-7de0-40e7-9c0c-782122865437. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1009.045825] env[62066]: DEBUG oslo_concurrency.lockutils [req-ca4a485a-ec0b-4758-bd4c-212af00683cb req-deb2052c-948a-49ff-96ef-df9e830dbc53 service nova] Acquiring lock "refresh_cache-7178eb83-41e7-4516-809d-a0972d7e6e23" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1009.046019] env[62066]: DEBUG oslo_concurrency.lockutils [req-ca4a485a-ec0b-4758-bd4c-212af00683cb req-deb2052c-948a-49ff-96ef-df9e830dbc53 service nova] Acquired lock "refresh_cache-7178eb83-41e7-4516-809d-a0972d7e6e23" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1009.046197] env[62066]: DEBUG nova.network.neutron [req-ca4a485a-ec0b-4758-bd4c-212af00683cb req-deb2052c-948a-49ff-96ef-df9e830dbc53 service nova] [instance: 7178eb83-41e7-4516-809d-a0972d7e6e23] Refreshing network info cache for port 63f16c96-7de0-40e7-9c0c-782122865437 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1009.199807] env[62066]: DEBUG oslo_concurrency.lockutils [None req-77cb8d13-ef45-4b61-b6f4-a82b4c16e2d2 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Lock "c8e0d47c-4421-4e00-9183-206fceeabc40" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.095s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1009.235663] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fef9f165-da01-40be-809c-1e3376a6fd9a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.244011] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24df0ae7-1e37-4823-b070-9c41f11c27bd {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.275358] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-196455eb-e4e3-4ef0-85e7-2f2a234b84af {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.283926] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a7e18fc-d95c-4d1a-8fb9-10b9a92a6bc7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.299069] env[62066]: DEBUG nova.compute.provider_tree [None req-c56f704c-7907-46df-ae47-acf7d6ca81d1 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1009.749580] env[62066]: DEBUG nova.network.neutron [req-ca4a485a-ec0b-4758-bd4c-212af00683cb req-deb2052c-948a-49ff-96ef-df9e830dbc53 service nova] [instance: 7178eb83-41e7-4516-809d-a0972d7e6e23] Updated VIF entry in instance network info cache for port 63f16c96-7de0-40e7-9c0c-782122865437. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1009.750038] env[62066]: DEBUG nova.network.neutron [req-ca4a485a-ec0b-4758-bd4c-212af00683cb req-deb2052c-948a-49ff-96ef-df9e830dbc53 service nova] [instance: 7178eb83-41e7-4516-809d-a0972d7e6e23] Updating instance_info_cache with network_info: [{"id": "63f16c96-7de0-40e7-9c0c-782122865437", "address": "fa:16:3e:c6:a7:3e", "network": {"id": "449dfe4e-bb49-43b6-9ba7-b57af74ebfb3", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-282029676-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0d4e33a99c7741fb8cdd97f4ec5dbbd8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db68bd64-5b56-49af-a075-13dcf85cb2e0", "external-id": "nsx-vlan-transportzone-590", "segmentation_id": 590, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap63f16c96-7d", "ovs_interfaceid": "63f16c96-7de0-40e7-9c0c-782122865437", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1009.803910] env[62066]: DEBUG nova.scheduler.client.report [None req-c56f704c-7907-46df-ae47-acf7d6ca81d1 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1009.835567] env[62066]: DEBUG nova.compute.manager [req-7154632a-df35-4c57-bb34-47febeef7eba req-04feb699-acbd-4515-8d1f-3946e66ee452 service nova] [instance: 7178eb83-41e7-4516-809d-a0972d7e6e23] Received event network-changed-63f16c96-7de0-40e7-9c0c-782122865437 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1009.835867] env[62066]: DEBUG nova.compute.manager [req-7154632a-df35-4c57-bb34-47febeef7eba req-04feb699-acbd-4515-8d1f-3946e66ee452 service nova] [instance: 7178eb83-41e7-4516-809d-a0972d7e6e23] Refreshing instance network info cache due to event network-changed-63f16c96-7de0-40e7-9c0c-782122865437. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1009.835947] env[62066]: DEBUG oslo_concurrency.lockutils [req-7154632a-df35-4c57-bb34-47febeef7eba req-04feb699-acbd-4515-8d1f-3946e66ee452 service nova] Acquiring lock "refresh_cache-7178eb83-41e7-4516-809d-a0972d7e6e23" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1009.836410] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4d5077a1-0ba8-4e51-9357-618c8fde121e tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Lock "a2c5277e-af49-4b31-8480-a74d354c7383" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 2.332s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1010.252309] env[62066]: DEBUG oslo_concurrency.lockutils [req-ca4a485a-ec0b-4758-bd4c-212af00683cb req-deb2052c-948a-49ff-96ef-df9e830dbc53 service nova] Releasing lock "refresh_cache-7178eb83-41e7-4516-809d-a0972d7e6e23" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1010.252771] env[62066]: DEBUG oslo_concurrency.lockutils [req-7154632a-df35-4c57-bb34-47febeef7eba req-04feb699-acbd-4515-8d1f-3946e66ee452 service nova] Acquired lock "refresh_cache-7178eb83-41e7-4516-809d-a0972d7e6e23" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1010.252976] env[62066]: DEBUG nova.network.neutron [req-7154632a-df35-4c57-bb34-47febeef7eba req-04feb699-acbd-4515-8d1f-3946e66ee452 service nova] [instance: 7178eb83-41e7-4516-809d-a0972d7e6e23] Refreshing network info cache for port 63f16c96-7de0-40e7-9c0c-782122865437 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1010.306889] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c56f704c-7907-46df-ae47-acf7d6ca81d1 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.637s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1010.309270] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c80a9ac9-d682-4646-84b4-272b36063bda tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.875s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1010.309531] env[62066]: DEBUG nova.objects.instance [None req-c80a9ac9-d682-4646-84b4-272b36063bda tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Lazy-loading 'resources' on Instance uuid 02fb3e92-5dd6-4b1e-a6e4-d60d3fc85b07 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1010.328767] env[62066]: INFO nova.scheduler.client.report [None req-c56f704c-7907-46df-ae47-acf7d6ca81d1 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Deleted allocations for instance 9dc7c179-4360-425d-915f-c2d7bc591b1e [ 1010.844881] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4d1bdf8b-de84-42ec-81b7-e654d596c4f0 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Acquiring lock "06983429-8afe-4bf9-ab65-aa4031568b9c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1010.845509] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4d1bdf8b-de84-42ec-81b7-e654d596c4f0 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Lock "06983429-8afe-4bf9-ab65-aa4031568b9c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1010.849098] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c56f704c-7907-46df-ae47-acf7d6ca81d1 tempest-ServerDiskConfigTestJSON-315316472 tempest-ServerDiskConfigTestJSON-315316472-project-member] Lock "9dc7c179-4360-425d-915f-c2d7bc591b1e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.366s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1010.906041] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3236d694-a14b-4be6-8b2a-9d133174ba6c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.915256] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2be763c2-eecf-4b5f-9792-dbf133b8b986 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.948623] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7a5f2e9-29be-492d-b4fd-1bfe66dd4e44 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.957296] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c39fc7c-3d12-4b09-ac12-fd10831f505f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.970863] env[62066]: DEBUG nova.compute.provider_tree [None req-c80a9ac9-d682-4646-84b4-272b36063bda tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1010.985719] env[62066]: DEBUG nova.network.neutron [req-7154632a-df35-4c57-bb34-47febeef7eba req-04feb699-acbd-4515-8d1f-3946e66ee452 service nova] [instance: 7178eb83-41e7-4516-809d-a0972d7e6e23] Updated VIF entry in instance network info cache for port 63f16c96-7de0-40e7-9c0c-782122865437. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1010.986109] env[62066]: DEBUG nova.network.neutron [req-7154632a-df35-4c57-bb34-47febeef7eba req-04feb699-acbd-4515-8d1f-3946e66ee452 service nova] [instance: 7178eb83-41e7-4516-809d-a0972d7e6e23] Updating instance_info_cache with network_info: [{"id": "63f16c96-7de0-40e7-9c0c-782122865437", "address": "fa:16:3e:c6:a7:3e", "network": {"id": "449dfe4e-bb49-43b6-9ba7-b57af74ebfb3", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-282029676-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0d4e33a99c7741fb8cdd97f4ec5dbbd8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db68bd64-5b56-49af-a075-13dcf85cb2e0", "external-id": "nsx-vlan-transportzone-590", "segmentation_id": 590, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap63f16c96-7d", "ovs_interfaceid": "63f16c96-7de0-40e7-9c0c-782122865437", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1011.351610] env[62066]: DEBUG oslo_concurrency.lockutils [None req-39c438a5-51a5-4a29-a4b4-c8b849e56234 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Acquiring lock "3ef1410d-c78b-47d9-807b-e80368c05902" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1011.351923] env[62066]: DEBUG oslo_concurrency.lockutils [None req-39c438a5-51a5-4a29-a4b4-c8b849e56234 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Lock "3ef1410d-c78b-47d9-807b-e80368c05902" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1011.355369] env[62066]: DEBUG nova.compute.manager [None req-4d1bdf8b-de84-42ec-81b7-e654d596c4f0 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: 06983429-8afe-4bf9-ab65-aa4031568b9c] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1011.474422] env[62066]: DEBUG nova.scheduler.client.report [None req-c80a9ac9-d682-4646-84b4-272b36063bda tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1011.488476] env[62066]: DEBUG oslo_concurrency.lockutils [req-7154632a-df35-4c57-bb34-47febeef7eba req-04feb699-acbd-4515-8d1f-3946e66ee452 service nova] Releasing lock "refresh_cache-7178eb83-41e7-4516-809d-a0972d7e6e23" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1011.488825] env[62066]: DEBUG nova.compute.manager [req-7154632a-df35-4c57-bb34-47febeef7eba req-04feb699-acbd-4515-8d1f-3946e66ee452 service nova] [instance: c8482d92-0180-44ee-a4f1-bf84786dad43] Received event network-changed-2f2f1786-934d-478f-be39-7509cfacf86f {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1011.489085] env[62066]: DEBUG nova.compute.manager [req-7154632a-df35-4c57-bb34-47febeef7eba req-04feb699-acbd-4515-8d1f-3946e66ee452 service nova] [instance: c8482d92-0180-44ee-a4f1-bf84786dad43] Refreshing instance network info cache due to event network-changed-2f2f1786-934d-478f-be39-7509cfacf86f. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1011.489360] env[62066]: DEBUG oslo_concurrency.lockutils [req-7154632a-df35-4c57-bb34-47febeef7eba req-04feb699-acbd-4515-8d1f-3946e66ee452 service nova] Acquiring lock "refresh_cache-c8482d92-0180-44ee-a4f1-bf84786dad43" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1011.489579] env[62066]: DEBUG oslo_concurrency.lockutils [req-7154632a-df35-4c57-bb34-47febeef7eba req-04feb699-acbd-4515-8d1f-3946e66ee452 service nova] Acquired lock "refresh_cache-c8482d92-0180-44ee-a4f1-bf84786dad43" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1011.490097] env[62066]: DEBUG nova.network.neutron [req-7154632a-df35-4c57-bb34-47febeef7eba req-04feb699-acbd-4515-8d1f-3946e66ee452 service nova] [instance: c8482d92-0180-44ee-a4f1-bf84786dad43] Refreshing network info cache for port 2f2f1786-934d-478f-be39-7509cfacf86f {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1011.856603] env[62066]: DEBUG nova.compute.manager [None req-39c438a5-51a5-4a29-a4b4-c8b849e56234 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: 3ef1410d-c78b-47d9-807b-e80368c05902] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1011.881464] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4d1bdf8b-de84-42ec-81b7-e654d596c4f0 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1011.979631] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c80a9ac9-d682-4646-84b4-272b36063bda tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.670s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1011.982286] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4d1bdf8b-de84-42ec-81b7-e654d596c4f0 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.101s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1011.983661] env[62066]: INFO nova.compute.claims [None req-4d1bdf8b-de84-42ec-81b7-e654d596c4f0 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: 06983429-8afe-4bf9-ab65-aa4031568b9c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1012.011013] env[62066]: INFO nova.scheduler.client.report [None req-c80a9ac9-d682-4646-84b4-272b36063bda tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Deleted allocations for instance 02fb3e92-5dd6-4b1e-a6e4-d60d3fc85b07 [ 1012.229033] env[62066]: DEBUG nova.network.neutron [req-7154632a-df35-4c57-bb34-47febeef7eba req-04feb699-acbd-4515-8d1f-3946e66ee452 service nova] [instance: c8482d92-0180-44ee-a4f1-bf84786dad43] Updated VIF entry in instance network info cache for port 2f2f1786-934d-478f-be39-7509cfacf86f. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1012.229463] env[62066]: DEBUG nova.network.neutron [req-7154632a-df35-4c57-bb34-47febeef7eba req-04feb699-acbd-4515-8d1f-3946e66ee452 service nova] [instance: c8482d92-0180-44ee-a4f1-bf84786dad43] Updating instance_info_cache with network_info: [{"id": "2f2f1786-934d-478f-be39-7509cfacf86f", "address": "fa:16:3e:24:41:2e", "network": {"id": "449dfe4e-bb49-43b6-9ba7-b57af74ebfb3", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-282029676-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.231", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0d4e33a99c7741fb8cdd97f4ec5dbbd8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db68bd64-5b56-49af-a075-13dcf85cb2e0", "external-id": "nsx-vlan-transportzone-590", "segmentation_id": 590, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2f2f1786-93", "ovs_interfaceid": "2f2f1786-934d-478f-be39-7509cfacf86f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1012.388478] env[62066]: DEBUG oslo_concurrency.lockutils [None req-39c438a5-51a5-4a29-a4b4-c8b849e56234 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1012.519927] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c80a9ac9-d682-4646-84b4-272b36063bda tempest-ServersTestJSON-363593134 tempest-ServersTestJSON-363593134-project-member] Lock "02fb3e92-5dd6-4b1e-a6e4-d60d3fc85b07" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.946s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1012.732365] env[62066]: DEBUG oslo_concurrency.lockutils [req-7154632a-df35-4c57-bb34-47febeef7eba req-04feb699-acbd-4515-8d1f-3946e66ee452 service nova] Releasing lock "refresh_cache-c8482d92-0180-44ee-a4f1-bf84786dad43" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1013.055842] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a72fdaad-f30c-46eb-be2e-ac4a6874d8d8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.064057] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1b62bbb-255d-43e7-a2d2-a63cb4902b4e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.096215] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cd01868-3da3-49f4-ba95-fa9f95087ce0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.104839] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d76b24b6-4b49-4d0a-9be6-0ae08b1af769 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.118429] env[62066]: DEBUG nova.compute.provider_tree [None req-4d1bdf8b-de84-42ec-81b7-e654d596c4f0 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1013.622174] env[62066]: DEBUG nova.scheduler.client.report [None req-4d1bdf8b-de84-42ec-81b7-e654d596c4f0 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1014.127913] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4d1bdf8b-de84-42ec-81b7-e654d596c4f0 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.146s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1014.128441] env[62066]: DEBUG nova.compute.manager [None req-4d1bdf8b-de84-42ec-81b7-e654d596c4f0 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: 06983429-8afe-4bf9-ab65-aa4031568b9c] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1014.131376] env[62066]: DEBUG oslo_concurrency.lockutils [None req-39c438a5-51a5-4a29-a4b4-c8b849e56234 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.743s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1014.132814] env[62066]: INFO nova.compute.claims [None req-39c438a5-51a5-4a29-a4b4-c8b849e56234 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: 3ef1410d-c78b-47d9-807b-e80368c05902] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1014.640128] env[62066]: DEBUG nova.compute.utils [None req-4d1bdf8b-de84-42ec-81b7-e654d596c4f0 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1014.641648] env[62066]: DEBUG nova.compute.manager [None req-4d1bdf8b-de84-42ec-81b7-e654d596c4f0 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: 06983429-8afe-4bf9-ab65-aa4031568b9c] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1014.641967] env[62066]: DEBUG nova.network.neutron [None req-4d1bdf8b-de84-42ec-81b7-e654d596c4f0 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: 06983429-8afe-4bf9-ab65-aa4031568b9c] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1014.700873] env[62066]: DEBUG nova.policy [None req-4d1bdf8b-de84-42ec-81b7-e654d596c4f0 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '588e53cee85f4ab484b76e7a59fcbe78', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8475ad5a900548cba568360999c846ea', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 1014.940209] env[62066]: DEBUG nova.network.neutron [None req-4d1bdf8b-de84-42ec-81b7-e654d596c4f0 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: 06983429-8afe-4bf9-ab65-aa4031568b9c] Successfully created port: 1be70afc-e68a-481b-909e-9bd764d93b21 {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1015.145894] env[62066]: DEBUG nova.compute.manager [None req-4d1bdf8b-de84-42ec-81b7-e654d596c4f0 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: 06983429-8afe-4bf9-ab65-aa4031568b9c] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1015.211841] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a993e3e1-81b0-4d89-91eb-29c5791ac028 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.219723] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70c6f2e1-c17e-45b5-a039-2a566db3061d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.250022] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27bacc83-4ad4-4078-bbdc-5c1ea8dcd5a2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.258294] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a595966a-2727-4b6e-865a-a5eefc6b99d1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.271499] env[62066]: DEBUG nova.compute.provider_tree [None req-39c438a5-51a5-4a29-a4b4-c8b849e56234 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1015.775137] env[62066]: DEBUG nova.scheduler.client.report [None req-39c438a5-51a5-4a29-a4b4-c8b849e56234 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1016.156100] env[62066]: DEBUG nova.compute.manager [None req-4d1bdf8b-de84-42ec-81b7-e654d596c4f0 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: 06983429-8afe-4bf9-ab65-aa4031568b9c] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1016.180575] env[62066]: DEBUG nova.virt.hardware [None req-4d1bdf8b-de84-42ec-81b7-e654d596c4f0 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-23T13:40:41Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-23T13:40:23Z,direct_url=,disk_format='vmdk',id=50ff584c-3b50-4395-af07-3e66769bc9f7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='eb52f7069a374c61ae946f052007c6d9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-23T13:40:24Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1016.180956] env[62066]: DEBUG nova.virt.hardware [None req-4d1bdf8b-de84-42ec-81b7-e654d596c4f0 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1016.181199] env[62066]: DEBUG nova.virt.hardware [None req-4d1bdf8b-de84-42ec-81b7-e654d596c4f0 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1016.181408] env[62066]: DEBUG nova.virt.hardware [None req-4d1bdf8b-de84-42ec-81b7-e654d596c4f0 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1016.181565] env[62066]: DEBUG nova.virt.hardware [None req-4d1bdf8b-de84-42ec-81b7-e654d596c4f0 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1016.181777] env[62066]: DEBUG nova.virt.hardware [None req-4d1bdf8b-de84-42ec-81b7-e654d596c4f0 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1016.182050] env[62066]: DEBUG nova.virt.hardware [None req-4d1bdf8b-de84-42ec-81b7-e654d596c4f0 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1016.182231] env[62066]: DEBUG nova.virt.hardware [None req-4d1bdf8b-de84-42ec-81b7-e654d596c4f0 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1016.182413] env[62066]: DEBUG nova.virt.hardware [None req-4d1bdf8b-de84-42ec-81b7-e654d596c4f0 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1016.182585] env[62066]: DEBUG nova.virt.hardware [None req-4d1bdf8b-de84-42ec-81b7-e654d596c4f0 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1016.182764] env[62066]: DEBUG nova.virt.hardware [None req-4d1bdf8b-de84-42ec-81b7-e654d596c4f0 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1016.183654] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fd3ab7c-1301-4b55-8413-e03b3c72a9b5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.211039] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e5aa496-6573-4542-bbbd-575f6d013db7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.280261] env[62066]: DEBUG oslo_concurrency.lockutils [None req-39c438a5-51a5-4a29-a4b4-c8b849e56234 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.149s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1016.280891] env[62066]: DEBUG nova.compute.manager [None req-39c438a5-51a5-4a29-a4b4-c8b849e56234 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: 3ef1410d-c78b-47d9-807b-e80368c05902] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1016.345388] env[62066]: DEBUG nova.compute.manager [req-28482c10-b405-42da-8bdc-10777bc5c500 req-4536127d-f965-4859-a280-5ab6309f1b47 service nova] [instance: 06983429-8afe-4bf9-ab65-aa4031568b9c] Received event network-vif-plugged-1be70afc-e68a-481b-909e-9bd764d93b21 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1016.345625] env[62066]: DEBUG oslo_concurrency.lockutils [req-28482c10-b405-42da-8bdc-10777bc5c500 req-4536127d-f965-4859-a280-5ab6309f1b47 service nova] Acquiring lock "06983429-8afe-4bf9-ab65-aa4031568b9c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1016.345820] env[62066]: DEBUG oslo_concurrency.lockutils [req-28482c10-b405-42da-8bdc-10777bc5c500 req-4536127d-f965-4859-a280-5ab6309f1b47 service nova] Lock "06983429-8afe-4bf9-ab65-aa4031568b9c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1016.345994] env[62066]: DEBUG oslo_concurrency.lockutils [req-28482c10-b405-42da-8bdc-10777bc5c500 req-4536127d-f965-4859-a280-5ab6309f1b47 service nova] Lock "06983429-8afe-4bf9-ab65-aa4031568b9c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1016.346189] env[62066]: DEBUG nova.compute.manager [req-28482c10-b405-42da-8bdc-10777bc5c500 req-4536127d-f965-4859-a280-5ab6309f1b47 service nova] [instance: 06983429-8afe-4bf9-ab65-aa4031568b9c] No waiting events found dispatching network-vif-plugged-1be70afc-e68a-481b-909e-9bd764d93b21 {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1016.346366] env[62066]: WARNING nova.compute.manager [req-28482c10-b405-42da-8bdc-10777bc5c500 req-4536127d-f965-4859-a280-5ab6309f1b47 service nova] [instance: 06983429-8afe-4bf9-ab65-aa4031568b9c] Received unexpected event network-vif-plugged-1be70afc-e68a-481b-909e-9bd764d93b21 for instance with vm_state building and task_state spawning. [ 1016.439749] env[62066]: DEBUG nova.network.neutron [None req-4d1bdf8b-de84-42ec-81b7-e654d596c4f0 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: 06983429-8afe-4bf9-ab65-aa4031568b9c] Successfully updated port: 1be70afc-e68a-481b-909e-9bd764d93b21 {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1016.786280] env[62066]: DEBUG nova.compute.utils [None req-39c438a5-51a5-4a29-a4b4-c8b849e56234 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1016.787836] env[62066]: DEBUG nova.compute.manager [None req-39c438a5-51a5-4a29-a4b4-c8b849e56234 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: 3ef1410d-c78b-47d9-807b-e80368c05902] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1016.788071] env[62066]: DEBUG nova.network.neutron [None req-39c438a5-51a5-4a29-a4b4-c8b849e56234 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: 3ef1410d-c78b-47d9-807b-e80368c05902] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1016.825057] env[62066]: DEBUG nova.policy [None req-39c438a5-51a5-4a29-a4b4-c8b849e56234 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd0d27da4c6fe4c9eb6e2a6e45c231a38', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2252d0f6f18a47fd8a1ddad8020a6db7', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 1016.943563] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4d1bdf8b-de84-42ec-81b7-e654d596c4f0 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Acquiring lock "refresh_cache-06983429-8afe-4bf9-ab65-aa4031568b9c" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1016.943720] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4d1bdf8b-de84-42ec-81b7-e654d596c4f0 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Acquired lock "refresh_cache-06983429-8afe-4bf9-ab65-aa4031568b9c" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1016.943876] env[62066]: DEBUG nova.network.neutron [None req-4d1bdf8b-de84-42ec-81b7-e654d596c4f0 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: 06983429-8afe-4bf9-ab65-aa4031568b9c] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1017.105748] env[62066]: DEBUG nova.network.neutron [None req-39c438a5-51a5-4a29-a4b4-c8b849e56234 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: 3ef1410d-c78b-47d9-807b-e80368c05902] Successfully created port: 3b1d21b6-842c-4761-867d-8b06174242db {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1017.290965] env[62066]: DEBUG nova.compute.manager [None req-39c438a5-51a5-4a29-a4b4-c8b849e56234 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: 3ef1410d-c78b-47d9-807b-e80368c05902] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1017.478251] env[62066]: DEBUG nova.network.neutron [None req-4d1bdf8b-de84-42ec-81b7-e654d596c4f0 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: 06983429-8afe-4bf9-ab65-aa4031568b9c] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1017.594770] env[62066]: DEBUG nova.network.neutron [None req-4d1bdf8b-de84-42ec-81b7-e654d596c4f0 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: 06983429-8afe-4bf9-ab65-aa4031568b9c] Updating instance_info_cache with network_info: [{"id": "1be70afc-e68a-481b-909e-9bd764d93b21", "address": "fa:16:3e:c4:ca:90", "network": {"id": "7617f887-bdd0-400a-8b5f-606d7785df78", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-558766873-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8475ad5a900548cba568360999c846ea", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2b6a4065-12af-4fb9-ac47-ec9143f7297e", "external-id": "nsx-vlan-transportzone-95", "segmentation_id": 95, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1be70afc-e6", "ovs_interfaceid": "1be70afc-e68a-481b-909e-9bd764d93b21", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1018.097591] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4d1bdf8b-de84-42ec-81b7-e654d596c4f0 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Releasing lock "refresh_cache-06983429-8afe-4bf9-ab65-aa4031568b9c" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1018.097921] env[62066]: DEBUG nova.compute.manager [None req-4d1bdf8b-de84-42ec-81b7-e654d596c4f0 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: 06983429-8afe-4bf9-ab65-aa4031568b9c] Instance network_info: |[{"id": "1be70afc-e68a-481b-909e-9bd764d93b21", "address": "fa:16:3e:c4:ca:90", "network": {"id": "7617f887-bdd0-400a-8b5f-606d7785df78", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-558766873-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8475ad5a900548cba568360999c846ea", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2b6a4065-12af-4fb9-ac47-ec9143f7297e", "external-id": "nsx-vlan-transportzone-95", "segmentation_id": 95, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1be70afc-e6", "ovs_interfaceid": "1be70afc-e68a-481b-909e-9bd764d93b21", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1018.098367] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-4d1bdf8b-de84-42ec-81b7-e654d596c4f0 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: 06983429-8afe-4bf9-ab65-aa4031568b9c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c4:ca:90', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '2b6a4065-12af-4fb9-ac47-ec9143f7297e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1be70afc-e68a-481b-909e-9bd764d93b21', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1018.106209] env[62066]: DEBUG oslo.service.loopingcall [None req-4d1bdf8b-de84-42ec-81b7-e654d596c4f0 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1018.106416] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 06983429-8afe-4bf9-ab65-aa4031568b9c] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1018.106637] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-dcfcf564-a243-4deb-8d85-f2ecac6bf0eb {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.127129] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1018.127129] env[62066]: value = "task-1341324" [ 1018.127129] env[62066]: _type = "Task" [ 1018.127129] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.134915] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1341324, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.300680] env[62066]: DEBUG nova.compute.manager [None req-39c438a5-51a5-4a29-a4b4-c8b849e56234 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: 3ef1410d-c78b-47d9-807b-e80368c05902] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1018.327104] env[62066]: DEBUG nova.virt.hardware [None req-39c438a5-51a5-4a29-a4b4-c8b849e56234 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-23T13:40:41Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-23T13:40:23Z,direct_url=,disk_format='vmdk',id=50ff584c-3b50-4395-af07-3e66769bc9f7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='eb52f7069a374c61ae946f052007c6d9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-23T13:40:24Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1018.327369] env[62066]: DEBUG nova.virt.hardware [None req-39c438a5-51a5-4a29-a4b4-c8b849e56234 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1018.327529] env[62066]: DEBUG nova.virt.hardware [None req-39c438a5-51a5-4a29-a4b4-c8b849e56234 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1018.327717] env[62066]: DEBUG nova.virt.hardware [None req-39c438a5-51a5-4a29-a4b4-c8b849e56234 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1018.327870] env[62066]: DEBUG nova.virt.hardware [None req-39c438a5-51a5-4a29-a4b4-c8b849e56234 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1018.328041] env[62066]: DEBUG nova.virt.hardware [None req-39c438a5-51a5-4a29-a4b4-c8b849e56234 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1018.328272] env[62066]: DEBUG nova.virt.hardware [None req-39c438a5-51a5-4a29-a4b4-c8b849e56234 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1018.328438] env[62066]: DEBUG nova.virt.hardware [None req-39c438a5-51a5-4a29-a4b4-c8b849e56234 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1018.328615] env[62066]: DEBUG nova.virt.hardware [None req-39c438a5-51a5-4a29-a4b4-c8b849e56234 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1018.328831] env[62066]: DEBUG nova.virt.hardware [None req-39c438a5-51a5-4a29-a4b4-c8b849e56234 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1018.329016] env[62066]: DEBUG nova.virt.hardware [None req-39c438a5-51a5-4a29-a4b4-c8b849e56234 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1018.329896] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a61e92b2-1131-48aa-b7c3-dca54a615fa6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.338412] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ba6441b-06c3-4622-b294-233cf317b4d8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.371016] env[62066]: DEBUG nova.compute.manager [req-5574a204-e23b-4a5e-b420-ceb3e084f62d req-a35469cd-19ca-4757-a004-119d69ef429a service nova] [instance: 06983429-8afe-4bf9-ab65-aa4031568b9c] Received event network-changed-1be70afc-e68a-481b-909e-9bd764d93b21 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1018.371016] env[62066]: DEBUG nova.compute.manager [req-5574a204-e23b-4a5e-b420-ceb3e084f62d req-a35469cd-19ca-4757-a004-119d69ef429a service nova] [instance: 06983429-8afe-4bf9-ab65-aa4031568b9c] Refreshing instance network info cache due to event network-changed-1be70afc-e68a-481b-909e-9bd764d93b21. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1018.371221] env[62066]: DEBUG oslo_concurrency.lockutils [req-5574a204-e23b-4a5e-b420-ceb3e084f62d req-a35469cd-19ca-4757-a004-119d69ef429a service nova] Acquiring lock "refresh_cache-06983429-8afe-4bf9-ab65-aa4031568b9c" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1018.371364] env[62066]: DEBUG oslo_concurrency.lockutils [req-5574a204-e23b-4a5e-b420-ceb3e084f62d req-a35469cd-19ca-4757-a004-119d69ef429a service nova] Acquired lock "refresh_cache-06983429-8afe-4bf9-ab65-aa4031568b9c" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1018.371521] env[62066]: DEBUG nova.network.neutron [req-5574a204-e23b-4a5e-b420-ceb3e084f62d req-a35469cd-19ca-4757-a004-119d69ef429a service nova] [instance: 06983429-8afe-4bf9-ab65-aa4031568b9c] Refreshing network info cache for port 1be70afc-e68a-481b-909e-9bd764d93b21 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1018.620431] env[62066]: DEBUG nova.network.neutron [None req-39c438a5-51a5-4a29-a4b4-c8b849e56234 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: 3ef1410d-c78b-47d9-807b-e80368c05902] Successfully updated port: 3b1d21b6-842c-4761-867d-8b06174242db {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1018.638337] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1341324, 'name': CreateVM_Task, 'duration_secs': 0.299995} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.638534] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 06983429-8afe-4bf9-ab65-aa4031568b9c] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1018.639216] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4d1bdf8b-de84-42ec-81b7-e654d596c4f0 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1018.639386] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4d1bdf8b-de84-42ec-81b7-e654d596c4f0 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Acquired lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1018.639716] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4d1bdf8b-de84-42ec-81b7-e654d596c4f0 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1018.639982] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-64a79cfd-d284-4020-998f-d4082450084a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.644662] env[62066]: DEBUG oslo_vmware.api [None req-4d1bdf8b-de84-42ec-81b7-e654d596c4f0 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Waiting for the task: (returnval){ [ 1018.644662] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]52668b0b-de47-c2e5-4965-d1b72f97baf0" [ 1018.644662] env[62066]: _type = "Task" [ 1018.644662] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.652678] env[62066]: DEBUG oslo_vmware.api [None req-4d1bdf8b-de84-42ec-81b7-e654d596c4f0 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52668b0b-de47-c2e5-4965-d1b72f97baf0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.054441] env[62066]: DEBUG nova.network.neutron [req-5574a204-e23b-4a5e-b420-ceb3e084f62d req-a35469cd-19ca-4757-a004-119d69ef429a service nova] [instance: 06983429-8afe-4bf9-ab65-aa4031568b9c] Updated VIF entry in instance network info cache for port 1be70afc-e68a-481b-909e-9bd764d93b21. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1019.054809] env[62066]: DEBUG nova.network.neutron [req-5574a204-e23b-4a5e-b420-ceb3e084f62d req-a35469cd-19ca-4757-a004-119d69ef429a service nova] [instance: 06983429-8afe-4bf9-ab65-aa4031568b9c] Updating instance_info_cache with network_info: [{"id": "1be70afc-e68a-481b-909e-9bd764d93b21", "address": "fa:16:3e:c4:ca:90", "network": {"id": "7617f887-bdd0-400a-8b5f-606d7785df78", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-558766873-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8475ad5a900548cba568360999c846ea", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2b6a4065-12af-4fb9-ac47-ec9143f7297e", "external-id": "nsx-vlan-transportzone-95", "segmentation_id": 95, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1be70afc-e6", "ovs_interfaceid": "1be70afc-e68a-481b-909e-9bd764d93b21", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1019.124502] env[62066]: DEBUG oslo_concurrency.lockutils [None req-39c438a5-51a5-4a29-a4b4-c8b849e56234 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Acquiring lock "refresh_cache-3ef1410d-c78b-47d9-807b-e80368c05902" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1019.124502] env[62066]: DEBUG oslo_concurrency.lockutils [None req-39c438a5-51a5-4a29-a4b4-c8b849e56234 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Acquired lock "refresh_cache-3ef1410d-c78b-47d9-807b-e80368c05902" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1019.124502] env[62066]: DEBUG nova.network.neutron [None req-39c438a5-51a5-4a29-a4b4-c8b849e56234 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: 3ef1410d-c78b-47d9-807b-e80368c05902] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1019.156939] env[62066]: DEBUG oslo_vmware.api [None req-4d1bdf8b-de84-42ec-81b7-e654d596c4f0 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52668b0b-de47-c2e5-4965-d1b72f97baf0, 'name': SearchDatastore_Task, 'duration_secs': 0.009882} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.157254] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4d1bdf8b-de84-42ec-81b7-e654d596c4f0 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Releasing lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1019.157487] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-4d1bdf8b-de84-42ec-81b7-e654d596c4f0 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: 06983429-8afe-4bf9-ab65-aa4031568b9c] Processing image 50ff584c-3b50-4395-af07-3e66769bc9f7 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1019.157720] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4d1bdf8b-de84-42ec-81b7-e654d596c4f0 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1019.157874] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4d1bdf8b-de84-42ec-81b7-e654d596c4f0 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Acquired lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1019.158075] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-4d1bdf8b-de84-42ec-81b7-e654d596c4f0 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1019.158337] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0fcf5aff-49ff-44be-88d1-a822703c2b55 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.166777] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-4d1bdf8b-de84-42ec-81b7-e654d596c4f0 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1019.166777] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-4d1bdf8b-de84-42ec-81b7-e654d596c4f0 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1019.167336] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8163bbe7-bc4e-4d0a-8cf1-2c97637388dd {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.172106] env[62066]: DEBUG oslo_vmware.api [None req-4d1bdf8b-de84-42ec-81b7-e654d596c4f0 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Waiting for the task: (returnval){ [ 1019.172106] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]52a63b1e-9ffe-4793-c88d-1a8f16e68cc3" [ 1019.172106] env[62066]: _type = "Task" [ 1019.172106] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.179492] env[62066]: DEBUG oslo_vmware.api [None req-4d1bdf8b-de84-42ec-81b7-e654d596c4f0 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52a63b1e-9ffe-4793-c88d-1a8f16e68cc3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.557706] env[62066]: DEBUG oslo_concurrency.lockutils [req-5574a204-e23b-4a5e-b420-ceb3e084f62d req-a35469cd-19ca-4757-a004-119d69ef429a service nova] Releasing lock "refresh_cache-06983429-8afe-4bf9-ab65-aa4031568b9c" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1019.653647] env[62066]: DEBUG nova.network.neutron [None req-39c438a5-51a5-4a29-a4b4-c8b849e56234 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: 3ef1410d-c78b-47d9-807b-e80368c05902] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1019.683043] env[62066]: DEBUG oslo_vmware.api [None req-4d1bdf8b-de84-42ec-81b7-e654d596c4f0 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52a63b1e-9ffe-4793-c88d-1a8f16e68cc3, 'name': SearchDatastore_Task, 'duration_secs': 0.007885} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.683632] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d1d673d5-9ecf-4e61-86ae-6f92cc91519e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.690510] env[62066]: DEBUG oslo_vmware.api [None req-4d1bdf8b-de84-42ec-81b7-e654d596c4f0 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Waiting for the task: (returnval){ [ 1019.690510] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]527e165c-3362-753d-52a3-5432ce7dcad6" [ 1019.690510] env[62066]: _type = "Task" [ 1019.690510] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.698239] env[62066]: DEBUG oslo_vmware.api [None req-4d1bdf8b-de84-42ec-81b7-e654d596c4f0 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]527e165c-3362-753d-52a3-5432ce7dcad6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.775206] env[62066]: DEBUG nova.network.neutron [None req-39c438a5-51a5-4a29-a4b4-c8b849e56234 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: 3ef1410d-c78b-47d9-807b-e80368c05902] Updating instance_info_cache with network_info: [{"id": "3b1d21b6-842c-4761-867d-8b06174242db", "address": "fa:16:3e:67:4b:4e", "network": {"id": "5b89b2d4-71dd-4cd1-9edb-f4040be26540", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-283918837-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2252d0f6f18a47fd8a1ddad8020a6db7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "352165bb-004f-4180-9627-3a275dbe18af", "external-id": "nsx-vlan-transportzone-926", "segmentation_id": 926, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3b1d21b6-84", "ovs_interfaceid": "3b1d21b6-842c-4761-867d-8b06174242db", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1020.201743] env[62066]: DEBUG oslo_vmware.api [None req-4d1bdf8b-de84-42ec-81b7-e654d596c4f0 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]527e165c-3362-753d-52a3-5432ce7dcad6, 'name': SearchDatastore_Task, 'duration_secs': 0.009317} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.202030] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4d1bdf8b-de84-42ec-81b7-e654d596c4f0 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Releasing lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1020.202304] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d1bdf8b-de84-42ec-81b7-e654d596c4f0 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk to [datastore2] 06983429-8afe-4bf9-ab65-aa4031568b9c/06983429-8afe-4bf9-ab65-aa4031568b9c.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1020.202570] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f1369fca-2151-4f99-8be2-4d4153c20bde {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.209790] env[62066]: DEBUG oslo_vmware.api [None req-4d1bdf8b-de84-42ec-81b7-e654d596c4f0 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Waiting for the task: (returnval){ [ 1020.209790] env[62066]: value = "task-1341325" [ 1020.209790] env[62066]: _type = "Task" [ 1020.209790] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.217357] env[62066]: DEBUG oslo_vmware.api [None req-4d1bdf8b-de84-42ec-81b7-e654d596c4f0 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341325, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.278136] env[62066]: DEBUG oslo_concurrency.lockutils [None req-39c438a5-51a5-4a29-a4b4-c8b849e56234 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Releasing lock "refresh_cache-3ef1410d-c78b-47d9-807b-e80368c05902" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1020.278469] env[62066]: DEBUG nova.compute.manager [None req-39c438a5-51a5-4a29-a4b4-c8b849e56234 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: 3ef1410d-c78b-47d9-807b-e80368c05902] Instance network_info: |[{"id": "3b1d21b6-842c-4761-867d-8b06174242db", "address": "fa:16:3e:67:4b:4e", "network": {"id": "5b89b2d4-71dd-4cd1-9edb-f4040be26540", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-283918837-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2252d0f6f18a47fd8a1ddad8020a6db7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "352165bb-004f-4180-9627-3a275dbe18af", "external-id": "nsx-vlan-transportzone-926", "segmentation_id": 926, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3b1d21b6-84", "ovs_interfaceid": "3b1d21b6-842c-4761-867d-8b06174242db", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1020.278978] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-39c438a5-51a5-4a29-a4b4-c8b849e56234 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: 3ef1410d-c78b-47d9-807b-e80368c05902] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:67:4b:4e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '352165bb-004f-4180-9627-3a275dbe18af', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3b1d21b6-842c-4761-867d-8b06174242db', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1020.287485] env[62066]: DEBUG oslo.service.loopingcall [None req-39c438a5-51a5-4a29-a4b4-c8b849e56234 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1020.287805] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3ef1410d-c78b-47d9-807b-e80368c05902] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1020.288115] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ccdbfecb-3f7b-4e5b-b9ac-6f8c93f96356 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.307962] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1020.307962] env[62066]: value = "task-1341326" [ 1020.307962] env[62066]: _type = "Task" [ 1020.307962] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.315422] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1341326, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.398961] env[62066]: DEBUG nova.compute.manager [req-838fe49f-74e6-4d22-aeba-19836242b19a req-75c641dd-caf3-4273-b653-656d52cf8733 service nova] [instance: 3ef1410d-c78b-47d9-807b-e80368c05902] Received event network-vif-plugged-3b1d21b6-842c-4761-867d-8b06174242db {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1020.399329] env[62066]: DEBUG oslo_concurrency.lockutils [req-838fe49f-74e6-4d22-aeba-19836242b19a req-75c641dd-caf3-4273-b653-656d52cf8733 service nova] Acquiring lock "3ef1410d-c78b-47d9-807b-e80368c05902-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1020.399630] env[62066]: DEBUG oslo_concurrency.lockutils [req-838fe49f-74e6-4d22-aeba-19836242b19a req-75c641dd-caf3-4273-b653-656d52cf8733 service nova] Lock "3ef1410d-c78b-47d9-807b-e80368c05902-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1020.399790] env[62066]: DEBUG oslo_concurrency.lockutils [req-838fe49f-74e6-4d22-aeba-19836242b19a req-75c641dd-caf3-4273-b653-656d52cf8733 service nova] Lock "3ef1410d-c78b-47d9-807b-e80368c05902-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1020.400014] env[62066]: DEBUG nova.compute.manager [req-838fe49f-74e6-4d22-aeba-19836242b19a req-75c641dd-caf3-4273-b653-656d52cf8733 service nova] [instance: 3ef1410d-c78b-47d9-807b-e80368c05902] No waiting events found dispatching network-vif-plugged-3b1d21b6-842c-4761-867d-8b06174242db {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1020.400230] env[62066]: WARNING nova.compute.manager [req-838fe49f-74e6-4d22-aeba-19836242b19a req-75c641dd-caf3-4273-b653-656d52cf8733 service nova] [instance: 3ef1410d-c78b-47d9-807b-e80368c05902] Received unexpected event network-vif-plugged-3b1d21b6-842c-4761-867d-8b06174242db for instance with vm_state building and task_state spawning. [ 1020.400437] env[62066]: DEBUG nova.compute.manager [req-838fe49f-74e6-4d22-aeba-19836242b19a req-75c641dd-caf3-4273-b653-656d52cf8733 service nova] [instance: 3ef1410d-c78b-47d9-807b-e80368c05902] Received event network-changed-3b1d21b6-842c-4761-867d-8b06174242db {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1020.400647] env[62066]: DEBUG nova.compute.manager [req-838fe49f-74e6-4d22-aeba-19836242b19a req-75c641dd-caf3-4273-b653-656d52cf8733 service nova] [instance: 3ef1410d-c78b-47d9-807b-e80368c05902] Refreshing instance network info cache due to event network-changed-3b1d21b6-842c-4761-867d-8b06174242db. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1020.400881] env[62066]: DEBUG oslo_concurrency.lockutils [req-838fe49f-74e6-4d22-aeba-19836242b19a req-75c641dd-caf3-4273-b653-656d52cf8733 service nova] Acquiring lock "refresh_cache-3ef1410d-c78b-47d9-807b-e80368c05902" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1020.401042] env[62066]: DEBUG oslo_concurrency.lockutils [req-838fe49f-74e6-4d22-aeba-19836242b19a req-75c641dd-caf3-4273-b653-656d52cf8733 service nova] Acquired lock "refresh_cache-3ef1410d-c78b-47d9-807b-e80368c05902" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1020.401258] env[62066]: DEBUG nova.network.neutron [req-838fe49f-74e6-4d22-aeba-19836242b19a req-75c641dd-caf3-4273-b653-656d52cf8733 service nova] [instance: 3ef1410d-c78b-47d9-807b-e80368c05902] Refreshing network info cache for port 3b1d21b6-842c-4761-867d-8b06174242db {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1020.720784] env[62066]: DEBUG oslo_vmware.api [None req-4d1bdf8b-de84-42ec-81b7-e654d596c4f0 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341325, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.469696} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.721153] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d1bdf8b-de84-42ec-81b7-e654d596c4f0 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk to [datastore2] 06983429-8afe-4bf9-ab65-aa4031568b9c/06983429-8afe-4bf9-ab65-aa4031568b9c.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1020.721298] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-4d1bdf8b-de84-42ec-81b7-e654d596c4f0 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: 06983429-8afe-4bf9-ab65-aa4031568b9c] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1020.721574] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c19efb26-f6d8-4d64-9be8-7d46a2e06b7c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.729784] env[62066]: DEBUG oslo_vmware.api [None req-4d1bdf8b-de84-42ec-81b7-e654d596c4f0 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Waiting for the task: (returnval){ [ 1020.729784] env[62066]: value = "task-1341327" [ 1020.729784] env[62066]: _type = "Task" [ 1020.729784] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.738197] env[62066]: DEBUG oslo_vmware.api [None req-4d1bdf8b-de84-42ec-81b7-e654d596c4f0 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341327, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.817531] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1341326, 'name': CreateVM_Task} progress is 25%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.087400] env[62066]: DEBUG nova.network.neutron [req-838fe49f-74e6-4d22-aeba-19836242b19a req-75c641dd-caf3-4273-b653-656d52cf8733 service nova] [instance: 3ef1410d-c78b-47d9-807b-e80368c05902] Updated VIF entry in instance network info cache for port 3b1d21b6-842c-4761-867d-8b06174242db. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1021.087758] env[62066]: DEBUG nova.network.neutron [req-838fe49f-74e6-4d22-aeba-19836242b19a req-75c641dd-caf3-4273-b653-656d52cf8733 service nova] [instance: 3ef1410d-c78b-47d9-807b-e80368c05902] Updating instance_info_cache with network_info: [{"id": "3b1d21b6-842c-4761-867d-8b06174242db", "address": "fa:16:3e:67:4b:4e", "network": {"id": "5b89b2d4-71dd-4cd1-9edb-f4040be26540", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-283918837-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2252d0f6f18a47fd8a1ddad8020a6db7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "352165bb-004f-4180-9627-3a275dbe18af", "external-id": "nsx-vlan-transportzone-926", "segmentation_id": 926, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3b1d21b6-84", "ovs_interfaceid": "3b1d21b6-842c-4761-867d-8b06174242db", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1021.245031] env[62066]: DEBUG oslo_vmware.api [None req-4d1bdf8b-de84-42ec-81b7-e654d596c4f0 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341327, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.059873} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.245469] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-4d1bdf8b-de84-42ec-81b7-e654d596c4f0 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: 06983429-8afe-4bf9-ab65-aa4031568b9c] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1021.246617] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a8c36f1-e7a5-4c5b-abdd-afa9775569bf {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.273864] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-4d1bdf8b-de84-42ec-81b7-e654d596c4f0 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: 06983429-8afe-4bf9-ab65-aa4031568b9c] Reconfiguring VM instance instance-00000066 to attach disk [datastore2] 06983429-8afe-4bf9-ab65-aa4031568b9c/06983429-8afe-4bf9-ab65-aa4031568b9c.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1021.274176] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b37a9680-3e3f-4b11-98bd-18d1294c71ea {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.294428] env[62066]: DEBUG oslo_vmware.api [None req-4d1bdf8b-de84-42ec-81b7-e654d596c4f0 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Waiting for the task: (returnval){ [ 1021.294428] env[62066]: value = "task-1341328" [ 1021.294428] env[62066]: _type = "Task" [ 1021.294428] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.302117] env[62066]: DEBUG oslo_vmware.api [None req-4d1bdf8b-de84-42ec-81b7-e654d596c4f0 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341328, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.319959] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1341326, 'name': CreateVM_Task, 'duration_secs': 0.621639} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.320193] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3ef1410d-c78b-47d9-807b-e80368c05902] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1021.320962] env[62066]: DEBUG oslo_concurrency.lockutils [None req-39c438a5-51a5-4a29-a4b4-c8b849e56234 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1021.321164] env[62066]: DEBUG oslo_concurrency.lockutils [None req-39c438a5-51a5-4a29-a4b4-c8b849e56234 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Acquired lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1021.321579] env[62066]: DEBUG oslo_concurrency.lockutils [None req-39c438a5-51a5-4a29-a4b4-c8b849e56234 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1021.321843] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a4dfbcf5-5211-4a6b-9423-b81beccda193 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.326651] env[62066]: DEBUG oslo_vmware.api [None req-39c438a5-51a5-4a29-a4b4-c8b849e56234 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Waiting for the task: (returnval){ [ 1021.326651] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]526ec4fd-ca74-0c26-0e93-d00b71e5f98e" [ 1021.326651] env[62066]: _type = "Task" [ 1021.326651] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.335371] env[62066]: DEBUG oslo_vmware.api [None req-39c438a5-51a5-4a29-a4b4-c8b849e56234 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]526ec4fd-ca74-0c26-0e93-d00b71e5f98e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.590787] env[62066]: DEBUG oslo_concurrency.lockutils [req-838fe49f-74e6-4d22-aeba-19836242b19a req-75c641dd-caf3-4273-b653-656d52cf8733 service nova] Releasing lock "refresh_cache-3ef1410d-c78b-47d9-807b-e80368c05902" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1021.684223] env[62066]: DEBUG oslo_service.periodic_task [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1021.684470] env[62066]: DEBUG oslo_service.periodic_task [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1021.684620] env[62066]: DEBUG nova.compute.manager [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Starting heal instance info cache {{(pid=62066) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1021.684741] env[62066]: DEBUG nova.compute.manager [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Rebuilding the list of instances to heal {{(pid=62066) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 1021.805400] env[62066]: DEBUG oslo_vmware.api [None req-4d1bdf8b-de84-42ec-81b7-e654d596c4f0 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341328, 'name': ReconfigVM_Task, 'duration_secs': 0.279663} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.805736] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-4d1bdf8b-de84-42ec-81b7-e654d596c4f0 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: 06983429-8afe-4bf9-ab65-aa4031568b9c] Reconfigured VM instance instance-00000066 to attach disk [datastore2] 06983429-8afe-4bf9-ab65-aa4031568b9c/06983429-8afe-4bf9-ab65-aa4031568b9c.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1021.806317] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3451afca-3d19-4bf0-a239-c8d6dcd42a62 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.813222] env[62066]: DEBUG oslo_vmware.api [None req-4d1bdf8b-de84-42ec-81b7-e654d596c4f0 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Waiting for the task: (returnval){ [ 1021.813222] env[62066]: value = "task-1341329" [ 1021.813222] env[62066]: _type = "Task" [ 1021.813222] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.821768] env[62066]: DEBUG oslo_vmware.api [None req-4d1bdf8b-de84-42ec-81b7-e654d596c4f0 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341329, 'name': Rename_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.835387] env[62066]: DEBUG oslo_vmware.api [None req-39c438a5-51a5-4a29-a4b4-c8b849e56234 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]526ec4fd-ca74-0c26-0e93-d00b71e5f98e, 'name': SearchDatastore_Task, 'duration_secs': 0.020059} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.835736] env[62066]: DEBUG oslo_concurrency.lockutils [None req-39c438a5-51a5-4a29-a4b4-c8b849e56234 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Releasing lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1021.836060] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-39c438a5-51a5-4a29-a4b4-c8b849e56234 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: 3ef1410d-c78b-47d9-807b-e80368c05902] Processing image 50ff584c-3b50-4395-af07-3e66769bc9f7 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1021.836395] env[62066]: DEBUG oslo_concurrency.lockutils [None req-39c438a5-51a5-4a29-a4b4-c8b849e56234 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1021.836557] env[62066]: DEBUG oslo_concurrency.lockutils [None req-39c438a5-51a5-4a29-a4b4-c8b849e56234 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Acquired lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1021.836820] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-39c438a5-51a5-4a29-a4b4-c8b849e56234 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1021.837142] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-705883ad-985f-4a99-8285-9f2ba0955cd7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.845807] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-39c438a5-51a5-4a29-a4b4-c8b849e56234 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1021.845984] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-39c438a5-51a5-4a29-a4b4-c8b849e56234 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1021.846766] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7afef6f5-a210-4eb2-b454-e8ced933ea28 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.851844] env[62066]: DEBUG oslo_vmware.api [None req-39c438a5-51a5-4a29-a4b4-c8b849e56234 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Waiting for the task: (returnval){ [ 1021.851844] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]5228bbd9-55c3-c012-d4a4-56af2647d798" [ 1021.851844] env[62066]: _type = "Task" [ 1021.851844] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.859340] env[62066]: DEBUG oslo_vmware.api [None req-39c438a5-51a5-4a29-a4b4-c8b849e56234 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]5228bbd9-55c3-c012-d4a4-56af2647d798, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.187884] env[62066]: DEBUG nova.compute.manager [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] [instance: 06983429-8afe-4bf9-ab65-aa4031568b9c] Skipping network cache update for instance because it is Building. {{(pid=62066) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 1022.188058] env[62066]: DEBUG nova.compute.manager [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] [instance: 3ef1410d-c78b-47d9-807b-e80368c05902] Skipping network cache update for instance because it is Building. {{(pid=62066) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 1022.214591] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Acquiring lock "refresh_cache-c8482d92-0180-44ee-a4f1-bf84786dad43" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1022.214733] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Acquired lock "refresh_cache-c8482d92-0180-44ee-a4f1-bf84786dad43" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1022.214875] env[62066]: DEBUG nova.network.neutron [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] [instance: c8482d92-0180-44ee-a4f1-bf84786dad43] Forcefully refreshing network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1022.215046] env[62066]: DEBUG nova.objects.instance [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Lazy-loading 'info_cache' on Instance uuid c8482d92-0180-44ee-a4f1-bf84786dad43 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1022.323619] env[62066]: DEBUG oslo_vmware.api [None req-4d1bdf8b-de84-42ec-81b7-e654d596c4f0 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341329, 'name': Rename_Task, 'duration_secs': 0.138545} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.323839] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d1bdf8b-de84-42ec-81b7-e654d596c4f0 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: 06983429-8afe-4bf9-ab65-aa4031568b9c] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1022.324106] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ae75d879-e91f-4f2a-8f62-25cfe412919c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.331145] env[62066]: DEBUG oslo_vmware.api [None req-4d1bdf8b-de84-42ec-81b7-e654d596c4f0 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Waiting for the task: (returnval){ [ 1022.331145] env[62066]: value = "task-1341330" [ 1022.331145] env[62066]: _type = "Task" [ 1022.331145] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.338615] env[62066]: DEBUG oslo_vmware.api [None req-4d1bdf8b-de84-42ec-81b7-e654d596c4f0 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341330, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.362373] env[62066]: DEBUG oslo_vmware.api [None req-39c438a5-51a5-4a29-a4b4-c8b849e56234 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]5228bbd9-55c3-c012-d4a4-56af2647d798, 'name': SearchDatastore_Task, 'duration_secs': 0.008848} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.363184] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dafbd1db-c486-4ff1-aead-1741e1a6274c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.368641] env[62066]: DEBUG oslo_vmware.api [None req-39c438a5-51a5-4a29-a4b4-c8b849e56234 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Waiting for the task: (returnval){ [ 1022.368641] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]52744fbd-f005-abf5-a47b-985dbca744da" [ 1022.368641] env[62066]: _type = "Task" [ 1022.368641] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.377994] env[62066]: DEBUG oslo_vmware.api [None req-39c438a5-51a5-4a29-a4b4-c8b849e56234 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52744fbd-f005-abf5-a47b-985dbca744da, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.841230] env[62066]: DEBUG oslo_vmware.api [None req-4d1bdf8b-de84-42ec-81b7-e654d596c4f0 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341330, 'name': PowerOnVM_Task, 'duration_secs': 0.434728} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.841534] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d1bdf8b-de84-42ec-81b7-e654d596c4f0 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: 06983429-8afe-4bf9-ab65-aa4031568b9c] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1022.841684] env[62066]: INFO nova.compute.manager [None req-4d1bdf8b-de84-42ec-81b7-e654d596c4f0 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: 06983429-8afe-4bf9-ab65-aa4031568b9c] Took 6.69 seconds to spawn the instance on the hypervisor. [ 1022.841892] env[62066]: DEBUG nova.compute.manager [None req-4d1bdf8b-de84-42ec-81b7-e654d596c4f0 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: 06983429-8afe-4bf9-ab65-aa4031568b9c] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1022.842643] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6e81b2e-24b9-4819-b533-32858fde9272 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.879818] env[62066]: DEBUG oslo_vmware.api [None req-39c438a5-51a5-4a29-a4b4-c8b849e56234 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52744fbd-f005-abf5-a47b-985dbca744da, 'name': SearchDatastore_Task, 'duration_secs': 0.009596} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.880106] env[62066]: DEBUG oslo_concurrency.lockutils [None req-39c438a5-51a5-4a29-a4b4-c8b849e56234 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Releasing lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1022.880367] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-39c438a5-51a5-4a29-a4b4-c8b849e56234 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk to [datastore2] 3ef1410d-c78b-47d9-807b-e80368c05902/3ef1410d-c78b-47d9-807b-e80368c05902.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1022.880637] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a8f2989a-aeee-4667-87e9-2100e54aa5bc {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.888604] env[62066]: DEBUG oslo_vmware.api [None req-39c438a5-51a5-4a29-a4b4-c8b849e56234 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Waiting for the task: (returnval){ [ 1022.888604] env[62066]: value = "task-1341331" [ 1022.888604] env[62066]: _type = "Task" [ 1022.888604] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.896816] env[62066]: DEBUG oslo_vmware.api [None req-39c438a5-51a5-4a29-a4b4-c8b849e56234 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Task: {'id': task-1341331, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.361978] env[62066]: INFO nova.compute.manager [None req-4d1bdf8b-de84-42ec-81b7-e654d596c4f0 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: 06983429-8afe-4bf9-ab65-aa4031568b9c] Took 11.50 seconds to build instance. [ 1023.398589] env[62066]: DEBUG oslo_vmware.api [None req-39c438a5-51a5-4a29-a4b4-c8b849e56234 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Task: {'id': task-1341331, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.440067} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.398856] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-39c438a5-51a5-4a29-a4b4-c8b849e56234 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk to [datastore2] 3ef1410d-c78b-47d9-807b-e80368c05902/3ef1410d-c78b-47d9-807b-e80368c05902.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1023.399090] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-39c438a5-51a5-4a29-a4b4-c8b849e56234 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: 3ef1410d-c78b-47d9-807b-e80368c05902] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1023.399347] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8d197f94-f13f-4c5f-abdf-d96f68bbdb02 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.406462] env[62066]: DEBUG oslo_vmware.api [None req-39c438a5-51a5-4a29-a4b4-c8b849e56234 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Waiting for the task: (returnval){ [ 1023.406462] env[62066]: value = "task-1341332" [ 1023.406462] env[62066]: _type = "Task" [ 1023.406462] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.415080] env[62066]: DEBUG oslo_vmware.api [None req-39c438a5-51a5-4a29-a4b4-c8b849e56234 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Task: {'id': task-1341332, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.863465] env[62066]: DEBUG oslo_concurrency.lockutils [None req-4d1bdf8b-de84-42ec-81b7-e654d596c4f0 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Lock "06983429-8afe-4bf9-ab65-aa4031568b9c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.018s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1023.917889] env[62066]: DEBUG oslo_vmware.api [None req-39c438a5-51a5-4a29-a4b4-c8b849e56234 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Task: {'id': task-1341332, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.061009} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.918221] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-39c438a5-51a5-4a29-a4b4-c8b849e56234 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: 3ef1410d-c78b-47d9-807b-e80368c05902] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1023.918971] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3b57b2b-a84d-4bf5-b9c0-8034c7ac3ae1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.941215] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-39c438a5-51a5-4a29-a4b4-c8b849e56234 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: 3ef1410d-c78b-47d9-807b-e80368c05902] Reconfiguring VM instance instance-00000067 to attach disk [datastore2] 3ef1410d-c78b-47d9-807b-e80368c05902/3ef1410d-c78b-47d9-807b-e80368c05902.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1023.941501] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2afc7d0f-2e03-4114-8ee9-3357d64c114b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.965366] env[62066]: DEBUG oslo_vmware.api [None req-39c438a5-51a5-4a29-a4b4-c8b849e56234 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Waiting for the task: (returnval){ [ 1023.965366] env[62066]: value = "task-1341333" [ 1023.965366] env[62066]: _type = "Task" [ 1023.965366] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.973792] env[62066]: DEBUG oslo_vmware.api [None req-39c438a5-51a5-4a29-a4b4-c8b849e56234 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Task: {'id': task-1341333, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.994108] env[62066]: DEBUG nova.network.neutron [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] [instance: c8482d92-0180-44ee-a4f1-bf84786dad43] Updating instance_info_cache with network_info: [{"id": "2f2f1786-934d-478f-be39-7509cfacf86f", "address": "fa:16:3e:24:41:2e", "network": {"id": "449dfe4e-bb49-43b6-9ba7-b57af74ebfb3", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-282029676-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.231", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0d4e33a99c7741fb8cdd97f4ec5dbbd8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db68bd64-5b56-49af-a075-13dcf85cb2e0", "external-id": "nsx-vlan-transportzone-590", "segmentation_id": 590, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2f2f1786-93", "ovs_interfaceid": "2f2f1786-934d-478f-be39-7509cfacf86f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1024.095205] env[62066]: DEBUG oslo_concurrency.lockutils [None req-aff71cbe-8521-4ba3-8073-85716f9ff0ee tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Acquiring lock "06983429-8afe-4bf9-ab65-aa4031568b9c" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1024.095403] env[62066]: DEBUG oslo_concurrency.lockutils [None req-aff71cbe-8521-4ba3-8073-85716f9ff0ee tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Lock "06983429-8afe-4bf9-ab65-aa4031568b9c" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1024.095591] env[62066]: DEBUG nova.compute.manager [None req-aff71cbe-8521-4ba3-8073-85716f9ff0ee tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: 06983429-8afe-4bf9-ab65-aa4031568b9c] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1024.097212] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f44e48d-f94e-4c11-a041-b1f0bfcf2948 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.104213] env[62066]: DEBUG nova.compute.manager [None req-aff71cbe-8521-4ba3-8073-85716f9ff0ee tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: 06983429-8afe-4bf9-ab65-aa4031568b9c] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62066) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1024.104783] env[62066]: DEBUG nova.objects.instance [None req-aff71cbe-8521-4ba3-8073-85716f9ff0ee tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Lazy-loading 'flavor' on Instance uuid 06983429-8afe-4bf9-ab65-aa4031568b9c {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1024.475088] env[62066]: DEBUG oslo_vmware.api [None req-39c438a5-51a5-4a29-a4b4-c8b849e56234 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Task: {'id': task-1341333, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.496761] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Releasing lock "refresh_cache-c8482d92-0180-44ee-a4f1-bf84786dad43" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1024.496959] env[62066]: DEBUG nova.compute.manager [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] [instance: c8482d92-0180-44ee-a4f1-bf84786dad43] Updated the network info_cache for instance {{(pid=62066) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 1024.497171] env[62066]: DEBUG oslo_service.periodic_task [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1024.497334] env[62066]: DEBUG oslo_service.periodic_task [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1024.497483] env[62066]: DEBUG oslo_service.periodic_task [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1024.497633] env[62066]: DEBUG oslo_service.periodic_task [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1024.497774] env[62066]: DEBUG oslo_service.periodic_task [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1024.497924] env[62066]: DEBUG oslo_service.periodic_task [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1024.498061] env[62066]: DEBUG nova.compute.manager [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62066) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1024.498212] env[62066]: DEBUG oslo_service.periodic_task [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Running periodic task ComputeManager.update_available_resource {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1024.609678] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-aff71cbe-8521-4ba3-8073-85716f9ff0ee tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: 06983429-8afe-4bf9-ab65-aa4031568b9c] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1024.609954] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d2bc017a-678c-4bce-a0c1-1b6c521e90da {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.618095] env[62066]: DEBUG oslo_vmware.api [None req-aff71cbe-8521-4ba3-8073-85716f9ff0ee tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Waiting for the task: (returnval){ [ 1024.618095] env[62066]: value = "task-1341334" [ 1024.618095] env[62066]: _type = "Task" [ 1024.618095] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.627427] env[62066]: DEBUG oslo_vmware.api [None req-aff71cbe-8521-4ba3-8073-85716f9ff0ee tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341334, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.976088] env[62066]: DEBUG oslo_vmware.api [None req-39c438a5-51a5-4a29-a4b4-c8b849e56234 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Task: {'id': task-1341333, 'name': ReconfigVM_Task, 'duration_secs': 0.514485} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.976516] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-39c438a5-51a5-4a29-a4b4-c8b849e56234 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: 3ef1410d-c78b-47d9-807b-e80368c05902] Reconfigured VM instance instance-00000067 to attach disk [datastore2] 3ef1410d-c78b-47d9-807b-e80368c05902/3ef1410d-c78b-47d9-807b-e80368c05902.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1024.977164] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4393c28f-361e-4066-814e-e7b250ae0196 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.985531] env[62066]: DEBUG oslo_vmware.api [None req-39c438a5-51a5-4a29-a4b4-c8b849e56234 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Waiting for the task: (returnval){ [ 1024.985531] env[62066]: value = "task-1341335" [ 1024.985531] env[62066]: _type = "Task" [ 1024.985531] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.994789] env[62066]: DEBUG oslo_vmware.api [None req-39c438a5-51a5-4a29-a4b4-c8b849e56234 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Task: {'id': task-1341335, 'name': Rename_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.000904] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1025.001127] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1025.001302] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1025.001454] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62066) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1025.002342] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3dc5e1cf-b6c6-42b8-80bc-6ef7fc60b724 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.010534] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26929741-637c-4d49-aff7-f281a9dcb931 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.025469] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6196fbeb-1f69-4e0b-a274-c33681918d63 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.032984] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b62b85f-67fb-4c2f-8eba-2b0d927d778b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.062841] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180652MB free_disk=154GB free_vcpus=48 pci_devices=None {{(pid=62066) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1025.062982] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1025.063272] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1025.128138] env[62066]: DEBUG oslo_vmware.api [None req-aff71cbe-8521-4ba3-8073-85716f9ff0ee tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341334, 'name': PowerOffVM_Task, 'duration_secs': 0.183796} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.128491] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-aff71cbe-8521-4ba3-8073-85716f9ff0ee tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: 06983429-8afe-4bf9-ab65-aa4031568b9c] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1025.128717] env[62066]: DEBUG nova.compute.manager [None req-aff71cbe-8521-4ba3-8073-85716f9ff0ee tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: 06983429-8afe-4bf9-ab65-aa4031568b9c] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1025.129643] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-313b0861-9d80-492d-aa4e-40831f067fe0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.497235] env[62066]: DEBUG oslo_vmware.api [None req-39c438a5-51a5-4a29-a4b4-c8b849e56234 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Task: {'id': task-1341335, 'name': Rename_Task, 'duration_secs': 0.156686} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.497515] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-39c438a5-51a5-4a29-a4b4-c8b849e56234 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: 3ef1410d-c78b-47d9-807b-e80368c05902] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1025.497763] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-26ce7f1e-33eb-4a2a-8d12-a563e4fc0f3b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.505523] env[62066]: DEBUG oslo_vmware.api [None req-39c438a5-51a5-4a29-a4b4-c8b849e56234 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Waiting for the task: (returnval){ [ 1025.505523] env[62066]: value = "task-1341336" [ 1025.505523] env[62066]: _type = "Task" [ 1025.505523] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.513636] env[62066]: DEBUG oslo_vmware.api [None req-39c438a5-51a5-4a29-a4b4-c8b849e56234 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Task: {'id': task-1341336, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.641768] env[62066]: DEBUG oslo_concurrency.lockutils [None req-aff71cbe-8521-4ba3-8073-85716f9ff0ee tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Lock "06983429-8afe-4bf9-ab65-aa4031568b9c" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.546s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1025.981804] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d0e86307-d1d5-4c1f-9fa1-62e41040f376 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Acquiring lock "interface-c8482d92-0180-44ee-a4f1-bf84786dad43-dda94cd6-0a2d-4c6b-bc0e-4fb38129159f" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1025.982130] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d0e86307-d1d5-4c1f-9fa1-62e41040f376 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Lock "interface-c8482d92-0180-44ee-a4f1-bf84786dad43-dda94cd6-0a2d-4c6b-bc0e-4fb38129159f" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1025.982461] env[62066]: DEBUG nova.objects.instance [None req-d0e86307-d1d5-4c1f-9fa1-62e41040f376 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Lazy-loading 'flavor' on Instance uuid c8482d92-0180-44ee-a4f1-bf84786dad43 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1026.016350] env[62066]: DEBUG oslo_vmware.api [None req-39c438a5-51a5-4a29-a4b4-c8b849e56234 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Task: {'id': task-1341336, 'name': PowerOnVM_Task, 'duration_secs': 0.475753} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.016464] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-39c438a5-51a5-4a29-a4b4-c8b849e56234 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: 3ef1410d-c78b-47d9-807b-e80368c05902] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1026.016661] env[62066]: INFO nova.compute.manager [None req-39c438a5-51a5-4a29-a4b4-c8b849e56234 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: 3ef1410d-c78b-47d9-807b-e80368c05902] Took 7.72 seconds to spawn the instance on the hypervisor. [ 1026.016847] env[62066]: DEBUG nova.compute.manager [None req-39c438a5-51a5-4a29-a4b4-c8b849e56234 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: 3ef1410d-c78b-47d9-807b-e80368c05902] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1026.017635] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b551f36-02da-488c-950a-989f7526b630 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.088037] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Instance c8482d92-0180-44ee-a4f1-bf84786dad43 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1026.088197] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Instance 7178eb83-41e7-4516-809d-a0972d7e6e23 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1026.088306] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Instance 06983429-8afe-4bf9-ab65-aa4031568b9c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1026.088419] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Instance 3ef1410d-c78b-47d9-807b-e80368c05902 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1026.088607] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Total usable vcpus: 48, total allocated vcpus: 4 {{(pid=62066) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1026.088804] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1280MB phys_disk=200GB used_disk=4GB total_vcpus=48 used_vcpus=4 pci_stats=[] {{(pid=62066) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1026.149302] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0797f19-11eb-4f9f-b1fd-771294325185 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.157057] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfb4f3b7-bdfe-421d-8674-1769a251cbe0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.186311] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e3b4b2e-a9f0-4327-a79d-f9cb9e388f1f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.195233] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37f5f8d6-270a-47f7-acc2-0a0d1fc5446f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.208044] env[62066]: DEBUG nova.compute.provider_tree [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Updating inventory in ProviderTree for provider 8e4f7194-1498-4f08-8723-ab7260524bcb with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1026.462204] env[62066]: DEBUG oslo_concurrency.lockutils [None req-775e70f5-1ec0-49af-b6c8-e68074f6f67f tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Acquiring lock "06983429-8afe-4bf9-ab65-aa4031568b9c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1026.462495] env[62066]: DEBUG oslo_concurrency.lockutils [None req-775e70f5-1ec0-49af-b6c8-e68074f6f67f tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Lock "06983429-8afe-4bf9-ab65-aa4031568b9c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1026.462770] env[62066]: DEBUG oslo_concurrency.lockutils [None req-775e70f5-1ec0-49af-b6c8-e68074f6f67f tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Acquiring lock "06983429-8afe-4bf9-ab65-aa4031568b9c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1026.462897] env[62066]: DEBUG oslo_concurrency.lockutils [None req-775e70f5-1ec0-49af-b6c8-e68074f6f67f tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Lock "06983429-8afe-4bf9-ab65-aa4031568b9c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1026.463086] env[62066]: DEBUG oslo_concurrency.lockutils [None req-775e70f5-1ec0-49af-b6c8-e68074f6f67f tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Lock "06983429-8afe-4bf9-ab65-aa4031568b9c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1026.465252] env[62066]: INFO nova.compute.manager [None req-775e70f5-1ec0-49af-b6c8-e68074f6f67f tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: 06983429-8afe-4bf9-ab65-aa4031568b9c] Terminating instance [ 1026.467338] env[62066]: DEBUG nova.compute.manager [None req-775e70f5-1ec0-49af-b6c8-e68074f6f67f tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: 06983429-8afe-4bf9-ab65-aa4031568b9c] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1026.467538] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-775e70f5-1ec0-49af-b6c8-e68074f6f67f tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: 06983429-8afe-4bf9-ab65-aa4031568b9c] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1026.468358] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f5fff05-206d-4722-9433-d96f2c5b8f61 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.476568] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-775e70f5-1ec0-49af-b6c8-e68074f6f67f tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: 06983429-8afe-4bf9-ab65-aa4031568b9c] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1026.476852] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-63e17a21-56a2-41fe-b05f-1f876760985a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.539060] env[62066]: INFO nova.compute.manager [None req-39c438a5-51a5-4a29-a4b4-c8b849e56234 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: 3ef1410d-c78b-47d9-807b-e80368c05902] Took 14.17 seconds to build instance. [ 1026.545793] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-775e70f5-1ec0-49af-b6c8-e68074f6f67f tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: 06983429-8afe-4bf9-ab65-aa4031568b9c] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1026.545999] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-775e70f5-1ec0-49af-b6c8-e68074f6f67f tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: 06983429-8afe-4bf9-ab65-aa4031568b9c] Deleting contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1026.546220] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-775e70f5-1ec0-49af-b6c8-e68074f6f67f tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Deleting the datastore file [datastore2] 06983429-8afe-4bf9-ab65-aa4031568b9c {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1026.546480] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ef0d59dc-3df4-483b-95b3-70652ea64e01 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.556387] env[62066]: DEBUG oslo_vmware.api [None req-775e70f5-1ec0-49af-b6c8-e68074f6f67f tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Waiting for the task: (returnval){ [ 1026.556387] env[62066]: value = "task-1341338" [ 1026.556387] env[62066]: _type = "Task" [ 1026.556387] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.565854] env[62066]: DEBUG oslo_vmware.api [None req-775e70f5-1ec0-49af-b6c8-e68074f6f67f tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341338, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.666254] env[62066]: DEBUG nova.objects.instance [None req-d0e86307-d1d5-4c1f-9fa1-62e41040f376 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Lazy-loading 'pci_requests' on Instance uuid c8482d92-0180-44ee-a4f1-bf84786dad43 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1026.680718] env[62066]: DEBUG nova.compute.manager [req-3f2a9b4e-f4ec-4d6a-bab8-b403d6a156fb req-46a16a52-1244-4e0e-8994-e634e820861f service nova] [instance: 3ef1410d-c78b-47d9-807b-e80368c05902] Received event network-changed-3b1d21b6-842c-4761-867d-8b06174242db {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1026.681014] env[62066]: DEBUG nova.compute.manager [req-3f2a9b4e-f4ec-4d6a-bab8-b403d6a156fb req-46a16a52-1244-4e0e-8994-e634e820861f service nova] [instance: 3ef1410d-c78b-47d9-807b-e80368c05902] Refreshing instance network info cache due to event network-changed-3b1d21b6-842c-4761-867d-8b06174242db. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1026.681249] env[62066]: DEBUG oslo_concurrency.lockutils [req-3f2a9b4e-f4ec-4d6a-bab8-b403d6a156fb req-46a16a52-1244-4e0e-8994-e634e820861f service nova] Acquiring lock "refresh_cache-3ef1410d-c78b-47d9-807b-e80368c05902" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1026.681399] env[62066]: DEBUG oslo_concurrency.lockutils [req-3f2a9b4e-f4ec-4d6a-bab8-b403d6a156fb req-46a16a52-1244-4e0e-8994-e634e820861f service nova] Acquired lock "refresh_cache-3ef1410d-c78b-47d9-807b-e80368c05902" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1026.682424] env[62066]: DEBUG nova.network.neutron [req-3f2a9b4e-f4ec-4d6a-bab8-b403d6a156fb req-46a16a52-1244-4e0e-8994-e634e820861f service nova] [instance: 3ef1410d-c78b-47d9-807b-e80368c05902] Refreshing network info cache for port 3b1d21b6-842c-4761-867d-8b06174242db {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1026.728932] env[62066]: ERROR nova.scheduler.client.report [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] [req-cee6fb07-c5fa-4316-8f58-ba0f6f03c473] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 8e4f7194-1498-4f08-8723-ab7260524bcb. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-cee6fb07-c5fa-4316-8f58-ba0f6f03c473"}]} [ 1026.744508] env[62066]: DEBUG nova.scheduler.client.report [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Refreshing inventories for resource provider 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 1026.757313] env[62066]: DEBUG nova.scheduler.client.report [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Updating ProviderTree inventory for provider 8e4f7194-1498-4f08-8723-ab7260524bcb from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 1026.757494] env[62066]: DEBUG nova.compute.provider_tree [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Updating inventory in ProviderTree for provider 8e4f7194-1498-4f08-8723-ab7260524bcb with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1026.769046] env[62066]: DEBUG nova.scheduler.client.report [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Refreshing aggregate associations for resource provider 8e4f7194-1498-4f08-8723-ab7260524bcb, aggregates: None {{(pid=62066) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 1026.788528] env[62066]: DEBUG nova.scheduler.client.report [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Refreshing trait associations for resource provider 8e4f7194-1498-4f08-8723-ab7260524bcb, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,HW_ARCH_X86_64,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=62066) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 1026.845059] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7723143-c098-4d19-9366-611746b265a0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.853348] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0273c2f7-1c4a-4c33-8a46-ea99a92bdc69 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.884315] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02dd69dc-e49f-4545-947b-469e01592c9a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.892600] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c55569c-4b3c-40c4-98d9-fb51538171e5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.906794] env[62066]: DEBUG nova.compute.provider_tree [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Updating inventory in ProviderTree for provider 8e4f7194-1498-4f08-8723-ab7260524bcb with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1027.042257] env[62066]: DEBUG oslo_concurrency.lockutils [None req-39c438a5-51a5-4a29-a4b4-c8b849e56234 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Lock "3ef1410d-c78b-47d9-807b-e80368c05902" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.690s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1027.066726] env[62066]: DEBUG oslo_vmware.api [None req-775e70f5-1ec0-49af-b6c8-e68074f6f67f tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341338, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.149331} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1027.067116] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-775e70f5-1ec0-49af-b6c8-e68074f6f67f tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1027.067435] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-775e70f5-1ec0-49af-b6c8-e68074f6f67f tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: 06983429-8afe-4bf9-ab65-aa4031568b9c] Deleted contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1027.067564] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-775e70f5-1ec0-49af-b6c8-e68074f6f67f tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: 06983429-8afe-4bf9-ab65-aa4031568b9c] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1027.067746] env[62066]: INFO nova.compute.manager [None req-775e70f5-1ec0-49af-b6c8-e68074f6f67f tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: 06983429-8afe-4bf9-ab65-aa4031568b9c] Took 0.60 seconds to destroy the instance on the hypervisor. [ 1027.067998] env[62066]: DEBUG oslo.service.loopingcall [None req-775e70f5-1ec0-49af-b6c8-e68074f6f67f tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1027.068217] env[62066]: DEBUG nova.compute.manager [-] [instance: 06983429-8afe-4bf9-ab65-aa4031568b9c] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1027.068313] env[62066]: DEBUG nova.network.neutron [-] [instance: 06983429-8afe-4bf9-ab65-aa4031568b9c] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1027.168590] env[62066]: DEBUG nova.objects.base [None req-d0e86307-d1d5-4c1f-9fa1-62e41040f376 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=62066) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1027.168819] env[62066]: DEBUG nova.network.neutron [None req-d0e86307-d1d5-4c1f-9fa1-62e41040f376 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: c8482d92-0180-44ee-a4f1-bf84786dad43] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1027.282376] env[62066]: DEBUG nova.policy [None req-d0e86307-d1d5-4c1f-9fa1-62e41040f376 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '747026d272d4467bbf14a0c30589c3b7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0d4e33a99c7741fb8cdd97f4ec5dbbd8', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 1027.439309] env[62066]: DEBUG nova.scheduler.client.report [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Updated inventory for provider 8e4f7194-1498-4f08-8723-ab7260524bcb with generation 120 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 1027.439659] env[62066]: DEBUG nova.compute.provider_tree [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Updating resource provider 8e4f7194-1498-4f08-8723-ab7260524bcb generation from 120 to 121 during operation: update_inventory {{(pid=62066) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1027.439820] env[62066]: DEBUG nova.compute.provider_tree [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Updating inventory in ProviderTree for provider 8e4f7194-1498-4f08-8723-ab7260524bcb with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1027.533478] env[62066]: DEBUG nova.network.neutron [req-3f2a9b4e-f4ec-4d6a-bab8-b403d6a156fb req-46a16a52-1244-4e0e-8994-e634e820861f service nova] [instance: 3ef1410d-c78b-47d9-807b-e80368c05902] Updated VIF entry in instance network info cache for port 3b1d21b6-842c-4761-867d-8b06174242db. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1027.533850] env[62066]: DEBUG nova.network.neutron [req-3f2a9b4e-f4ec-4d6a-bab8-b403d6a156fb req-46a16a52-1244-4e0e-8994-e634e820861f service nova] [instance: 3ef1410d-c78b-47d9-807b-e80368c05902] Updating instance_info_cache with network_info: [{"id": "3b1d21b6-842c-4761-867d-8b06174242db", "address": "fa:16:3e:67:4b:4e", "network": {"id": "5b89b2d4-71dd-4cd1-9edb-f4040be26540", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-283918837-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.149", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2252d0f6f18a47fd8a1ddad8020a6db7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "352165bb-004f-4180-9627-3a275dbe18af", "external-id": "nsx-vlan-transportzone-926", "segmentation_id": 926, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3b1d21b6-84", "ovs_interfaceid": "3b1d21b6-842c-4761-867d-8b06174242db", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1027.893836] env[62066]: DEBUG nova.network.neutron [-] [instance: 06983429-8afe-4bf9-ab65-aa4031568b9c] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1027.944934] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62066) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1027.945282] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.882s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1028.036160] env[62066]: DEBUG oslo_concurrency.lockutils [req-3f2a9b4e-f4ec-4d6a-bab8-b403d6a156fb req-46a16a52-1244-4e0e-8994-e634e820861f service nova] Releasing lock "refresh_cache-3ef1410d-c78b-47d9-807b-e80368c05902" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1028.396840] env[62066]: INFO nova.compute.manager [-] [instance: 06983429-8afe-4bf9-ab65-aa4031568b9c] Took 1.33 seconds to deallocate network for instance. [ 1028.706963] env[62066]: DEBUG nova.compute.manager [req-3b60c3e3-e939-4322-b534-9ec0c71cdb75 req-eb4c6012-f36a-41ba-ab9a-f97c71c189f3 service nova] [instance: 06983429-8afe-4bf9-ab65-aa4031568b9c] Received event network-vif-deleted-1be70afc-e68a-481b-909e-9bd764d93b21 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1028.789625] env[62066]: DEBUG nova.compute.manager [req-aa39f6f3-1d15-4545-905a-d2a39bf9fde7 req-f1ecbab2-53e0-4e45-84c8-1ee9f7b97128 service nova] [instance: c8482d92-0180-44ee-a4f1-bf84786dad43] Received event network-vif-plugged-dda94cd6-0a2d-4c6b-bc0e-4fb38129159f {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1028.789862] env[62066]: DEBUG oslo_concurrency.lockutils [req-aa39f6f3-1d15-4545-905a-d2a39bf9fde7 req-f1ecbab2-53e0-4e45-84c8-1ee9f7b97128 service nova] Acquiring lock "c8482d92-0180-44ee-a4f1-bf84786dad43-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1028.790278] env[62066]: DEBUG oslo_concurrency.lockutils [req-aa39f6f3-1d15-4545-905a-d2a39bf9fde7 req-f1ecbab2-53e0-4e45-84c8-1ee9f7b97128 service nova] Lock "c8482d92-0180-44ee-a4f1-bf84786dad43-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1028.790495] env[62066]: DEBUG oslo_concurrency.lockutils [req-aa39f6f3-1d15-4545-905a-d2a39bf9fde7 req-f1ecbab2-53e0-4e45-84c8-1ee9f7b97128 service nova] Lock "c8482d92-0180-44ee-a4f1-bf84786dad43-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1028.790680] env[62066]: DEBUG nova.compute.manager [req-aa39f6f3-1d15-4545-905a-d2a39bf9fde7 req-f1ecbab2-53e0-4e45-84c8-1ee9f7b97128 service nova] [instance: c8482d92-0180-44ee-a4f1-bf84786dad43] No waiting events found dispatching network-vif-plugged-dda94cd6-0a2d-4c6b-bc0e-4fb38129159f {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1028.790940] env[62066]: WARNING nova.compute.manager [req-aa39f6f3-1d15-4545-905a-d2a39bf9fde7 req-f1ecbab2-53e0-4e45-84c8-1ee9f7b97128 service nova] [instance: c8482d92-0180-44ee-a4f1-bf84786dad43] Received unexpected event network-vif-plugged-dda94cd6-0a2d-4c6b-bc0e-4fb38129159f for instance with vm_state active and task_state None. [ 1028.881851] env[62066]: DEBUG nova.network.neutron [None req-d0e86307-d1d5-4c1f-9fa1-62e41040f376 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: c8482d92-0180-44ee-a4f1-bf84786dad43] Successfully updated port: dda94cd6-0a2d-4c6b-bc0e-4fb38129159f {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1028.904385] env[62066]: DEBUG oslo_concurrency.lockutils [None req-775e70f5-1ec0-49af-b6c8-e68074f6f67f tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1028.904665] env[62066]: DEBUG oslo_concurrency.lockutils [None req-775e70f5-1ec0-49af-b6c8-e68074f6f67f tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1028.904947] env[62066]: DEBUG nova.objects.instance [None req-775e70f5-1ec0-49af-b6c8-e68074f6f67f tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Lazy-loading 'resources' on Instance uuid 06983429-8afe-4bf9-ab65-aa4031568b9c {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1029.384715] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d0e86307-d1d5-4c1f-9fa1-62e41040f376 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Acquiring lock "refresh_cache-c8482d92-0180-44ee-a4f1-bf84786dad43" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1029.387335] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d0e86307-d1d5-4c1f-9fa1-62e41040f376 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Acquired lock "refresh_cache-c8482d92-0180-44ee-a4f1-bf84786dad43" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1029.387335] env[62066]: DEBUG nova.network.neutron [None req-d0e86307-d1d5-4c1f-9fa1-62e41040f376 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: c8482d92-0180-44ee-a4f1-bf84786dad43] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1029.466538] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d150e4ef-3f4f-4fdd-b11d-992408291877 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.474763] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd8d7dc2-4466-48d0-8aeb-67368e6db8b7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.503815] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d418126e-ac68-41e2-8bae-bedadf99bf3a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.511114] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2ca5d3b-a191-446c-a8c2-98b3d970df16 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.525103] env[62066]: DEBUG nova.compute.provider_tree [None req-775e70f5-1ec0-49af-b6c8-e68074f6f67f tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1029.920109] env[62066]: WARNING nova.network.neutron [None req-d0e86307-d1d5-4c1f-9fa1-62e41040f376 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: c8482d92-0180-44ee-a4f1-bf84786dad43] 449dfe4e-bb49-43b6-9ba7-b57af74ebfb3 already exists in list: networks containing: ['449dfe4e-bb49-43b6-9ba7-b57af74ebfb3']. ignoring it [ 1030.028010] env[62066]: DEBUG nova.scheduler.client.report [None req-775e70f5-1ec0-49af-b6c8-e68074f6f67f tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1030.182058] env[62066]: DEBUG nova.network.neutron [None req-d0e86307-d1d5-4c1f-9fa1-62e41040f376 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: c8482d92-0180-44ee-a4f1-bf84786dad43] Updating instance_info_cache with network_info: [{"id": "2f2f1786-934d-478f-be39-7509cfacf86f", "address": "fa:16:3e:24:41:2e", "network": {"id": "449dfe4e-bb49-43b6-9ba7-b57af74ebfb3", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-282029676-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.231", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0d4e33a99c7741fb8cdd97f4ec5dbbd8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db68bd64-5b56-49af-a075-13dcf85cb2e0", "external-id": "nsx-vlan-transportzone-590", "segmentation_id": 590, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2f2f1786-93", "ovs_interfaceid": "2f2f1786-934d-478f-be39-7509cfacf86f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "dda94cd6-0a2d-4c6b-bc0e-4fb38129159f", "address": "fa:16:3e:60:d6:92", "network": {"id": "449dfe4e-bb49-43b6-9ba7-b57af74ebfb3", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-282029676-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0d4e33a99c7741fb8cdd97f4ec5dbbd8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db68bd64-5b56-49af-a075-13dcf85cb2e0", "external-id": "nsx-vlan-transportzone-590", "segmentation_id": 590, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdda94cd6-0a", "ovs_interfaceid": "dda94cd6-0a2d-4c6b-bc0e-4fb38129159f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1030.533171] env[62066]: DEBUG oslo_concurrency.lockutils [None req-775e70f5-1ec0-49af-b6c8-e68074f6f67f tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.628s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1030.555883] env[62066]: INFO nova.scheduler.client.report [None req-775e70f5-1ec0-49af-b6c8-e68074f6f67f tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Deleted allocations for instance 06983429-8afe-4bf9-ab65-aa4031568b9c [ 1030.684431] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d0e86307-d1d5-4c1f-9fa1-62e41040f376 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Releasing lock "refresh_cache-c8482d92-0180-44ee-a4f1-bf84786dad43" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1030.685160] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d0e86307-d1d5-4c1f-9fa1-62e41040f376 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Acquiring lock "c8482d92-0180-44ee-a4f1-bf84786dad43" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1030.685361] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d0e86307-d1d5-4c1f-9fa1-62e41040f376 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Acquired lock "c8482d92-0180-44ee-a4f1-bf84786dad43" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1030.686232] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bb1ffd3-409c-460d-8390-1acae523a68c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.703586] env[62066]: DEBUG nova.virt.hardware [None req-d0e86307-d1d5-4c1f-9fa1-62e41040f376 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-23T13:40:41Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1030.703811] env[62066]: DEBUG nova.virt.hardware [None req-d0e86307-d1d5-4c1f-9fa1-62e41040f376 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1030.703974] env[62066]: DEBUG nova.virt.hardware [None req-d0e86307-d1d5-4c1f-9fa1-62e41040f376 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1030.704181] env[62066]: DEBUG nova.virt.hardware [None req-d0e86307-d1d5-4c1f-9fa1-62e41040f376 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1030.704334] env[62066]: DEBUG nova.virt.hardware [None req-d0e86307-d1d5-4c1f-9fa1-62e41040f376 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1030.704510] env[62066]: DEBUG nova.virt.hardware [None req-d0e86307-d1d5-4c1f-9fa1-62e41040f376 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1030.704688] env[62066]: DEBUG nova.virt.hardware [None req-d0e86307-d1d5-4c1f-9fa1-62e41040f376 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1030.704853] env[62066]: DEBUG nova.virt.hardware [None req-d0e86307-d1d5-4c1f-9fa1-62e41040f376 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1030.705040] env[62066]: DEBUG nova.virt.hardware [None req-d0e86307-d1d5-4c1f-9fa1-62e41040f376 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1030.705217] env[62066]: DEBUG nova.virt.hardware [None req-d0e86307-d1d5-4c1f-9fa1-62e41040f376 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1030.705396] env[62066]: DEBUG nova.virt.hardware [None req-d0e86307-d1d5-4c1f-9fa1-62e41040f376 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1030.711721] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-d0e86307-d1d5-4c1f-9fa1-62e41040f376 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: c8482d92-0180-44ee-a4f1-bf84786dad43] Reconfiguring VM to attach interface {{(pid=62066) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 1030.712038] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3d6b843b-2d9e-4dc1-a37f-808f30b351fc {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.728798] env[62066]: DEBUG oslo_vmware.api [None req-d0e86307-d1d5-4c1f-9fa1-62e41040f376 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Waiting for the task: (returnval){ [ 1030.728798] env[62066]: value = "task-1341339" [ 1030.728798] env[62066]: _type = "Task" [ 1030.728798] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.736679] env[62066]: DEBUG oslo_vmware.api [None req-d0e86307-d1d5-4c1f-9fa1-62e41040f376 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': task-1341339, 'name': ReconfigVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.814844] env[62066]: DEBUG nova.compute.manager [req-0def89de-1176-40a0-a364-6e11384c5cc2 req-d5fe8ad7-cde4-43f3-882b-073c8414cc3d service nova] [instance: c8482d92-0180-44ee-a4f1-bf84786dad43] Received event network-changed-dda94cd6-0a2d-4c6b-bc0e-4fb38129159f {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1030.815064] env[62066]: DEBUG nova.compute.manager [req-0def89de-1176-40a0-a364-6e11384c5cc2 req-d5fe8ad7-cde4-43f3-882b-073c8414cc3d service nova] [instance: c8482d92-0180-44ee-a4f1-bf84786dad43] Refreshing instance network info cache due to event network-changed-dda94cd6-0a2d-4c6b-bc0e-4fb38129159f. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1030.815314] env[62066]: DEBUG oslo_concurrency.lockutils [req-0def89de-1176-40a0-a364-6e11384c5cc2 req-d5fe8ad7-cde4-43f3-882b-073c8414cc3d service nova] Acquiring lock "refresh_cache-c8482d92-0180-44ee-a4f1-bf84786dad43" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1030.815462] env[62066]: DEBUG oslo_concurrency.lockutils [req-0def89de-1176-40a0-a364-6e11384c5cc2 req-d5fe8ad7-cde4-43f3-882b-073c8414cc3d service nova] Acquired lock "refresh_cache-c8482d92-0180-44ee-a4f1-bf84786dad43" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1030.815702] env[62066]: DEBUG nova.network.neutron [req-0def89de-1176-40a0-a364-6e11384c5cc2 req-d5fe8ad7-cde4-43f3-882b-073c8414cc3d service nova] [instance: c8482d92-0180-44ee-a4f1-bf84786dad43] Refreshing network info cache for port dda94cd6-0a2d-4c6b-bc0e-4fb38129159f {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1031.065387] env[62066]: DEBUG oslo_concurrency.lockutils [None req-775e70f5-1ec0-49af-b6c8-e68074f6f67f tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Lock "06983429-8afe-4bf9-ab65-aa4031568b9c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.603s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1031.239985] env[62066]: DEBUG oslo_vmware.api [None req-d0e86307-d1d5-4c1f-9fa1-62e41040f376 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': task-1341339, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.553603] env[62066]: DEBUG nova.network.neutron [req-0def89de-1176-40a0-a364-6e11384c5cc2 req-d5fe8ad7-cde4-43f3-882b-073c8414cc3d service nova] [instance: c8482d92-0180-44ee-a4f1-bf84786dad43] Updated VIF entry in instance network info cache for port dda94cd6-0a2d-4c6b-bc0e-4fb38129159f. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1031.554071] env[62066]: DEBUG nova.network.neutron [req-0def89de-1176-40a0-a364-6e11384c5cc2 req-d5fe8ad7-cde4-43f3-882b-073c8414cc3d service nova] [instance: c8482d92-0180-44ee-a4f1-bf84786dad43] Updating instance_info_cache with network_info: [{"id": "2f2f1786-934d-478f-be39-7509cfacf86f", "address": "fa:16:3e:24:41:2e", "network": {"id": "449dfe4e-bb49-43b6-9ba7-b57af74ebfb3", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-282029676-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.231", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0d4e33a99c7741fb8cdd97f4ec5dbbd8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db68bd64-5b56-49af-a075-13dcf85cb2e0", "external-id": "nsx-vlan-transportzone-590", "segmentation_id": 590, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2f2f1786-93", "ovs_interfaceid": "2f2f1786-934d-478f-be39-7509cfacf86f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "dda94cd6-0a2d-4c6b-bc0e-4fb38129159f", "address": "fa:16:3e:60:d6:92", "network": {"id": "449dfe4e-bb49-43b6-9ba7-b57af74ebfb3", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-282029676-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0d4e33a99c7741fb8cdd97f4ec5dbbd8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db68bd64-5b56-49af-a075-13dcf85cb2e0", "external-id": "nsx-vlan-transportzone-590", "segmentation_id": 590, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdda94cd6-0a", "ovs_interfaceid": "dda94cd6-0a2d-4c6b-bc0e-4fb38129159f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1031.739666] env[62066]: DEBUG oslo_vmware.api [None req-d0e86307-d1d5-4c1f-9fa1-62e41040f376 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': task-1341339, 'name': ReconfigVM_Task, 'duration_secs': 0.621513} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.740171] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d0e86307-d1d5-4c1f-9fa1-62e41040f376 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Releasing lock "c8482d92-0180-44ee-a4f1-bf84786dad43" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1031.740394] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-d0e86307-d1d5-4c1f-9fa1-62e41040f376 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: c8482d92-0180-44ee-a4f1-bf84786dad43] Reconfigured VM to attach interface {{(pid=62066) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 1032.057288] env[62066]: DEBUG oslo_concurrency.lockutils [req-0def89de-1176-40a0-a364-6e11384c5cc2 req-d5fe8ad7-cde4-43f3-882b-073c8414cc3d service nova] Releasing lock "refresh_cache-c8482d92-0180-44ee-a4f1-bf84786dad43" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1032.246270] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d0e86307-d1d5-4c1f-9fa1-62e41040f376 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Lock "interface-c8482d92-0180-44ee-a4f1-bf84786dad43-dda94cd6-0a2d-4c6b-bc0e-4fb38129159f" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 6.264s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1032.473074] env[62066]: DEBUG oslo_concurrency.lockutils [None req-dcc1833f-1bf8-4a90-a602-d99226e73a39 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Acquiring lock "e5f06116-2f49-4cef-8755-454ce36f163d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1032.473329] env[62066]: DEBUG oslo_concurrency.lockutils [None req-dcc1833f-1bf8-4a90-a602-d99226e73a39 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Lock "e5f06116-2f49-4cef-8755-454ce36f163d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1032.975585] env[62066]: DEBUG nova.compute.manager [None req-dcc1833f-1bf8-4a90-a602-d99226e73a39 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: e5f06116-2f49-4cef-8755-454ce36f163d] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1033.496808] env[62066]: DEBUG oslo_concurrency.lockutils [None req-dcc1833f-1bf8-4a90-a602-d99226e73a39 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1033.497091] env[62066]: DEBUG oslo_concurrency.lockutils [None req-dcc1833f-1bf8-4a90-a602-d99226e73a39 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1033.498614] env[62066]: INFO nova.compute.claims [None req-dcc1833f-1bf8-4a90-a602-d99226e73a39 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: e5f06116-2f49-4cef-8755-454ce36f163d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1033.670235] env[62066]: DEBUG oslo_concurrency.lockutils [None req-731bf532-59a8-46e0-b4de-b41dad6bd527 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Acquiring lock "interface-c8482d92-0180-44ee-a4f1-bf84786dad43-dda94cd6-0a2d-4c6b-bc0e-4fb38129159f" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1033.670485] env[62066]: DEBUG oslo_concurrency.lockutils [None req-731bf532-59a8-46e0-b4de-b41dad6bd527 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Lock "interface-c8482d92-0180-44ee-a4f1-bf84786dad43-dda94cd6-0a2d-4c6b-bc0e-4fb38129159f" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1034.173973] env[62066]: DEBUG oslo_concurrency.lockutils [None req-731bf532-59a8-46e0-b4de-b41dad6bd527 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Acquiring lock "c8482d92-0180-44ee-a4f1-bf84786dad43" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1034.174301] env[62066]: DEBUG oslo_concurrency.lockutils [None req-731bf532-59a8-46e0-b4de-b41dad6bd527 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Acquired lock "c8482d92-0180-44ee-a4f1-bf84786dad43" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1034.175016] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df75277e-a061-4e0c-bea8-bf77ec201501 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.192820] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd32405e-0391-4e81-b43b-d9e445e73cdd {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.218844] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-731bf532-59a8-46e0-b4de-b41dad6bd527 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: c8482d92-0180-44ee-a4f1-bf84786dad43] Reconfiguring VM to detach interface {{(pid=62066) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 1034.219143] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bb7abbde-3049-4633-b15a-2e94599b3d72 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.237147] env[62066]: DEBUG oslo_vmware.api [None req-731bf532-59a8-46e0-b4de-b41dad6bd527 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Waiting for the task: (returnval){ [ 1034.237147] env[62066]: value = "task-1341340" [ 1034.237147] env[62066]: _type = "Task" [ 1034.237147] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.244673] env[62066]: DEBUG oslo_vmware.api [None req-731bf532-59a8-46e0-b4de-b41dad6bd527 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': task-1341340, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.566339] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e981655b-f349-41c5-a6a1-80b39debda43 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.574812] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3fc25a8-9392-4e92-89b3-dc7a0d914e43 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.603892] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45994434-3fa8-4889-bd0c-0985f7b759f2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.611186] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a867bfa-17c5-434c-a323-52779a5e54b0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.624063] env[62066]: DEBUG nova.compute.provider_tree [None req-dcc1833f-1bf8-4a90-a602-d99226e73a39 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1034.747160] env[62066]: DEBUG oslo_vmware.api [None req-731bf532-59a8-46e0-b4de-b41dad6bd527 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': task-1341340, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.126961] env[62066]: DEBUG nova.scheduler.client.report [None req-dcc1833f-1bf8-4a90-a602-d99226e73a39 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1035.247683] env[62066]: DEBUG oslo_vmware.api [None req-731bf532-59a8-46e0-b4de-b41dad6bd527 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': task-1341340, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.632026] env[62066]: DEBUG oslo_concurrency.lockutils [None req-dcc1833f-1bf8-4a90-a602-d99226e73a39 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.135s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1035.632641] env[62066]: DEBUG nova.compute.manager [None req-dcc1833f-1bf8-4a90-a602-d99226e73a39 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: e5f06116-2f49-4cef-8755-454ce36f163d] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1035.747525] env[62066]: DEBUG oslo_vmware.api [None req-731bf532-59a8-46e0-b4de-b41dad6bd527 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': task-1341340, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.137199] env[62066]: DEBUG nova.compute.utils [None req-dcc1833f-1bf8-4a90-a602-d99226e73a39 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1036.138601] env[62066]: DEBUG nova.compute.manager [None req-dcc1833f-1bf8-4a90-a602-d99226e73a39 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: e5f06116-2f49-4cef-8755-454ce36f163d] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1036.138778] env[62066]: DEBUG nova.network.neutron [None req-dcc1833f-1bf8-4a90-a602-d99226e73a39 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: e5f06116-2f49-4cef-8755-454ce36f163d] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1036.178172] env[62066]: DEBUG nova.policy [None req-dcc1833f-1bf8-4a90-a602-d99226e73a39 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '588e53cee85f4ab484b76e7a59fcbe78', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8475ad5a900548cba568360999c846ea', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 1036.248692] env[62066]: DEBUG oslo_vmware.api [None req-731bf532-59a8-46e0-b4de-b41dad6bd527 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': task-1341340, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.418082] env[62066]: DEBUG nova.network.neutron [None req-dcc1833f-1bf8-4a90-a602-d99226e73a39 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: e5f06116-2f49-4cef-8755-454ce36f163d] Successfully created port: 1bd7f7b9-9fe0-48c2-a958-c75db1378c5d {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1036.642711] env[62066]: DEBUG nova.compute.manager [None req-dcc1833f-1bf8-4a90-a602-d99226e73a39 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: e5f06116-2f49-4cef-8755-454ce36f163d] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1036.749543] env[62066]: DEBUG oslo_vmware.api [None req-731bf532-59a8-46e0-b4de-b41dad6bd527 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': task-1341340, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.249589] env[62066]: DEBUG oslo_vmware.api [None req-731bf532-59a8-46e0-b4de-b41dad6bd527 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': task-1341340, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.651922] env[62066]: DEBUG nova.compute.manager [None req-dcc1833f-1bf8-4a90-a602-d99226e73a39 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: e5f06116-2f49-4cef-8755-454ce36f163d] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1037.677480] env[62066]: DEBUG nova.virt.hardware [None req-dcc1833f-1bf8-4a90-a602-d99226e73a39 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-23T13:40:41Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-23T13:40:23Z,direct_url=,disk_format='vmdk',id=50ff584c-3b50-4395-af07-3e66769bc9f7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='eb52f7069a374c61ae946f052007c6d9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-23T13:40:24Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1037.677754] env[62066]: DEBUG nova.virt.hardware [None req-dcc1833f-1bf8-4a90-a602-d99226e73a39 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1037.677916] env[62066]: DEBUG nova.virt.hardware [None req-dcc1833f-1bf8-4a90-a602-d99226e73a39 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1037.678124] env[62066]: DEBUG nova.virt.hardware [None req-dcc1833f-1bf8-4a90-a602-d99226e73a39 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1037.678314] env[62066]: DEBUG nova.virt.hardware [None req-dcc1833f-1bf8-4a90-a602-d99226e73a39 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1037.678488] env[62066]: DEBUG nova.virt.hardware [None req-dcc1833f-1bf8-4a90-a602-d99226e73a39 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1037.678707] env[62066]: DEBUG nova.virt.hardware [None req-dcc1833f-1bf8-4a90-a602-d99226e73a39 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1037.678873] env[62066]: DEBUG nova.virt.hardware [None req-dcc1833f-1bf8-4a90-a602-d99226e73a39 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1037.679058] env[62066]: DEBUG nova.virt.hardware [None req-dcc1833f-1bf8-4a90-a602-d99226e73a39 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1037.679232] env[62066]: DEBUG nova.virt.hardware [None req-dcc1833f-1bf8-4a90-a602-d99226e73a39 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1037.679408] env[62066]: DEBUG nova.virt.hardware [None req-dcc1833f-1bf8-4a90-a602-d99226e73a39 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1037.680297] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-223c27a1-6a36-4577-9a28-a9993946d285 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.688546] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb9b13d3-8173-4232-8434-3d2cbaa14360 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.749201] env[62066]: DEBUG oslo_vmware.api [None req-731bf532-59a8-46e0-b4de-b41dad6bd527 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': task-1341340, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.787287] env[62066]: DEBUG nova.compute.manager [req-ba187a14-93a8-4885-819d-b1761f531da6 req-16dd5f8f-8e5a-4aeb-9395-9fb6aadf66d1 service nova] [instance: e5f06116-2f49-4cef-8755-454ce36f163d] Received event network-vif-plugged-1bd7f7b9-9fe0-48c2-a958-c75db1378c5d {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1037.787510] env[62066]: DEBUG oslo_concurrency.lockutils [req-ba187a14-93a8-4885-819d-b1761f531da6 req-16dd5f8f-8e5a-4aeb-9395-9fb6aadf66d1 service nova] Acquiring lock "e5f06116-2f49-4cef-8755-454ce36f163d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1037.787725] env[62066]: DEBUG oslo_concurrency.lockutils [req-ba187a14-93a8-4885-819d-b1761f531da6 req-16dd5f8f-8e5a-4aeb-9395-9fb6aadf66d1 service nova] Lock "e5f06116-2f49-4cef-8755-454ce36f163d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1037.787899] env[62066]: DEBUG oslo_concurrency.lockutils [req-ba187a14-93a8-4885-819d-b1761f531da6 req-16dd5f8f-8e5a-4aeb-9395-9fb6aadf66d1 service nova] Lock "e5f06116-2f49-4cef-8755-454ce36f163d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1037.788084] env[62066]: DEBUG nova.compute.manager [req-ba187a14-93a8-4885-819d-b1761f531da6 req-16dd5f8f-8e5a-4aeb-9395-9fb6aadf66d1 service nova] [instance: e5f06116-2f49-4cef-8755-454ce36f163d] No waiting events found dispatching network-vif-plugged-1bd7f7b9-9fe0-48c2-a958-c75db1378c5d {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1037.788259] env[62066]: WARNING nova.compute.manager [req-ba187a14-93a8-4885-819d-b1761f531da6 req-16dd5f8f-8e5a-4aeb-9395-9fb6aadf66d1 service nova] [instance: e5f06116-2f49-4cef-8755-454ce36f163d] Received unexpected event network-vif-plugged-1bd7f7b9-9fe0-48c2-a958-c75db1378c5d for instance with vm_state building and task_state spawning. [ 1037.869785] env[62066]: DEBUG nova.network.neutron [None req-dcc1833f-1bf8-4a90-a602-d99226e73a39 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: e5f06116-2f49-4cef-8755-454ce36f163d] Successfully updated port: 1bd7f7b9-9fe0-48c2-a958-c75db1378c5d {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1038.250797] env[62066]: DEBUG oslo_vmware.api [None req-731bf532-59a8-46e0-b4de-b41dad6bd527 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': task-1341340, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.375271] env[62066]: DEBUG oslo_concurrency.lockutils [None req-dcc1833f-1bf8-4a90-a602-d99226e73a39 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Acquiring lock "refresh_cache-e5f06116-2f49-4cef-8755-454ce36f163d" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1038.375438] env[62066]: DEBUG oslo_concurrency.lockutils [None req-dcc1833f-1bf8-4a90-a602-d99226e73a39 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Acquired lock "refresh_cache-e5f06116-2f49-4cef-8755-454ce36f163d" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1038.375589] env[62066]: DEBUG nova.network.neutron [None req-dcc1833f-1bf8-4a90-a602-d99226e73a39 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: e5f06116-2f49-4cef-8755-454ce36f163d] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1038.751597] env[62066]: DEBUG oslo_vmware.api [None req-731bf532-59a8-46e0-b4de-b41dad6bd527 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': task-1341340, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.906544] env[62066]: DEBUG nova.network.neutron [None req-dcc1833f-1bf8-4a90-a602-d99226e73a39 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: e5f06116-2f49-4cef-8755-454ce36f163d] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1039.037254] env[62066]: DEBUG nova.network.neutron [None req-dcc1833f-1bf8-4a90-a602-d99226e73a39 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: e5f06116-2f49-4cef-8755-454ce36f163d] Updating instance_info_cache with network_info: [{"id": "1bd7f7b9-9fe0-48c2-a958-c75db1378c5d", "address": "fa:16:3e:5a:8c:ec", "network": {"id": "7617f887-bdd0-400a-8b5f-606d7785df78", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-558766873-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8475ad5a900548cba568360999c846ea", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2b6a4065-12af-4fb9-ac47-ec9143f7297e", "external-id": "nsx-vlan-transportzone-95", "segmentation_id": 95, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1bd7f7b9-9f", "ovs_interfaceid": "1bd7f7b9-9fe0-48c2-a958-c75db1378c5d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1039.251953] env[62066]: DEBUG oslo_vmware.api [None req-731bf532-59a8-46e0-b4de-b41dad6bd527 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': task-1341340, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.540473] env[62066]: DEBUG oslo_concurrency.lockutils [None req-dcc1833f-1bf8-4a90-a602-d99226e73a39 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Releasing lock "refresh_cache-e5f06116-2f49-4cef-8755-454ce36f163d" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1039.540811] env[62066]: DEBUG nova.compute.manager [None req-dcc1833f-1bf8-4a90-a602-d99226e73a39 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: e5f06116-2f49-4cef-8755-454ce36f163d] Instance network_info: |[{"id": "1bd7f7b9-9fe0-48c2-a958-c75db1378c5d", "address": "fa:16:3e:5a:8c:ec", "network": {"id": "7617f887-bdd0-400a-8b5f-606d7785df78", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-558766873-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8475ad5a900548cba568360999c846ea", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2b6a4065-12af-4fb9-ac47-ec9143f7297e", "external-id": "nsx-vlan-transportzone-95", "segmentation_id": 95, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1bd7f7b9-9f", "ovs_interfaceid": "1bd7f7b9-9fe0-48c2-a958-c75db1378c5d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1039.541379] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-dcc1833f-1bf8-4a90-a602-d99226e73a39 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: e5f06116-2f49-4cef-8755-454ce36f163d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:5a:8c:ec', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '2b6a4065-12af-4fb9-ac47-ec9143f7297e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1bd7f7b9-9fe0-48c2-a958-c75db1378c5d', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1039.548823] env[62066]: DEBUG oslo.service.loopingcall [None req-dcc1833f-1bf8-4a90-a602-d99226e73a39 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1039.549077] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e5f06116-2f49-4cef-8755-454ce36f163d] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1039.549313] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-dc50f6c0-d1fc-4fcd-be9a-fcacc60ecf04 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.568979] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1039.568979] env[62066]: value = "task-1341341" [ 1039.568979] env[62066]: _type = "Task" [ 1039.568979] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1039.576249] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1341341, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.752711] env[62066]: DEBUG oslo_vmware.api [None req-731bf532-59a8-46e0-b4de-b41dad6bd527 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': task-1341340, 'name': ReconfigVM_Task} progress is 18%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.813726] env[62066]: DEBUG nova.compute.manager [req-8c3cc490-9fa4-47df-b91c-e05d8bb35436 req-711b57ec-2582-47f1-b16b-7622cf6a84a5 service nova] [instance: e5f06116-2f49-4cef-8755-454ce36f163d] Received event network-changed-1bd7f7b9-9fe0-48c2-a958-c75db1378c5d {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1039.813938] env[62066]: DEBUG nova.compute.manager [req-8c3cc490-9fa4-47df-b91c-e05d8bb35436 req-711b57ec-2582-47f1-b16b-7622cf6a84a5 service nova] [instance: e5f06116-2f49-4cef-8755-454ce36f163d] Refreshing instance network info cache due to event network-changed-1bd7f7b9-9fe0-48c2-a958-c75db1378c5d. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1039.814202] env[62066]: DEBUG oslo_concurrency.lockutils [req-8c3cc490-9fa4-47df-b91c-e05d8bb35436 req-711b57ec-2582-47f1-b16b-7622cf6a84a5 service nova] Acquiring lock "refresh_cache-e5f06116-2f49-4cef-8755-454ce36f163d" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1039.814377] env[62066]: DEBUG oslo_concurrency.lockutils [req-8c3cc490-9fa4-47df-b91c-e05d8bb35436 req-711b57ec-2582-47f1-b16b-7622cf6a84a5 service nova] Acquired lock "refresh_cache-e5f06116-2f49-4cef-8755-454ce36f163d" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1039.814550] env[62066]: DEBUG nova.network.neutron [req-8c3cc490-9fa4-47df-b91c-e05d8bb35436 req-711b57ec-2582-47f1-b16b-7622cf6a84a5 service nova] [instance: e5f06116-2f49-4cef-8755-454ce36f163d] Refreshing network info cache for port 1bd7f7b9-9fe0-48c2-a958-c75db1378c5d {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1040.079137] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1341341, 'name': CreateVM_Task, 'duration_secs': 0.317563} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1040.079295] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e5f06116-2f49-4cef-8755-454ce36f163d] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1040.079977] env[62066]: DEBUG oslo_concurrency.lockutils [None req-dcc1833f-1bf8-4a90-a602-d99226e73a39 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1040.080167] env[62066]: DEBUG oslo_concurrency.lockutils [None req-dcc1833f-1bf8-4a90-a602-d99226e73a39 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Acquired lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1040.080576] env[62066]: DEBUG oslo_concurrency.lockutils [None req-dcc1833f-1bf8-4a90-a602-d99226e73a39 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1040.080837] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4a919592-0d2a-42c2-a011-a43df306e640 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.085353] env[62066]: DEBUG oslo_vmware.api [None req-dcc1833f-1bf8-4a90-a602-d99226e73a39 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Waiting for the task: (returnval){ [ 1040.085353] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]52b33726-2679-be4e-0a65-f3a7843f018a" [ 1040.085353] env[62066]: _type = "Task" [ 1040.085353] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1040.092367] env[62066]: DEBUG oslo_vmware.api [None req-dcc1833f-1bf8-4a90-a602-d99226e73a39 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52b33726-2679-be4e-0a65-f3a7843f018a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.253208] env[62066]: DEBUG oslo_vmware.api [None req-731bf532-59a8-46e0-b4de-b41dad6bd527 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': task-1341340, 'name': ReconfigVM_Task, 'duration_secs': 5.740678} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1040.253557] env[62066]: DEBUG oslo_concurrency.lockutils [None req-731bf532-59a8-46e0-b4de-b41dad6bd527 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Releasing lock "c8482d92-0180-44ee-a4f1-bf84786dad43" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1040.253604] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-731bf532-59a8-46e0-b4de-b41dad6bd527 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: c8482d92-0180-44ee-a4f1-bf84786dad43] Reconfigured VM to detach interface {{(pid=62066) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 1040.596693] env[62066]: DEBUG oslo_vmware.api [None req-dcc1833f-1bf8-4a90-a602-d99226e73a39 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52b33726-2679-be4e-0a65-f3a7843f018a, 'name': SearchDatastore_Task, 'duration_secs': 0.008594} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1040.597186] env[62066]: DEBUG oslo_concurrency.lockutils [None req-dcc1833f-1bf8-4a90-a602-d99226e73a39 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Releasing lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1040.597361] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-dcc1833f-1bf8-4a90-a602-d99226e73a39 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: e5f06116-2f49-4cef-8755-454ce36f163d] Processing image 50ff584c-3b50-4395-af07-3e66769bc9f7 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1040.597602] env[62066]: DEBUG oslo_concurrency.lockutils [None req-dcc1833f-1bf8-4a90-a602-d99226e73a39 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1040.597755] env[62066]: DEBUG oslo_concurrency.lockutils [None req-dcc1833f-1bf8-4a90-a602-d99226e73a39 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Acquired lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1040.597942] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-dcc1833f-1bf8-4a90-a602-d99226e73a39 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1040.598255] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-879aba11-a6d5-41da-b4d9-8c03de7cb0e3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.606773] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-dcc1833f-1bf8-4a90-a602-d99226e73a39 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1040.606956] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-dcc1833f-1bf8-4a90-a602-d99226e73a39 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1040.608031] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ce7c1d41-b276-4bb7-9674-fb83d3ed6c00 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.612766] env[62066]: DEBUG oslo_vmware.api [None req-dcc1833f-1bf8-4a90-a602-d99226e73a39 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Waiting for the task: (returnval){ [ 1040.612766] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]526ad9ab-1d15-6d62-8da7-ba2b2adbe1e3" [ 1040.612766] env[62066]: _type = "Task" [ 1040.612766] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1040.620125] env[62066]: DEBUG oslo_vmware.api [None req-dcc1833f-1bf8-4a90-a602-d99226e73a39 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]526ad9ab-1d15-6d62-8da7-ba2b2adbe1e3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.643107] env[62066]: DEBUG nova.network.neutron [req-8c3cc490-9fa4-47df-b91c-e05d8bb35436 req-711b57ec-2582-47f1-b16b-7622cf6a84a5 service nova] [instance: e5f06116-2f49-4cef-8755-454ce36f163d] Updated VIF entry in instance network info cache for port 1bd7f7b9-9fe0-48c2-a958-c75db1378c5d. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1040.643461] env[62066]: DEBUG nova.network.neutron [req-8c3cc490-9fa4-47df-b91c-e05d8bb35436 req-711b57ec-2582-47f1-b16b-7622cf6a84a5 service nova] [instance: e5f06116-2f49-4cef-8755-454ce36f163d] Updating instance_info_cache with network_info: [{"id": "1bd7f7b9-9fe0-48c2-a958-c75db1378c5d", "address": "fa:16:3e:5a:8c:ec", "network": {"id": "7617f887-bdd0-400a-8b5f-606d7785df78", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-558766873-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8475ad5a900548cba568360999c846ea", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2b6a4065-12af-4fb9-ac47-ec9143f7297e", "external-id": "nsx-vlan-transportzone-95", "segmentation_id": 95, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1bd7f7b9-9f", "ovs_interfaceid": "1bd7f7b9-9fe0-48c2-a958-c75db1378c5d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1041.122795] env[62066]: DEBUG oslo_vmware.api [None req-dcc1833f-1bf8-4a90-a602-d99226e73a39 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]526ad9ab-1d15-6d62-8da7-ba2b2adbe1e3, 'name': SearchDatastore_Task, 'duration_secs': 0.008048} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1041.123697] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d14b0f8b-3f11-4603-96ef-559bbb8f1558 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.128990] env[62066]: DEBUG oslo_vmware.api [None req-dcc1833f-1bf8-4a90-a602-d99226e73a39 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Waiting for the task: (returnval){ [ 1041.128990] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]52382e71-77b9-fd52-abb8-9732c9a7efa9" [ 1041.128990] env[62066]: _type = "Task" [ 1041.128990] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1041.136408] env[62066]: DEBUG oslo_vmware.api [None req-dcc1833f-1bf8-4a90-a602-d99226e73a39 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52382e71-77b9-fd52-abb8-9732c9a7efa9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.146056] env[62066]: DEBUG oslo_concurrency.lockutils [req-8c3cc490-9fa4-47df-b91c-e05d8bb35436 req-711b57ec-2582-47f1-b16b-7622cf6a84a5 service nova] Releasing lock "refresh_cache-e5f06116-2f49-4cef-8755-454ce36f163d" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1041.639822] env[62066]: DEBUG oslo_vmware.api [None req-dcc1833f-1bf8-4a90-a602-d99226e73a39 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52382e71-77b9-fd52-abb8-9732c9a7efa9, 'name': SearchDatastore_Task, 'duration_secs': 0.008789} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1041.639822] env[62066]: DEBUG oslo_concurrency.lockutils [None req-dcc1833f-1bf8-4a90-a602-d99226e73a39 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Releasing lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1041.639822] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-dcc1833f-1bf8-4a90-a602-d99226e73a39 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk to [datastore2] e5f06116-2f49-4cef-8755-454ce36f163d/e5f06116-2f49-4cef-8755-454ce36f163d.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1041.640236] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-006f05aa-f60f-495c-a243-da91618ac89a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.646791] env[62066]: DEBUG oslo_vmware.api [None req-dcc1833f-1bf8-4a90-a602-d99226e73a39 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Waiting for the task: (returnval){ [ 1041.646791] env[62066]: value = "task-1341342" [ 1041.646791] env[62066]: _type = "Task" [ 1041.646791] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1041.654223] env[62066]: DEBUG oslo_vmware.api [None req-dcc1833f-1bf8-4a90-a602-d99226e73a39 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341342, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.695071] env[62066]: DEBUG oslo_concurrency.lockutils [None req-731bf532-59a8-46e0-b4de-b41dad6bd527 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Acquiring lock "refresh_cache-c8482d92-0180-44ee-a4f1-bf84786dad43" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1041.695234] env[62066]: DEBUG oslo_concurrency.lockutils [None req-731bf532-59a8-46e0-b4de-b41dad6bd527 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Acquired lock "refresh_cache-c8482d92-0180-44ee-a4f1-bf84786dad43" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1041.695423] env[62066]: DEBUG nova.network.neutron [None req-731bf532-59a8-46e0-b4de-b41dad6bd527 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: c8482d92-0180-44ee-a4f1-bf84786dad43] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1042.156118] env[62066]: DEBUG oslo_vmware.api [None req-dcc1833f-1bf8-4a90-a602-d99226e73a39 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341342, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.452433} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1042.156449] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-dcc1833f-1bf8-4a90-a602-d99226e73a39 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk to [datastore2] e5f06116-2f49-4cef-8755-454ce36f163d/e5f06116-2f49-4cef-8755-454ce36f163d.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1042.156705] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-dcc1833f-1bf8-4a90-a602-d99226e73a39 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: e5f06116-2f49-4cef-8755-454ce36f163d] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1042.156963] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-59ff3b80-c42d-4b9d-a653-8cf06ba4cc32 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.163432] env[62066]: DEBUG oslo_vmware.api [None req-dcc1833f-1bf8-4a90-a602-d99226e73a39 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Waiting for the task: (returnval){ [ 1042.163432] env[62066]: value = "task-1341343" [ 1042.163432] env[62066]: _type = "Task" [ 1042.163432] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1042.170969] env[62066]: DEBUG oslo_vmware.api [None req-dcc1833f-1bf8-4a90-a602-d99226e73a39 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341343, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.388649] env[62066]: INFO nova.network.neutron [None req-731bf532-59a8-46e0-b4de-b41dad6bd527 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: c8482d92-0180-44ee-a4f1-bf84786dad43] Port dda94cd6-0a2d-4c6b-bc0e-4fb38129159f from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 1042.389069] env[62066]: DEBUG nova.network.neutron [None req-731bf532-59a8-46e0-b4de-b41dad6bd527 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: c8482d92-0180-44ee-a4f1-bf84786dad43] Updating instance_info_cache with network_info: [{"id": "2f2f1786-934d-478f-be39-7509cfacf86f", "address": "fa:16:3e:24:41:2e", "network": {"id": "449dfe4e-bb49-43b6-9ba7-b57af74ebfb3", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-282029676-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0d4e33a99c7741fb8cdd97f4ec5dbbd8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db68bd64-5b56-49af-a075-13dcf85cb2e0", "external-id": "nsx-vlan-transportzone-590", "segmentation_id": 590, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2f2f1786-93", "ovs_interfaceid": "2f2f1786-934d-478f-be39-7509cfacf86f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1042.436656] env[62066]: DEBUG nova.compute.manager [req-cabbb501-a107-41b0-aec7-0487ef3c8ef3 req-19c58bc1-1cbe-4769-aeef-c562796c58f8 service nova] [instance: c8482d92-0180-44ee-a4f1-bf84786dad43] Received event network-changed-2f2f1786-934d-478f-be39-7509cfacf86f {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1042.436835] env[62066]: DEBUG nova.compute.manager [req-cabbb501-a107-41b0-aec7-0487ef3c8ef3 req-19c58bc1-1cbe-4769-aeef-c562796c58f8 service nova] [instance: c8482d92-0180-44ee-a4f1-bf84786dad43] Refreshing instance network info cache due to event network-changed-2f2f1786-934d-478f-be39-7509cfacf86f. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1042.437055] env[62066]: DEBUG oslo_concurrency.lockutils [req-cabbb501-a107-41b0-aec7-0487ef3c8ef3 req-19c58bc1-1cbe-4769-aeef-c562796c58f8 service nova] Acquiring lock "refresh_cache-c8482d92-0180-44ee-a4f1-bf84786dad43" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1042.673165] env[62066]: DEBUG oslo_vmware.api [None req-dcc1833f-1bf8-4a90-a602-d99226e73a39 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341343, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068339} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1042.673548] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-dcc1833f-1bf8-4a90-a602-d99226e73a39 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: e5f06116-2f49-4cef-8755-454ce36f163d] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1042.674273] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db5d865b-71cf-4179-bc1f-f8c0607a3b21 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.696921] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-dcc1833f-1bf8-4a90-a602-d99226e73a39 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: e5f06116-2f49-4cef-8755-454ce36f163d] Reconfiguring VM instance instance-00000068 to attach disk [datastore2] e5f06116-2f49-4cef-8755-454ce36f163d/e5f06116-2f49-4cef-8755-454ce36f163d.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1042.697159] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b9b310d7-a956-468b-89d9-064bf772a845 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.716350] env[62066]: DEBUG oslo_vmware.api [None req-dcc1833f-1bf8-4a90-a602-d99226e73a39 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Waiting for the task: (returnval){ [ 1042.716350] env[62066]: value = "task-1341344" [ 1042.716350] env[62066]: _type = "Task" [ 1042.716350] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1042.723722] env[62066]: DEBUG oslo_vmware.api [None req-dcc1833f-1bf8-4a90-a602-d99226e73a39 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341344, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.891489] env[62066]: DEBUG oslo_concurrency.lockutils [None req-731bf532-59a8-46e0-b4de-b41dad6bd527 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Releasing lock "refresh_cache-c8482d92-0180-44ee-a4f1-bf84786dad43" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1042.894097] env[62066]: DEBUG oslo_concurrency.lockutils [req-cabbb501-a107-41b0-aec7-0487ef3c8ef3 req-19c58bc1-1cbe-4769-aeef-c562796c58f8 service nova] Acquired lock "refresh_cache-c8482d92-0180-44ee-a4f1-bf84786dad43" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1042.894326] env[62066]: DEBUG nova.network.neutron [req-cabbb501-a107-41b0-aec7-0487ef3c8ef3 req-19c58bc1-1cbe-4769-aeef-c562796c58f8 service nova] [instance: c8482d92-0180-44ee-a4f1-bf84786dad43] Refreshing network info cache for port 2f2f1786-934d-478f-be39-7509cfacf86f {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1043.104322] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f5aa422b-966c-4ca1-829d-4066608698d0 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Acquiring lock "interface-7178eb83-41e7-4516-809d-a0972d7e6e23-dda94cd6-0a2d-4c6b-bc0e-4fb38129159f" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1043.104322] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f5aa422b-966c-4ca1-829d-4066608698d0 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Lock "interface-7178eb83-41e7-4516-809d-a0972d7e6e23-dda94cd6-0a2d-4c6b-bc0e-4fb38129159f" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1043.104564] env[62066]: DEBUG nova.objects.instance [None req-f5aa422b-966c-4ca1-829d-4066608698d0 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Lazy-loading 'flavor' on Instance uuid 7178eb83-41e7-4516-809d-a0972d7e6e23 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1043.225588] env[62066]: DEBUG oslo_vmware.api [None req-dcc1833f-1bf8-4a90-a602-d99226e73a39 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341344, 'name': ReconfigVM_Task, 'duration_secs': 0.271517} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1043.225865] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-dcc1833f-1bf8-4a90-a602-d99226e73a39 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: e5f06116-2f49-4cef-8755-454ce36f163d] Reconfigured VM instance instance-00000068 to attach disk [datastore2] e5f06116-2f49-4cef-8755-454ce36f163d/e5f06116-2f49-4cef-8755-454ce36f163d.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1043.226502] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-75416a45-9c73-4fd3-ab31-19a0a1f1f761 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.232887] env[62066]: DEBUG oslo_vmware.api [None req-dcc1833f-1bf8-4a90-a602-d99226e73a39 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Waiting for the task: (returnval){ [ 1043.232887] env[62066]: value = "task-1341345" [ 1043.232887] env[62066]: _type = "Task" [ 1043.232887] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1043.240321] env[62066]: DEBUG oslo_vmware.api [None req-dcc1833f-1bf8-4a90-a602-d99226e73a39 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341345, 'name': Rename_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.397595] env[62066]: DEBUG oslo_concurrency.lockutils [None req-731bf532-59a8-46e0-b4de-b41dad6bd527 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Lock "interface-c8482d92-0180-44ee-a4f1-bf84786dad43-dda94cd6-0a2d-4c6b-bc0e-4fb38129159f" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 9.727s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1043.597476] env[62066]: DEBUG nova.network.neutron [req-cabbb501-a107-41b0-aec7-0487ef3c8ef3 req-19c58bc1-1cbe-4769-aeef-c562796c58f8 service nova] [instance: c8482d92-0180-44ee-a4f1-bf84786dad43] Updated VIF entry in instance network info cache for port 2f2f1786-934d-478f-be39-7509cfacf86f. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1043.597852] env[62066]: DEBUG nova.network.neutron [req-cabbb501-a107-41b0-aec7-0487ef3c8ef3 req-19c58bc1-1cbe-4769-aeef-c562796c58f8 service nova] [instance: c8482d92-0180-44ee-a4f1-bf84786dad43] Updating instance_info_cache with network_info: [{"id": "2f2f1786-934d-478f-be39-7509cfacf86f", "address": "fa:16:3e:24:41:2e", "network": {"id": "449dfe4e-bb49-43b6-9ba7-b57af74ebfb3", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-282029676-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0d4e33a99c7741fb8cdd97f4ec5dbbd8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db68bd64-5b56-49af-a075-13dcf85cb2e0", "external-id": "nsx-vlan-transportzone-590", "segmentation_id": 590, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2f2f1786-93", "ovs_interfaceid": "2f2f1786-934d-478f-be39-7509cfacf86f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1043.695070] env[62066]: DEBUG nova.objects.instance [None req-f5aa422b-966c-4ca1-829d-4066608698d0 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Lazy-loading 'pci_requests' on Instance uuid 7178eb83-41e7-4516-809d-a0972d7e6e23 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1043.742398] env[62066]: DEBUG oslo_vmware.api [None req-dcc1833f-1bf8-4a90-a602-d99226e73a39 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341345, 'name': Rename_Task, 'duration_secs': 0.274319} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1043.742654] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-dcc1833f-1bf8-4a90-a602-d99226e73a39 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: e5f06116-2f49-4cef-8755-454ce36f163d] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1043.742897] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f99631de-baad-4326-b3ec-cd09c8a3954a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.749486] env[62066]: DEBUG oslo_vmware.api [None req-dcc1833f-1bf8-4a90-a602-d99226e73a39 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Waiting for the task: (returnval){ [ 1043.749486] env[62066]: value = "task-1341346" [ 1043.749486] env[62066]: _type = "Task" [ 1043.749486] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1043.756916] env[62066]: DEBUG oslo_vmware.api [None req-dcc1833f-1bf8-4a90-a602-d99226e73a39 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341346, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.100271] env[62066]: DEBUG oslo_concurrency.lockutils [req-cabbb501-a107-41b0-aec7-0487ef3c8ef3 req-19c58bc1-1cbe-4769-aeef-c562796c58f8 service nova] Releasing lock "refresh_cache-c8482d92-0180-44ee-a4f1-bf84786dad43" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1044.197355] env[62066]: DEBUG nova.objects.base [None req-f5aa422b-966c-4ca1-829d-4066608698d0 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Object Instance<7178eb83-41e7-4516-809d-a0972d7e6e23> lazy-loaded attributes: flavor,pci_requests {{(pid=62066) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1044.197651] env[62066]: DEBUG nova.network.neutron [None req-f5aa422b-966c-4ca1-829d-4066608698d0 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: 7178eb83-41e7-4516-809d-a0972d7e6e23] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1044.259657] env[62066]: DEBUG oslo_vmware.api [None req-dcc1833f-1bf8-4a90-a602-d99226e73a39 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341346, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.264647] env[62066]: DEBUG nova.policy [None req-f5aa422b-966c-4ca1-829d-4066608698d0 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '747026d272d4467bbf14a0c30589c3b7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0d4e33a99c7741fb8cdd97f4ec5dbbd8', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 1044.462641] env[62066]: DEBUG nova.compute.manager [req-13be28ee-bb99-4fff-9186-c05e38d7c32d req-2206b38c-eaac-4cbd-99cf-2292b338bfe9 service nova] [instance: 7178eb83-41e7-4516-809d-a0972d7e6e23] Received event network-changed-63f16c96-7de0-40e7-9c0c-782122865437 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1044.462848] env[62066]: DEBUG nova.compute.manager [req-13be28ee-bb99-4fff-9186-c05e38d7c32d req-2206b38c-eaac-4cbd-99cf-2292b338bfe9 service nova] [instance: 7178eb83-41e7-4516-809d-a0972d7e6e23] Refreshing instance network info cache due to event network-changed-63f16c96-7de0-40e7-9c0c-782122865437. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1044.463084] env[62066]: DEBUG oslo_concurrency.lockutils [req-13be28ee-bb99-4fff-9186-c05e38d7c32d req-2206b38c-eaac-4cbd-99cf-2292b338bfe9 service nova] Acquiring lock "refresh_cache-7178eb83-41e7-4516-809d-a0972d7e6e23" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1044.463238] env[62066]: DEBUG oslo_concurrency.lockutils [req-13be28ee-bb99-4fff-9186-c05e38d7c32d req-2206b38c-eaac-4cbd-99cf-2292b338bfe9 service nova] Acquired lock "refresh_cache-7178eb83-41e7-4516-809d-a0972d7e6e23" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1044.463409] env[62066]: DEBUG nova.network.neutron [req-13be28ee-bb99-4fff-9186-c05e38d7c32d req-2206b38c-eaac-4cbd-99cf-2292b338bfe9 service nova] [instance: 7178eb83-41e7-4516-809d-a0972d7e6e23] Refreshing network info cache for port 63f16c96-7de0-40e7-9c0c-782122865437 {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1044.759900] env[62066]: DEBUG oslo_vmware.api [None req-dcc1833f-1bf8-4a90-a602-d99226e73a39 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341346, 'name': PowerOnVM_Task, 'duration_secs': 0.604464} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1044.760253] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-dcc1833f-1bf8-4a90-a602-d99226e73a39 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: e5f06116-2f49-4cef-8755-454ce36f163d] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1044.760415] env[62066]: INFO nova.compute.manager [None req-dcc1833f-1bf8-4a90-a602-d99226e73a39 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: e5f06116-2f49-4cef-8755-454ce36f163d] Took 7.11 seconds to spawn the instance on the hypervisor. [ 1044.760603] env[62066]: DEBUG nova.compute.manager [None req-dcc1833f-1bf8-4a90-a602-d99226e73a39 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: e5f06116-2f49-4cef-8755-454ce36f163d] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1044.761395] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0729501a-838f-4dfe-a6ca-ac002dd366c7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.171252] env[62066]: DEBUG nova.network.neutron [req-13be28ee-bb99-4fff-9186-c05e38d7c32d req-2206b38c-eaac-4cbd-99cf-2292b338bfe9 service nova] [instance: 7178eb83-41e7-4516-809d-a0972d7e6e23] Updated VIF entry in instance network info cache for port 63f16c96-7de0-40e7-9c0c-782122865437. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1045.171627] env[62066]: DEBUG nova.network.neutron [req-13be28ee-bb99-4fff-9186-c05e38d7c32d req-2206b38c-eaac-4cbd-99cf-2292b338bfe9 service nova] [instance: 7178eb83-41e7-4516-809d-a0972d7e6e23] Updating instance_info_cache with network_info: [{"id": "63f16c96-7de0-40e7-9c0c-782122865437", "address": "fa:16:3e:c6:a7:3e", "network": {"id": "449dfe4e-bb49-43b6-9ba7-b57af74ebfb3", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-282029676-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.231", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0d4e33a99c7741fb8cdd97f4ec5dbbd8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db68bd64-5b56-49af-a075-13dcf85cb2e0", "external-id": "nsx-vlan-transportzone-590", "segmentation_id": 590, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap63f16c96-7d", "ovs_interfaceid": "63f16c96-7de0-40e7-9c0c-782122865437", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1045.276527] env[62066]: INFO nova.compute.manager [None req-dcc1833f-1bf8-4a90-a602-d99226e73a39 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: e5f06116-2f49-4cef-8755-454ce36f163d] Took 11.80 seconds to build instance. [ 1045.478439] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d86c50e-9c04-4d87-84dd-7a9a33b75aa8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.485595] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-76862022-2e28-4927-ada4-f1c8dc91a3ee tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: e5f06116-2f49-4cef-8755-454ce36f163d] Suspending the VM {{(pid=62066) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 1045.485833] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-e67b73be-aa50-4463-b456-891360499cee {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.492807] env[62066]: DEBUG oslo_vmware.api [None req-76862022-2e28-4927-ada4-f1c8dc91a3ee tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Waiting for the task: (returnval){ [ 1045.492807] env[62066]: value = "task-1341347" [ 1045.492807] env[62066]: _type = "Task" [ 1045.492807] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.500401] env[62066]: DEBUG oslo_vmware.api [None req-76862022-2e28-4927-ada4-f1c8dc91a3ee tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341347, 'name': SuspendVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.645716] env[62066]: DEBUG nova.compute.manager [req-9b5c3000-623b-47ae-8d72-75e1bf8594f2 req-902becea-d836-4ccc-96da-f9075eda10cd service nova] [instance: 7178eb83-41e7-4516-809d-a0972d7e6e23] Received event network-vif-plugged-dda94cd6-0a2d-4c6b-bc0e-4fb38129159f {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1045.646011] env[62066]: DEBUG oslo_concurrency.lockutils [req-9b5c3000-623b-47ae-8d72-75e1bf8594f2 req-902becea-d836-4ccc-96da-f9075eda10cd service nova] Acquiring lock "7178eb83-41e7-4516-809d-a0972d7e6e23-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1045.646325] env[62066]: DEBUG oslo_concurrency.lockutils [req-9b5c3000-623b-47ae-8d72-75e1bf8594f2 req-902becea-d836-4ccc-96da-f9075eda10cd service nova] Lock "7178eb83-41e7-4516-809d-a0972d7e6e23-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1045.646509] env[62066]: DEBUG oslo_concurrency.lockutils [req-9b5c3000-623b-47ae-8d72-75e1bf8594f2 req-902becea-d836-4ccc-96da-f9075eda10cd service nova] Lock "7178eb83-41e7-4516-809d-a0972d7e6e23-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1045.646709] env[62066]: DEBUG nova.compute.manager [req-9b5c3000-623b-47ae-8d72-75e1bf8594f2 req-902becea-d836-4ccc-96da-f9075eda10cd service nova] [instance: 7178eb83-41e7-4516-809d-a0972d7e6e23] No waiting events found dispatching network-vif-plugged-dda94cd6-0a2d-4c6b-bc0e-4fb38129159f {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1045.646883] env[62066]: WARNING nova.compute.manager [req-9b5c3000-623b-47ae-8d72-75e1bf8594f2 req-902becea-d836-4ccc-96da-f9075eda10cd service nova] [instance: 7178eb83-41e7-4516-809d-a0972d7e6e23] Received unexpected event network-vif-plugged-dda94cd6-0a2d-4c6b-bc0e-4fb38129159f for instance with vm_state active and task_state None. [ 1045.674414] env[62066]: DEBUG oslo_concurrency.lockutils [req-13be28ee-bb99-4fff-9186-c05e38d7c32d req-2206b38c-eaac-4cbd-99cf-2292b338bfe9 service nova] Releasing lock "refresh_cache-7178eb83-41e7-4516-809d-a0972d7e6e23" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1045.734560] env[62066]: DEBUG nova.network.neutron [None req-f5aa422b-966c-4ca1-829d-4066608698d0 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: 7178eb83-41e7-4516-809d-a0972d7e6e23] Successfully updated port: dda94cd6-0a2d-4c6b-bc0e-4fb38129159f {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1045.778686] env[62066]: DEBUG oslo_concurrency.lockutils [None req-dcc1833f-1bf8-4a90-a602-d99226e73a39 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Lock "e5f06116-2f49-4cef-8755-454ce36f163d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.305s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1046.003072] env[62066]: DEBUG oslo_vmware.api [None req-76862022-2e28-4927-ada4-f1c8dc91a3ee tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341347, 'name': SuspendVM_Task} progress is 66%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.237188] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f5aa422b-966c-4ca1-829d-4066608698d0 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Acquiring lock "refresh_cache-7178eb83-41e7-4516-809d-a0972d7e6e23" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1046.237474] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f5aa422b-966c-4ca1-829d-4066608698d0 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Acquired lock "refresh_cache-7178eb83-41e7-4516-809d-a0972d7e6e23" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1046.237781] env[62066]: DEBUG nova.network.neutron [None req-f5aa422b-966c-4ca1-829d-4066608698d0 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: 7178eb83-41e7-4516-809d-a0972d7e6e23] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1046.504144] env[62066]: DEBUG oslo_vmware.api [None req-76862022-2e28-4927-ada4-f1c8dc91a3ee tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341347, 'name': SuspendVM_Task, 'duration_secs': 0.881804} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1046.504750] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-76862022-2e28-4927-ada4-f1c8dc91a3ee tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: e5f06116-2f49-4cef-8755-454ce36f163d] Suspended the VM {{(pid=62066) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 1046.504945] env[62066]: DEBUG nova.compute.manager [None req-76862022-2e28-4927-ada4-f1c8dc91a3ee tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: e5f06116-2f49-4cef-8755-454ce36f163d] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1046.505728] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-953e384d-bab0-4b94-b0f5-fb9dc9038c93 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.773606] env[62066]: WARNING nova.network.neutron [None req-f5aa422b-966c-4ca1-829d-4066608698d0 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: 7178eb83-41e7-4516-809d-a0972d7e6e23] 449dfe4e-bb49-43b6-9ba7-b57af74ebfb3 already exists in list: networks containing: ['449dfe4e-bb49-43b6-9ba7-b57af74ebfb3']. ignoring it [ 1047.036189] env[62066]: DEBUG nova.network.neutron [None req-f5aa422b-966c-4ca1-829d-4066608698d0 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: 7178eb83-41e7-4516-809d-a0972d7e6e23] Updating instance_info_cache with network_info: [{"id": "63f16c96-7de0-40e7-9c0c-782122865437", "address": "fa:16:3e:c6:a7:3e", "network": {"id": "449dfe4e-bb49-43b6-9ba7-b57af74ebfb3", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-282029676-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.231", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0d4e33a99c7741fb8cdd97f4ec5dbbd8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db68bd64-5b56-49af-a075-13dcf85cb2e0", "external-id": "nsx-vlan-transportzone-590", "segmentation_id": 590, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap63f16c96-7d", "ovs_interfaceid": "63f16c96-7de0-40e7-9c0c-782122865437", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "dda94cd6-0a2d-4c6b-bc0e-4fb38129159f", "address": "fa:16:3e:60:d6:92", "network": {"id": "449dfe4e-bb49-43b6-9ba7-b57af74ebfb3", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-282029676-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0d4e33a99c7741fb8cdd97f4ec5dbbd8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db68bd64-5b56-49af-a075-13dcf85cb2e0", "external-id": "nsx-vlan-transportzone-590", "segmentation_id": 590, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdda94cd6-0a", "ovs_interfaceid": "dda94cd6-0a2d-4c6b-bc0e-4fb38129159f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1047.538505] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f5aa422b-966c-4ca1-829d-4066608698d0 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Releasing lock "refresh_cache-7178eb83-41e7-4516-809d-a0972d7e6e23" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1047.539168] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f5aa422b-966c-4ca1-829d-4066608698d0 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Acquiring lock "7178eb83-41e7-4516-809d-a0972d7e6e23" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1047.539359] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f5aa422b-966c-4ca1-829d-4066608698d0 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Acquired lock "7178eb83-41e7-4516-809d-a0972d7e6e23" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1047.540270] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-072a6f86-3539-4888-958f-ae602137f531 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.556498] env[62066]: DEBUG nova.virt.hardware [None req-f5aa422b-966c-4ca1-829d-4066608698d0 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-23T13:40:41Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1047.556711] env[62066]: DEBUG nova.virt.hardware [None req-f5aa422b-966c-4ca1-829d-4066608698d0 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1047.556873] env[62066]: DEBUG nova.virt.hardware [None req-f5aa422b-966c-4ca1-829d-4066608698d0 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1047.557073] env[62066]: DEBUG nova.virt.hardware [None req-f5aa422b-966c-4ca1-829d-4066608698d0 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1047.557226] env[62066]: DEBUG nova.virt.hardware [None req-f5aa422b-966c-4ca1-829d-4066608698d0 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1047.557380] env[62066]: DEBUG nova.virt.hardware [None req-f5aa422b-966c-4ca1-829d-4066608698d0 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1047.557588] env[62066]: DEBUG nova.virt.hardware [None req-f5aa422b-966c-4ca1-829d-4066608698d0 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1047.557750] env[62066]: DEBUG nova.virt.hardware [None req-f5aa422b-966c-4ca1-829d-4066608698d0 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1047.557918] env[62066]: DEBUG nova.virt.hardware [None req-f5aa422b-966c-4ca1-829d-4066608698d0 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1047.558096] env[62066]: DEBUG nova.virt.hardware [None req-f5aa422b-966c-4ca1-829d-4066608698d0 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1047.558280] env[62066]: DEBUG nova.virt.hardware [None req-f5aa422b-966c-4ca1-829d-4066608698d0 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1047.564479] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-f5aa422b-966c-4ca1-829d-4066608698d0 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: 7178eb83-41e7-4516-809d-a0972d7e6e23] Reconfiguring VM to attach interface {{(pid=62066) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 1047.564765] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2b141f9d-2888-4e12-9e61-1c048460166a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.581687] env[62066]: DEBUG oslo_vmware.api [None req-f5aa422b-966c-4ca1-829d-4066608698d0 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Waiting for the task: (returnval){ [ 1047.581687] env[62066]: value = "task-1341348" [ 1047.581687] env[62066]: _type = "Task" [ 1047.581687] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.589080] env[62066]: DEBUG oslo_vmware.api [None req-f5aa422b-966c-4ca1-829d-4066608698d0 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': task-1341348, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.670846] env[62066]: DEBUG nova.compute.manager [req-5ef5e825-0de4-420a-bfcb-86666e3a03c2 req-2f378f20-db78-409e-9223-5329221839cf service nova] [instance: 7178eb83-41e7-4516-809d-a0972d7e6e23] Received event network-changed-dda94cd6-0a2d-4c6b-bc0e-4fb38129159f {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1047.671153] env[62066]: DEBUG nova.compute.manager [req-5ef5e825-0de4-420a-bfcb-86666e3a03c2 req-2f378f20-db78-409e-9223-5329221839cf service nova] [instance: 7178eb83-41e7-4516-809d-a0972d7e6e23] Refreshing instance network info cache due to event network-changed-dda94cd6-0a2d-4c6b-bc0e-4fb38129159f. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1047.671388] env[62066]: DEBUG oslo_concurrency.lockutils [req-5ef5e825-0de4-420a-bfcb-86666e3a03c2 req-2f378f20-db78-409e-9223-5329221839cf service nova] Acquiring lock "refresh_cache-7178eb83-41e7-4516-809d-a0972d7e6e23" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1047.671546] env[62066]: DEBUG oslo_concurrency.lockutils [req-5ef5e825-0de4-420a-bfcb-86666e3a03c2 req-2f378f20-db78-409e-9223-5329221839cf service nova] Acquired lock "refresh_cache-7178eb83-41e7-4516-809d-a0972d7e6e23" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1047.671774] env[62066]: DEBUG nova.network.neutron [req-5ef5e825-0de4-420a-bfcb-86666e3a03c2 req-2f378f20-db78-409e-9223-5329221839cf service nova] [instance: 7178eb83-41e7-4516-809d-a0972d7e6e23] Refreshing network info cache for port dda94cd6-0a2d-4c6b-bc0e-4fb38129159f {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1047.823195] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d297bc9c-670c-49a3-a699-807be5042092 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Acquiring lock "e5f06116-2f49-4cef-8755-454ce36f163d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1047.823446] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d297bc9c-670c-49a3-a699-807be5042092 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Lock "e5f06116-2f49-4cef-8755-454ce36f163d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1047.823638] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d297bc9c-670c-49a3-a699-807be5042092 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Acquiring lock "e5f06116-2f49-4cef-8755-454ce36f163d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1047.823826] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d297bc9c-670c-49a3-a699-807be5042092 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Lock "e5f06116-2f49-4cef-8755-454ce36f163d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1047.824019] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d297bc9c-670c-49a3-a699-807be5042092 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Lock "e5f06116-2f49-4cef-8755-454ce36f163d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1047.826388] env[62066]: INFO nova.compute.manager [None req-d297bc9c-670c-49a3-a699-807be5042092 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: e5f06116-2f49-4cef-8755-454ce36f163d] Terminating instance [ 1047.828124] env[62066]: DEBUG nova.compute.manager [None req-d297bc9c-670c-49a3-a699-807be5042092 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: e5f06116-2f49-4cef-8755-454ce36f163d] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1047.828353] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-d297bc9c-670c-49a3-a699-807be5042092 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: e5f06116-2f49-4cef-8755-454ce36f163d] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1047.829217] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5beddc58-be8f-47a3-a08a-b4a69276a4a0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.836600] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-d297bc9c-670c-49a3-a699-807be5042092 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: e5f06116-2f49-4cef-8755-454ce36f163d] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1047.836834] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d7026c19-c483-4221-9b9e-84fb539eb2fd {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.899390] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-d297bc9c-670c-49a3-a699-807be5042092 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: e5f06116-2f49-4cef-8755-454ce36f163d] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1047.899670] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-d297bc9c-670c-49a3-a699-807be5042092 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: e5f06116-2f49-4cef-8755-454ce36f163d] Deleting contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1047.899992] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-d297bc9c-670c-49a3-a699-807be5042092 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Deleting the datastore file [datastore2] e5f06116-2f49-4cef-8755-454ce36f163d {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1047.900430] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5faddc75-7e63-4b97-84b6-e80d0f257079 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.908494] env[62066]: DEBUG oslo_vmware.api [None req-d297bc9c-670c-49a3-a699-807be5042092 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Waiting for the task: (returnval){ [ 1047.908494] env[62066]: value = "task-1341350" [ 1047.908494] env[62066]: _type = "Task" [ 1047.908494] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.916434] env[62066]: DEBUG oslo_vmware.api [None req-d297bc9c-670c-49a3-a699-807be5042092 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341350, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.091281] env[62066]: DEBUG oslo_vmware.api [None req-f5aa422b-966c-4ca1-829d-4066608698d0 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': task-1341348, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.418698] env[62066]: DEBUG oslo_vmware.api [None req-d297bc9c-670c-49a3-a699-807be5042092 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341350, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.211433} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.420965] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-d297bc9c-670c-49a3-a699-807be5042092 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1048.421221] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-d297bc9c-670c-49a3-a699-807be5042092 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: e5f06116-2f49-4cef-8755-454ce36f163d] Deleted contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1048.421412] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-d297bc9c-670c-49a3-a699-807be5042092 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: e5f06116-2f49-4cef-8755-454ce36f163d] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1048.421595] env[62066]: INFO nova.compute.manager [None req-d297bc9c-670c-49a3-a699-807be5042092 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: e5f06116-2f49-4cef-8755-454ce36f163d] Took 0.59 seconds to destroy the instance on the hypervisor. [ 1048.421839] env[62066]: DEBUG oslo.service.loopingcall [None req-d297bc9c-670c-49a3-a699-807be5042092 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1048.422056] env[62066]: DEBUG nova.compute.manager [-] [instance: e5f06116-2f49-4cef-8755-454ce36f163d] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1048.422160] env[62066]: DEBUG nova.network.neutron [-] [instance: e5f06116-2f49-4cef-8755-454ce36f163d] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1048.592558] env[62066]: DEBUG oslo_vmware.api [None req-f5aa422b-966c-4ca1-829d-4066608698d0 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': task-1341348, 'name': ReconfigVM_Task, 'duration_secs': 0.542145} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.593045] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f5aa422b-966c-4ca1-829d-4066608698d0 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Releasing lock "7178eb83-41e7-4516-809d-a0972d7e6e23" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1048.593266] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-f5aa422b-966c-4ca1-829d-4066608698d0 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: 7178eb83-41e7-4516-809d-a0972d7e6e23] Reconfigured VM to attach interface {{(pid=62066) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 1048.634237] env[62066]: DEBUG nova.network.neutron [req-5ef5e825-0de4-420a-bfcb-86666e3a03c2 req-2f378f20-db78-409e-9223-5329221839cf service nova] [instance: 7178eb83-41e7-4516-809d-a0972d7e6e23] Updated VIF entry in instance network info cache for port dda94cd6-0a2d-4c6b-bc0e-4fb38129159f. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1048.634648] env[62066]: DEBUG nova.network.neutron [req-5ef5e825-0de4-420a-bfcb-86666e3a03c2 req-2f378f20-db78-409e-9223-5329221839cf service nova] [instance: 7178eb83-41e7-4516-809d-a0972d7e6e23] Updating instance_info_cache with network_info: [{"id": "63f16c96-7de0-40e7-9c0c-782122865437", "address": "fa:16:3e:c6:a7:3e", "network": {"id": "449dfe4e-bb49-43b6-9ba7-b57af74ebfb3", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-282029676-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.231", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0d4e33a99c7741fb8cdd97f4ec5dbbd8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db68bd64-5b56-49af-a075-13dcf85cb2e0", "external-id": "nsx-vlan-transportzone-590", "segmentation_id": 590, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap63f16c96-7d", "ovs_interfaceid": "63f16c96-7de0-40e7-9c0c-782122865437", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "dda94cd6-0a2d-4c6b-bc0e-4fb38129159f", "address": "fa:16:3e:60:d6:92", "network": {"id": "449dfe4e-bb49-43b6-9ba7-b57af74ebfb3", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-282029676-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0d4e33a99c7741fb8cdd97f4ec5dbbd8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db68bd64-5b56-49af-a075-13dcf85cb2e0", "external-id": "nsx-vlan-transportzone-590", "segmentation_id": 590, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdda94cd6-0a", "ovs_interfaceid": "dda94cd6-0a2d-4c6b-bc0e-4fb38129159f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1049.098172] env[62066]: DEBUG oslo_concurrency.lockutils [None req-f5aa422b-966c-4ca1-829d-4066608698d0 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Lock "interface-7178eb83-41e7-4516-809d-a0972d7e6e23-dda94cd6-0a2d-4c6b-bc0e-4fb38129159f" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 5.994s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1049.137216] env[62066]: DEBUG oslo_concurrency.lockutils [req-5ef5e825-0de4-420a-bfcb-86666e3a03c2 req-2f378f20-db78-409e-9223-5329221839cf service nova] Releasing lock "refresh_cache-7178eb83-41e7-4516-809d-a0972d7e6e23" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1049.252455] env[62066]: DEBUG nova.network.neutron [-] [instance: e5f06116-2f49-4cef-8755-454ce36f163d] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1049.330375] env[62066]: DEBUG oslo_service.periodic_task [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1049.330601] env[62066]: DEBUG oslo_service.periodic_task [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1049.694569] env[62066]: DEBUG nova.compute.manager [req-8eae5751-e61e-4cc9-b009-873082ddd6a3 req-f8475981-44a3-4491-844f-d38597a37097 service nova] [instance: e5f06116-2f49-4cef-8755-454ce36f163d] Received event network-vif-deleted-1bd7f7b9-9fe0-48c2-a958-c75db1378c5d {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1049.755372] env[62066]: INFO nova.compute.manager [-] [instance: e5f06116-2f49-4cef-8755-454ce36f163d] Took 1.33 seconds to deallocate network for instance. [ 1049.836726] env[62066]: DEBUG oslo_service.periodic_task [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1049.836726] env[62066]: DEBUG oslo_service.periodic_task [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1049.836726] env[62066]: DEBUG oslo_service.periodic_task [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1049.836948] env[62066]: DEBUG oslo_service.periodic_task [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1049.836948] env[62066]: DEBUG nova.compute.manager [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62066) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1049.837071] env[62066]: DEBUG oslo_service.periodic_task [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1049.837204] env[62066]: DEBUG nova.compute.manager [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Cleaning up deleted instances {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11228}} [ 1050.261922] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d297bc9c-670c-49a3-a699-807be5042092 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1050.262320] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d297bc9c-670c-49a3-a699-807be5042092 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1050.262527] env[62066]: DEBUG nova.objects.instance [None req-d297bc9c-670c-49a3-a699-807be5042092 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Lazy-loading 'resources' on Instance uuid e5f06116-2f49-4cef-8755-454ce36f163d {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1050.272627] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1d3c474f-5ef6-4ee0-ab35-0ec48db7e18b tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Acquiring lock "interface-7178eb83-41e7-4516-809d-a0972d7e6e23-dda94cd6-0a2d-4c6b-bc0e-4fb38129159f" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1050.272856] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1d3c474f-5ef6-4ee0-ab35-0ec48db7e18b tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Lock "interface-7178eb83-41e7-4516-809d-a0972d7e6e23-dda94cd6-0a2d-4c6b-bc0e-4fb38129159f" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1050.350243] env[62066]: DEBUG nova.compute.manager [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] There are 45 instances to clean {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11237}} [ 1050.350534] env[62066]: DEBUG nova.compute.manager [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] [instance: 06983429-8afe-4bf9-ab65-aa4031568b9c] Instance has had 0 of 5 cleanup attempts {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1050.775441] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1d3c474f-5ef6-4ee0-ab35-0ec48db7e18b tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Acquiring lock "7178eb83-41e7-4516-809d-a0972d7e6e23" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1050.775633] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1d3c474f-5ef6-4ee0-ab35-0ec48db7e18b tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Acquired lock "7178eb83-41e7-4516-809d-a0972d7e6e23" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1050.776514] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad10f659-c596-4421-8506-0a78389e371a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.797545] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce0dc736-4332-40c7-9b4f-4a4a4008f794 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.822885] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-1d3c474f-5ef6-4ee0-ab35-0ec48db7e18b tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: 7178eb83-41e7-4516-809d-a0972d7e6e23] Reconfiguring VM to detach interface {{(pid=62066) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 1050.825472] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-71a98b03-3083-4b05-88a5-2e43f0ba80e2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.843509] env[62066]: DEBUG oslo_vmware.api [None req-1d3c474f-5ef6-4ee0-ab35-0ec48db7e18b tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Waiting for the task: (returnval){ [ 1050.843509] env[62066]: value = "task-1341351" [ 1050.843509] env[62066]: _type = "Task" [ 1050.843509] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.851421] env[62066]: DEBUG oslo_vmware.api [None req-1d3c474f-5ef6-4ee0-ab35-0ec48db7e18b tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': task-1341351, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.853885] env[62066]: DEBUG nova.compute.manager [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] [instance: 9dc7c179-4360-425d-915f-c2d7bc591b1e] Instance has had 0 of 5 cleanup attempts {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1050.871949] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed831cd2-22b8-4aed-a193-afae873b1666 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.878661] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fec01dd2-0075-47e3-baa6-32d406547f80 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.907290] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee18776f-f728-40a7-8492-3a4957cd73bb {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.913671] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af456a2d-0f13-462f-a2bf-7b0290359f54 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.926153] env[62066]: DEBUG nova.compute.provider_tree [None req-d297bc9c-670c-49a3-a699-807be5042092 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Updating inventory in ProviderTree for provider 8e4f7194-1498-4f08-8723-ab7260524bcb with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1051.355817] env[62066]: DEBUG oslo_vmware.api [None req-1d3c474f-5ef6-4ee0-ab35-0ec48db7e18b tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': task-1341351, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.356244] env[62066]: DEBUG nova.compute.manager [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] [instance: a4ef97ad-bbb5-47fa-b5c1-2867a0d56435] Instance has had 0 of 5 cleanup attempts {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1051.445243] env[62066]: ERROR nova.scheduler.client.report [None req-d297bc9c-670c-49a3-a699-807be5042092 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [req-adb0af23-7a0c-4cf9-8209-e2b2a9876e4c] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 8e4f7194-1498-4f08-8723-ab7260524bcb. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-adb0af23-7a0c-4cf9-8209-e2b2a9876e4c"}]} [ 1051.462522] env[62066]: DEBUG nova.scheduler.client.report [None req-d297bc9c-670c-49a3-a699-807be5042092 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Refreshing inventories for resource provider 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 1051.476303] env[62066]: DEBUG nova.scheduler.client.report [None req-d297bc9c-670c-49a3-a699-807be5042092 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Updating ProviderTree inventory for provider 8e4f7194-1498-4f08-8723-ab7260524bcb from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 1051.476526] env[62066]: DEBUG nova.compute.provider_tree [None req-d297bc9c-670c-49a3-a699-807be5042092 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Updating inventory in ProviderTree for provider 8e4f7194-1498-4f08-8723-ab7260524bcb with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1051.486473] env[62066]: DEBUG nova.scheduler.client.report [None req-d297bc9c-670c-49a3-a699-807be5042092 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Refreshing aggregate associations for resource provider 8e4f7194-1498-4f08-8723-ab7260524bcb, aggregates: None {{(pid=62066) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 1051.504072] env[62066]: DEBUG nova.scheduler.client.report [None req-d297bc9c-670c-49a3-a699-807be5042092 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Refreshing trait associations for resource provider 8e4f7194-1498-4f08-8723-ab7260524bcb, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,HW_ARCH_X86_64,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=62066) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 1051.552832] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8266af2-5afc-4c7a-a1dd-576e783310e9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.560153] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f11f1f87-b514-4fd5-a333-80c55e666c74 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.588938] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60bd57b6-b6d5-47a1-9328-389f3407f7cf {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.595800] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56e00335-c62c-496e-965a-1237b9e10899 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.608527] env[62066]: DEBUG nova.compute.provider_tree [None req-d297bc9c-670c-49a3-a699-807be5042092 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Updating inventory in ProviderTree for provider 8e4f7194-1498-4f08-8723-ab7260524bcb with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1051.853538] env[62066]: DEBUG oslo_vmware.api [None req-1d3c474f-5ef6-4ee0-ab35-0ec48db7e18b tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': task-1341351, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.859094] env[62066]: DEBUG nova.compute.manager [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] [instance: 09e3ae18-1517-470d-8cc9-a8b50ee774c7] Instance has had 0 of 5 cleanup attempts {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1052.137669] env[62066]: DEBUG nova.scheduler.client.report [None req-d297bc9c-670c-49a3-a699-807be5042092 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Updated inventory for provider 8e4f7194-1498-4f08-8723-ab7260524bcb with generation 122 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 1052.137957] env[62066]: DEBUG nova.compute.provider_tree [None req-d297bc9c-670c-49a3-a699-807be5042092 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Updating resource provider 8e4f7194-1498-4f08-8723-ab7260524bcb generation from 122 to 123 during operation: update_inventory {{(pid=62066) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1052.138163] env[62066]: DEBUG nova.compute.provider_tree [None req-d297bc9c-670c-49a3-a699-807be5042092 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Updating inventory in ProviderTree for provider 8e4f7194-1498-4f08-8723-ab7260524bcb with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1052.354694] env[62066]: DEBUG oslo_vmware.api [None req-1d3c474f-5ef6-4ee0-ab35-0ec48db7e18b tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': task-1341351, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.362258] env[62066]: DEBUG nova.compute.manager [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] [instance: ca504be5-209c-456c-af00-d403bfa2d634] Instance has had 0 of 5 cleanup attempts {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1052.642733] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d297bc9c-670c-49a3-a699-807be5042092 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.380s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1052.666828] env[62066]: INFO nova.scheduler.client.report [None req-d297bc9c-670c-49a3-a699-807be5042092 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Deleted allocations for instance e5f06116-2f49-4cef-8755-454ce36f163d [ 1052.854836] env[62066]: DEBUG oslo_vmware.api [None req-1d3c474f-5ef6-4ee0-ab35-0ec48db7e18b tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': task-1341351, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.865471] env[62066]: DEBUG nova.compute.manager [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] [instance: 5693eea2-1cac-45d5-aa27-58526b65136e] Instance has had 0 of 5 cleanup attempts {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1053.174846] env[62066]: DEBUG oslo_concurrency.lockutils [None req-d297bc9c-670c-49a3-a699-807be5042092 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Lock "e5f06116-2f49-4cef-8755-454ce36f163d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.351s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1053.355918] env[62066]: DEBUG oslo_vmware.api [None req-1d3c474f-5ef6-4ee0-ab35-0ec48db7e18b tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': task-1341351, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.368450] env[62066]: DEBUG nova.compute.manager [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] [instance: 651e5142-7ab0-4040-97b4-4599bd4e419a] Instance has had 0 of 5 cleanup attempts {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1053.684881] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ba4c824b-ebe4-4eeb-b8ce-cabfc4cf56f6 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Acquiring lock "f1d29693-8010-44ce-989c-0063421f6c91" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1053.685148] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ba4c824b-ebe4-4eeb-b8ce-cabfc4cf56f6 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Lock "f1d29693-8010-44ce-989c-0063421f6c91" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1053.856410] env[62066]: DEBUG oslo_vmware.api [None req-1d3c474f-5ef6-4ee0-ab35-0ec48db7e18b tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': task-1341351, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.870973] env[62066]: DEBUG nova.compute.manager [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] [instance: 20987cb6-e12c-48c5-8fae-d990b1bb6b8a] Instance has had 0 of 5 cleanup attempts {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1054.187586] env[62066]: DEBUG nova.compute.manager [None req-ba4c824b-ebe4-4eeb-b8ce-cabfc4cf56f6 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: f1d29693-8010-44ce-989c-0063421f6c91] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1054.356734] env[62066]: DEBUG oslo_vmware.api [None req-1d3c474f-5ef6-4ee0-ab35-0ec48db7e18b tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': task-1341351, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.374317] env[62066]: DEBUG nova.compute.manager [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] [instance: a10a6dc0-7df7-40a1-888a-d1414dbeb1f7] Instance has had 0 of 5 cleanup attempts {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1054.709250] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ba4c824b-ebe4-4eeb-b8ce-cabfc4cf56f6 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1054.709457] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ba4c824b-ebe4-4eeb-b8ce-cabfc4cf56f6 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1054.710890] env[62066]: INFO nova.compute.claims [None req-ba4c824b-ebe4-4eeb-b8ce-cabfc4cf56f6 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: f1d29693-8010-44ce-989c-0063421f6c91] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1054.856755] env[62066]: DEBUG oslo_vmware.api [None req-1d3c474f-5ef6-4ee0-ab35-0ec48db7e18b tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': task-1341351, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.877674] env[62066]: DEBUG nova.compute.manager [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] [instance: 15afaea0-65de-4f32-851c-365003e5498f] Instance has had 0 of 5 cleanup attempts {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1055.358212] env[62066]: DEBUG oslo_vmware.api [None req-1d3c474f-5ef6-4ee0-ab35-0ec48db7e18b tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': task-1341351, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.380774] env[62066]: DEBUG nova.compute.manager [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] [instance: 4e39b9b0-4ab0-4f48-8a3f-6bc1895597eb] Instance has had 0 of 5 cleanup attempts {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1055.773394] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fc98706-8fe9-453f-aca4-01e9936cef7e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.780170] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42ace979-310a-48b1-acbf-09f6caf307b0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.808311] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0384202a-f23a-4e73-9777-4ad941484541 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.815138] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fefa5fe-3741-4469-aed5-50f1f8c8345d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.827290] env[62066]: DEBUG nova.compute.provider_tree [None req-ba4c824b-ebe4-4eeb-b8ce-cabfc4cf56f6 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1055.857962] env[62066]: DEBUG oslo_vmware.api [None req-1d3c474f-5ef6-4ee0-ab35-0ec48db7e18b tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': task-1341351, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.883821] env[62066]: DEBUG nova.compute.manager [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] [instance: f4a84152-f222-46a2-9a35-8a0a0078fc12] Instance has had 0 of 5 cleanup attempts {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1056.330402] env[62066]: DEBUG nova.scheduler.client.report [None req-ba4c824b-ebe4-4eeb-b8ce-cabfc4cf56f6 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1056.359472] env[62066]: DEBUG oslo_vmware.api [None req-1d3c474f-5ef6-4ee0-ab35-0ec48db7e18b tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': task-1341351, 'name': ReconfigVM_Task} progress is 18%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.386640] env[62066]: DEBUG nova.compute.manager [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] [instance: 90c33434-c127-450b-9ff0-75181b4ac385] Instance has had 0 of 5 cleanup attempts {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1056.835119] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ba4c824b-ebe4-4eeb-b8ce-cabfc4cf56f6 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.125s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1056.835649] env[62066]: DEBUG nova.compute.manager [None req-ba4c824b-ebe4-4eeb-b8ce-cabfc4cf56f6 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: f1d29693-8010-44ce-989c-0063421f6c91] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1056.859120] env[62066]: DEBUG oslo_vmware.api [None req-1d3c474f-5ef6-4ee0-ab35-0ec48db7e18b tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': task-1341351, 'name': ReconfigVM_Task, 'duration_secs': 5.739798} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1056.859349] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1d3c474f-5ef6-4ee0-ab35-0ec48db7e18b tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Releasing lock "7178eb83-41e7-4516-809d-a0972d7e6e23" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1056.859558] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-1d3c474f-5ef6-4ee0-ab35-0ec48db7e18b tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: 7178eb83-41e7-4516-809d-a0972d7e6e23] Reconfigured VM to detach interface {{(pid=62066) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 1056.889565] env[62066]: DEBUG nova.compute.manager [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] [instance: 9c989152-ec69-478e-a0b0-62983852576c] Instance has had 0 of 5 cleanup attempts {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1057.340148] env[62066]: DEBUG nova.compute.utils [None req-ba4c824b-ebe4-4eeb-b8ce-cabfc4cf56f6 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1057.341636] env[62066]: DEBUG nova.compute.manager [None req-ba4c824b-ebe4-4eeb-b8ce-cabfc4cf56f6 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: f1d29693-8010-44ce-989c-0063421f6c91] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1057.341810] env[62066]: DEBUG nova.network.neutron [None req-ba4c824b-ebe4-4eeb-b8ce-cabfc4cf56f6 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: f1d29693-8010-44ce-989c-0063421f6c91] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1057.386876] env[62066]: DEBUG nova.policy [None req-ba4c824b-ebe4-4eeb-b8ce-cabfc4cf56f6 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '588e53cee85f4ab484b76e7a59fcbe78', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8475ad5a900548cba568360999c846ea', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 1057.392468] env[62066]: DEBUG nova.compute.manager [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] [instance: b718b09b-a7fa-445b-8be3-5b3ebca210a2] Instance has had 0 of 5 cleanup attempts {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1057.649787] env[62066]: DEBUG nova.network.neutron [None req-ba4c824b-ebe4-4eeb-b8ce-cabfc4cf56f6 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: f1d29693-8010-44ce-989c-0063421f6c91] Successfully created port: 47784aec-0880-4e36-8692-67e75ce627db {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1057.846779] env[62066]: DEBUG nova.compute.manager [None req-ba4c824b-ebe4-4eeb-b8ce-cabfc4cf56f6 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: f1d29693-8010-44ce-989c-0063421f6c91] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1057.896771] env[62066]: DEBUG nova.compute.manager [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] [instance: 8807ce4d-532f-469c-a302-464c61c7efeb] Instance has had 0 of 5 cleanup attempts {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1058.168597] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1d3c474f-5ef6-4ee0-ab35-0ec48db7e18b tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Acquiring lock "refresh_cache-7178eb83-41e7-4516-809d-a0972d7e6e23" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1058.168763] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1d3c474f-5ef6-4ee0-ab35-0ec48db7e18b tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Acquired lock "refresh_cache-7178eb83-41e7-4516-809d-a0972d7e6e23" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1058.168967] env[62066]: DEBUG nova.network.neutron [None req-1d3c474f-5ef6-4ee0-ab35-0ec48db7e18b tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: 7178eb83-41e7-4516-809d-a0972d7e6e23] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1058.401843] env[62066]: DEBUG nova.compute.manager [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] [instance: afaca059-fc4e-4de2-8fa0-d2f226ea7051] Instance has had 0 of 5 cleanup attempts {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1058.660528] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9350449a-b162-414c-8112-6af365c96c5e tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Acquiring lock "7178eb83-41e7-4516-809d-a0972d7e6e23" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1058.660824] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9350449a-b162-414c-8112-6af365c96c5e tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Lock "7178eb83-41e7-4516-809d-a0972d7e6e23" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1058.661060] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9350449a-b162-414c-8112-6af365c96c5e tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Acquiring lock "7178eb83-41e7-4516-809d-a0972d7e6e23-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1058.661255] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9350449a-b162-414c-8112-6af365c96c5e tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Lock "7178eb83-41e7-4516-809d-a0972d7e6e23-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1058.661623] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9350449a-b162-414c-8112-6af365c96c5e tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Lock "7178eb83-41e7-4516-809d-a0972d7e6e23-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1058.663719] env[62066]: INFO nova.compute.manager [None req-9350449a-b162-414c-8112-6af365c96c5e tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: 7178eb83-41e7-4516-809d-a0972d7e6e23] Terminating instance [ 1058.665464] env[62066]: DEBUG nova.compute.manager [None req-9350449a-b162-414c-8112-6af365c96c5e tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: 7178eb83-41e7-4516-809d-a0972d7e6e23] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1058.665669] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-9350449a-b162-414c-8112-6af365c96c5e tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: 7178eb83-41e7-4516-809d-a0972d7e6e23] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1058.666523] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b01e9e6-8c02-48c9-8487-eb8b266518c9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.675127] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-9350449a-b162-414c-8112-6af365c96c5e tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: 7178eb83-41e7-4516-809d-a0972d7e6e23] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1058.676020] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a7e0119c-1e10-4cde-98f7-777bc0c5643e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.683543] env[62066]: DEBUG oslo_vmware.api [None req-9350449a-b162-414c-8112-6af365c96c5e tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Waiting for the task: (returnval){ [ 1058.683543] env[62066]: value = "task-1341352" [ 1058.683543] env[62066]: _type = "Task" [ 1058.683543] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1058.691765] env[62066]: DEBUG oslo_vmware.api [None req-9350449a-b162-414c-8112-6af365c96c5e tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': task-1341352, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.857226] env[62066]: DEBUG nova.compute.manager [None req-ba4c824b-ebe4-4eeb-b8ce-cabfc4cf56f6 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: f1d29693-8010-44ce-989c-0063421f6c91] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1058.885867] env[62066]: DEBUG nova.virt.hardware [None req-ba4c824b-ebe4-4eeb-b8ce-cabfc4cf56f6 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-23T13:40:41Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-23T13:40:23Z,direct_url=,disk_format='vmdk',id=50ff584c-3b50-4395-af07-3e66769bc9f7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='eb52f7069a374c61ae946f052007c6d9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-23T13:40:24Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1058.886267] env[62066]: DEBUG nova.virt.hardware [None req-ba4c824b-ebe4-4eeb-b8ce-cabfc4cf56f6 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1058.886453] env[62066]: DEBUG nova.virt.hardware [None req-ba4c824b-ebe4-4eeb-b8ce-cabfc4cf56f6 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1058.886778] env[62066]: DEBUG nova.virt.hardware [None req-ba4c824b-ebe4-4eeb-b8ce-cabfc4cf56f6 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1058.887027] env[62066]: DEBUG nova.virt.hardware [None req-ba4c824b-ebe4-4eeb-b8ce-cabfc4cf56f6 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1058.887206] env[62066]: DEBUG nova.virt.hardware [None req-ba4c824b-ebe4-4eeb-b8ce-cabfc4cf56f6 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1058.887560] env[62066]: DEBUG nova.virt.hardware [None req-ba4c824b-ebe4-4eeb-b8ce-cabfc4cf56f6 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1058.887837] env[62066]: DEBUG nova.virt.hardware [None req-ba4c824b-ebe4-4eeb-b8ce-cabfc4cf56f6 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1058.888087] env[62066]: DEBUG nova.virt.hardware [None req-ba4c824b-ebe4-4eeb-b8ce-cabfc4cf56f6 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1058.888311] env[62066]: DEBUG nova.virt.hardware [None req-ba4c824b-ebe4-4eeb-b8ce-cabfc4cf56f6 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1058.888473] env[62066]: DEBUG nova.virt.hardware [None req-ba4c824b-ebe4-4eeb-b8ce-cabfc4cf56f6 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1058.889500] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52967e54-677e-4624-b83f-575d9daa8568 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.899409] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05332d1a-7354-4267-a34e-759287c18bb0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.905466] env[62066]: DEBUG nova.compute.manager [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] [instance: 170e184a-f009-42cf-82dc-eb9696a00a24] Instance has had 0 of 5 cleanup attempts {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1058.916070] env[62066]: INFO nova.network.neutron [None req-1d3c474f-5ef6-4ee0-ab35-0ec48db7e18b tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: 7178eb83-41e7-4516-809d-a0972d7e6e23] Port dda94cd6-0a2d-4c6b-bc0e-4fb38129159f from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 1058.916443] env[62066]: DEBUG nova.network.neutron [None req-1d3c474f-5ef6-4ee0-ab35-0ec48db7e18b tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: 7178eb83-41e7-4516-809d-a0972d7e6e23] Updating instance_info_cache with network_info: [{"id": "63f16c96-7de0-40e7-9c0c-782122865437", "address": "fa:16:3e:c6:a7:3e", "network": {"id": "449dfe4e-bb49-43b6-9ba7-b57af74ebfb3", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-282029676-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.231", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0d4e33a99c7741fb8cdd97f4ec5dbbd8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db68bd64-5b56-49af-a075-13dcf85cb2e0", "external-id": "nsx-vlan-transportzone-590", "segmentation_id": 590, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap63f16c96-7d", "ovs_interfaceid": "63f16c96-7de0-40e7-9c0c-782122865437", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1059.025162] env[62066]: DEBUG nova.compute.manager [req-70402fb9-b125-458c-809c-4bae8af4b53b req-071d8cee-ab82-4064-83ca-fa004131d835 service nova] [instance: f1d29693-8010-44ce-989c-0063421f6c91] Received event network-vif-plugged-47784aec-0880-4e36-8692-67e75ce627db {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1059.025400] env[62066]: DEBUG oslo_concurrency.lockutils [req-70402fb9-b125-458c-809c-4bae8af4b53b req-071d8cee-ab82-4064-83ca-fa004131d835 service nova] Acquiring lock "f1d29693-8010-44ce-989c-0063421f6c91-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1059.025649] env[62066]: DEBUG oslo_concurrency.lockutils [req-70402fb9-b125-458c-809c-4bae8af4b53b req-071d8cee-ab82-4064-83ca-fa004131d835 service nova] Lock "f1d29693-8010-44ce-989c-0063421f6c91-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1059.025782] env[62066]: DEBUG oslo_concurrency.lockutils [req-70402fb9-b125-458c-809c-4bae8af4b53b req-071d8cee-ab82-4064-83ca-fa004131d835 service nova] Lock "f1d29693-8010-44ce-989c-0063421f6c91-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1059.025955] env[62066]: DEBUG nova.compute.manager [req-70402fb9-b125-458c-809c-4bae8af4b53b req-071d8cee-ab82-4064-83ca-fa004131d835 service nova] [instance: f1d29693-8010-44ce-989c-0063421f6c91] No waiting events found dispatching network-vif-plugged-47784aec-0880-4e36-8692-67e75ce627db {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1059.026346] env[62066]: WARNING nova.compute.manager [req-70402fb9-b125-458c-809c-4bae8af4b53b req-071d8cee-ab82-4064-83ca-fa004131d835 service nova] [instance: f1d29693-8010-44ce-989c-0063421f6c91] Received unexpected event network-vif-plugged-47784aec-0880-4e36-8692-67e75ce627db for instance with vm_state building and task_state spawning. [ 1059.194064] env[62066]: DEBUG oslo_vmware.api [None req-9350449a-b162-414c-8112-6af365c96c5e tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': task-1341352, 'name': PowerOffVM_Task, 'duration_secs': 0.275909} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1059.194349] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-9350449a-b162-414c-8112-6af365c96c5e tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: 7178eb83-41e7-4516-809d-a0972d7e6e23] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1059.194526] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-9350449a-b162-414c-8112-6af365c96c5e tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: 7178eb83-41e7-4516-809d-a0972d7e6e23] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1059.194777] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-119df7cc-3bae-4593-a6b6-10e488029858 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.257333] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-9350449a-b162-414c-8112-6af365c96c5e tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: 7178eb83-41e7-4516-809d-a0972d7e6e23] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1059.257574] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-9350449a-b162-414c-8112-6af365c96c5e tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: 7178eb83-41e7-4516-809d-a0972d7e6e23] Deleting contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1059.257758] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-9350449a-b162-414c-8112-6af365c96c5e tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Deleting the datastore file [datastore2] 7178eb83-41e7-4516-809d-a0972d7e6e23 {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1059.258035] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c45d9099-8d50-4155-b266-134680e0bf8e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.263971] env[62066]: DEBUG oslo_vmware.api [None req-9350449a-b162-414c-8112-6af365c96c5e tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Waiting for the task: (returnval){ [ 1059.263971] env[62066]: value = "task-1341354" [ 1059.263971] env[62066]: _type = "Task" [ 1059.263971] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.271965] env[62066]: DEBUG oslo_vmware.api [None req-9350449a-b162-414c-8112-6af365c96c5e tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': task-1341354, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.419279] env[62066]: DEBUG nova.compute.manager [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] [instance: 2634fcc0-96bd-4513-8f92-515f83fe23d7] Instance has had 0 of 5 cleanup attempts {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1059.421602] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1d3c474f-5ef6-4ee0-ab35-0ec48db7e18b tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Releasing lock "refresh_cache-7178eb83-41e7-4516-809d-a0972d7e6e23" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1059.581823] env[62066]: DEBUG nova.network.neutron [None req-ba4c824b-ebe4-4eeb-b8ce-cabfc4cf56f6 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: f1d29693-8010-44ce-989c-0063421f6c91] Successfully updated port: 47784aec-0880-4e36-8692-67e75ce627db {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1059.603243] env[62066]: DEBUG nova.compute.manager [req-d970efc9-30dc-4a9e-8703-ac7ccab066d8 req-6a35170b-b5c3-47fc-b664-2fa736979e32 service nova] [instance: f1d29693-8010-44ce-989c-0063421f6c91] Received event network-changed-47784aec-0880-4e36-8692-67e75ce627db {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1059.603523] env[62066]: DEBUG nova.compute.manager [req-d970efc9-30dc-4a9e-8703-ac7ccab066d8 req-6a35170b-b5c3-47fc-b664-2fa736979e32 service nova] [instance: f1d29693-8010-44ce-989c-0063421f6c91] Refreshing instance network info cache due to event network-changed-47784aec-0880-4e36-8692-67e75ce627db. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1059.603695] env[62066]: DEBUG oslo_concurrency.lockutils [req-d970efc9-30dc-4a9e-8703-ac7ccab066d8 req-6a35170b-b5c3-47fc-b664-2fa736979e32 service nova] Acquiring lock "refresh_cache-f1d29693-8010-44ce-989c-0063421f6c91" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1059.603806] env[62066]: DEBUG oslo_concurrency.lockutils [req-d970efc9-30dc-4a9e-8703-ac7ccab066d8 req-6a35170b-b5c3-47fc-b664-2fa736979e32 service nova] Acquired lock "refresh_cache-f1d29693-8010-44ce-989c-0063421f6c91" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1059.603970] env[62066]: DEBUG nova.network.neutron [req-d970efc9-30dc-4a9e-8703-ac7ccab066d8 req-6a35170b-b5c3-47fc-b664-2fa736979e32 service nova] [instance: f1d29693-8010-44ce-989c-0063421f6c91] Refreshing network info cache for port 47784aec-0880-4e36-8692-67e75ce627db {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1059.775306] env[62066]: DEBUG oslo_vmware.api [None req-9350449a-b162-414c-8112-6af365c96c5e tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': task-1341354, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.180374} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1059.775481] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-9350449a-b162-414c-8112-6af365c96c5e tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1059.775668] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-9350449a-b162-414c-8112-6af365c96c5e tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: 7178eb83-41e7-4516-809d-a0972d7e6e23] Deleted contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1059.775846] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-9350449a-b162-414c-8112-6af365c96c5e tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: 7178eb83-41e7-4516-809d-a0972d7e6e23] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1059.776036] env[62066]: INFO nova.compute.manager [None req-9350449a-b162-414c-8112-6af365c96c5e tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: 7178eb83-41e7-4516-809d-a0972d7e6e23] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1059.776282] env[62066]: DEBUG oslo.service.loopingcall [None req-9350449a-b162-414c-8112-6af365c96c5e tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1059.776477] env[62066]: DEBUG nova.compute.manager [-] [instance: 7178eb83-41e7-4516-809d-a0972d7e6e23] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1059.776570] env[62066]: DEBUG nova.network.neutron [-] [instance: 7178eb83-41e7-4516-809d-a0972d7e6e23] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1059.925208] env[62066]: DEBUG nova.compute.manager [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] [instance: 76306c3f-0674-40fe-9864-d82d11e4bb5e] Instance has had 0 of 5 cleanup attempts {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1059.928050] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1d3c474f-5ef6-4ee0-ab35-0ec48db7e18b tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Lock "interface-7178eb83-41e7-4516-809d-a0972d7e6e23-dda94cd6-0a2d-4c6b-bc0e-4fb38129159f" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 9.655s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1060.087024] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ba4c824b-ebe4-4eeb-b8ce-cabfc4cf56f6 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Acquiring lock "refresh_cache-f1d29693-8010-44ce-989c-0063421f6c91" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1060.150271] env[62066]: DEBUG nova.network.neutron [req-d970efc9-30dc-4a9e-8703-ac7ccab066d8 req-6a35170b-b5c3-47fc-b664-2fa736979e32 service nova] [instance: f1d29693-8010-44ce-989c-0063421f6c91] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1060.279546] env[62066]: DEBUG nova.network.neutron [req-d970efc9-30dc-4a9e-8703-ac7ccab066d8 req-6a35170b-b5c3-47fc-b664-2fa736979e32 service nova] [instance: f1d29693-8010-44ce-989c-0063421f6c91] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1060.430007] env[62066]: DEBUG nova.compute.manager [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] [instance: c8e0d47c-4421-4e00-9183-206fceeabc40] Instance has had 0 of 5 cleanup attempts {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1060.781951] env[62066]: DEBUG oslo_concurrency.lockutils [req-d970efc9-30dc-4a9e-8703-ac7ccab066d8 req-6a35170b-b5c3-47fc-b664-2fa736979e32 service nova] Releasing lock "refresh_cache-f1d29693-8010-44ce-989c-0063421f6c91" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1060.782365] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ba4c824b-ebe4-4eeb-b8ce-cabfc4cf56f6 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Acquired lock "refresh_cache-f1d29693-8010-44ce-989c-0063421f6c91" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1060.782670] env[62066]: DEBUG nova.network.neutron [None req-ba4c824b-ebe4-4eeb-b8ce-cabfc4cf56f6 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: f1d29693-8010-44ce-989c-0063421f6c91] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1060.934193] env[62066]: DEBUG nova.compute.manager [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] [instance: 0bc31a8b-a5fc-4f6e-a4c0-024bf4260755] Instance has had 0 of 5 cleanup attempts {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1060.963604] env[62066]: DEBUG nova.network.neutron [-] [instance: 7178eb83-41e7-4516-809d-a0972d7e6e23] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1061.321859] env[62066]: DEBUG nova.network.neutron [None req-ba4c824b-ebe4-4eeb-b8ce-cabfc4cf56f6 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: f1d29693-8010-44ce-989c-0063421f6c91] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1061.437347] env[62066]: DEBUG nova.compute.manager [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] [instance: 3120c9a5-0558-4e83-b0fb-994a989398fb] Instance has had 0 of 5 cleanup attempts {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1061.451576] env[62066]: DEBUG nova.network.neutron [None req-ba4c824b-ebe4-4eeb-b8ce-cabfc4cf56f6 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: f1d29693-8010-44ce-989c-0063421f6c91] Updating instance_info_cache with network_info: [{"id": "47784aec-0880-4e36-8692-67e75ce627db", "address": "fa:16:3e:88:b4:76", "network": {"id": "7617f887-bdd0-400a-8b5f-606d7785df78", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-558766873-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8475ad5a900548cba568360999c846ea", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2b6a4065-12af-4fb9-ac47-ec9143f7297e", "external-id": "nsx-vlan-transportzone-95", "segmentation_id": 95, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap47784aec-08", "ovs_interfaceid": "47784aec-0880-4e36-8692-67e75ce627db", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1061.466252] env[62066]: INFO nova.compute.manager [-] [instance: 7178eb83-41e7-4516-809d-a0972d7e6e23] Took 1.69 seconds to deallocate network for instance. [ 1061.627542] env[62066]: DEBUG nova.compute.manager [req-f0ead5ed-1109-4101-962e-17b8ad9e953e req-784c0795-68f4-4ae2-8a85-21cd33d4279e service nova] [instance: 7178eb83-41e7-4516-809d-a0972d7e6e23] Received event network-vif-deleted-63f16c96-7de0-40e7-9c0c-782122865437 {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1061.940141] env[62066]: DEBUG nova.compute.manager [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] [instance: d77e3405-d158-4a8c-9f54-d5fbd26fed48] Instance has had 0 of 5 cleanup attempts {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1061.954041] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ba4c824b-ebe4-4eeb-b8ce-cabfc4cf56f6 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Releasing lock "refresh_cache-f1d29693-8010-44ce-989c-0063421f6c91" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1061.954150] env[62066]: DEBUG nova.compute.manager [None req-ba4c824b-ebe4-4eeb-b8ce-cabfc4cf56f6 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: f1d29693-8010-44ce-989c-0063421f6c91] Instance network_info: |[{"id": "47784aec-0880-4e36-8692-67e75ce627db", "address": "fa:16:3e:88:b4:76", "network": {"id": "7617f887-bdd0-400a-8b5f-606d7785df78", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-558766873-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8475ad5a900548cba568360999c846ea", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2b6a4065-12af-4fb9-ac47-ec9143f7297e", "external-id": "nsx-vlan-transportzone-95", "segmentation_id": 95, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap47784aec-08", "ovs_interfaceid": "47784aec-0880-4e36-8692-67e75ce627db", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1061.954540] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-ba4c824b-ebe4-4eeb-b8ce-cabfc4cf56f6 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: f1d29693-8010-44ce-989c-0063421f6c91] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:88:b4:76', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '2b6a4065-12af-4fb9-ac47-ec9143f7297e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '47784aec-0880-4e36-8692-67e75ce627db', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1061.962026] env[62066]: DEBUG oslo.service.loopingcall [None req-ba4c824b-ebe4-4eeb-b8ce-cabfc4cf56f6 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1061.962806] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f1d29693-8010-44ce-989c-0063421f6c91] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1061.963045] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-73b7f4bf-fd01-4ed5-b4ba-b186d5734a70 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.977870] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9350449a-b162-414c-8112-6af365c96c5e tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1061.978110] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9350449a-b162-414c-8112-6af365c96c5e tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1061.978328] env[62066]: DEBUG nova.objects.instance [None req-9350449a-b162-414c-8112-6af365c96c5e tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Lazy-loading 'resources' on Instance uuid 7178eb83-41e7-4516-809d-a0972d7e6e23 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1061.984299] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1061.984299] env[62066]: value = "task-1341355" [ 1061.984299] env[62066]: _type = "Task" [ 1061.984299] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1061.992129] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1341355, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.443344] env[62066]: DEBUG nova.compute.manager [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] [instance: b08f27b1-429e-4d05-9a95-f45fd0cb2904] Instance has had 0 of 5 cleanup attempts {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1062.494828] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1341355, 'name': CreateVM_Task, 'duration_secs': 0.298994} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1062.495029] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f1d29693-8010-44ce-989c-0063421f6c91] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1062.495668] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ba4c824b-ebe4-4eeb-b8ce-cabfc4cf56f6 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1062.495839] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ba4c824b-ebe4-4eeb-b8ce-cabfc4cf56f6 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Acquired lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1062.496175] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ba4c824b-ebe4-4eeb-b8ce-cabfc4cf56f6 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1062.496424] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9e35e571-1ac7-41d0-b44f-aeb89a554740 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.500809] env[62066]: DEBUG oslo_vmware.api [None req-ba4c824b-ebe4-4eeb-b8ce-cabfc4cf56f6 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Waiting for the task: (returnval){ [ 1062.500809] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]52190fd6-9307-6b1e-3798-dfe876995fcb" [ 1062.500809] env[62066]: _type = "Task" [ 1062.500809] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1062.508126] env[62066]: DEBUG oslo_vmware.api [None req-ba4c824b-ebe4-4eeb-b8ce-cabfc4cf56f6 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52190fd6-9307-6b1e-3798-dfe876995fcb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.666151] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bda99202-6e45-4799-bd71-f0d7c54521f1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.673706] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27af94da-f629-4a4b-a0fb-81685ef45dda {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.702584] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebbaef31-8747-45d4-bd15-2be54a80cacd {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.709584] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d52998aa-d6eb-46ae-baff-8633d67d58c4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.721979] env[62066]: DEBUG nova.compute.provider_tree [None req-9350449a-b162-414c-8112-6af365c96c5e tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Updating inventory in ProviderTree for provider 8e4f7194-1498-4f08-8723-ab7260524bcb with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1062.946377] env[62066]: DEBUG nova.compute.manager [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] [instance: 86a92b61-f3e9-48f9-8ee2-756669d558ef] Instance has had 0 of 5 cleanup attempts {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1063.010955] env[62066]: DEBUG oslo_vmware.api [None req-ba4c824b-ebe4-4eeb-b8ce-cabfc4cf56f6 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52190fd6-9307-6b1e-3798-dfe876995fcb, 'name': SearchDatastore_Task, 'duration_secs': 0.008864} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1063.011240] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ba4c824b-ebe4-4eeb-b8ce-cabfc4cf56f6 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Releasing lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1063.011501] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-ba4c824b-ebe4-4eeb-b8ce-cabfc4cf56f6 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: f1d29693-8010-44ce-989c-0063421f6c91] Processing image 50ff584c-3b50-4395-af07-3e66769bc9f7 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1063.011741] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ba4c824b-ebe4-4eeb-b8ce-cabfc4cf56f6 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1063.011894] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ba4c824b-ebe4-4eeb-b8ce-cabfc4cf56f6 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Acquired lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1063.012092] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-ba4c824b-ebe4-4eeb-b8ce-cabfc4cf56f6 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1063.012351] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-81804e25-4072-4c3d-8d6a-0c83f65c9f59 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.019742] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-ba4c824b-ebe4-4eeb-b8ce-cabfc4cf56f6 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1063.019919] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-ba4c824b-ebe4-4eeb-b8ce-cabfc4cf56f6 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1063.020592] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c7a405a0-2b8c-4435-b467-cfe367a24e22 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.025287] env[62066]: DEBUG oslo_vmware.api [None req-ba4c824b-ebe4-4eeb-b8ce-cabfc4cf56f6 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Waiting for the task: (returnval){ [ 1063.025287] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]52361b08-b7fd-1451-c17f-da73924ed995" [ 1063.025287] env[62066]: _type = "Task" [ 1063.025287] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1063.033357] env[62066]: DEBUG oslo_vmware.api [None req-ba4c824b-ebe4-4eeb-b8ce-cabfc4cf56f6 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52361b08-b7fd-1451-c17f-da73924ed995, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.241891] env[62066]: ERROR nova.scheduler.client.report [None req-9350449a-b162-414c-8112-6af365c96c5e tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [req-8753fc67-1172-4575-b264-512686f8fe8b] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 8e4f7194-1498-4f08-8723-ab7260524bcb. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-8753fc67-1172-4575-b264-512686f8fe8b"}]} [ 1063.256795] env[62066]: DEBUG nova.scheduler.client.report [None req-9350449a-b162-414c-8112-6af365c96c5e tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Refreshing inventories for resource provider 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 1063.269178] env[62066]: DEBUG nova.scheduler.client.report [None req-9350449a-b162-414c-8112-6af365c96c5e tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Updating ProviderTree inventory for provider 8e4f7194-1498-4f08-8723-ab7260524bcb from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 1063.269382] env[62066]: DEBUG nova.compute.provider_tree [None req-9350449a-b162-414c-8112-6af365c96c5e tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Updating inventory in ProviderTree for provider 8e4f7194-1498-4f08-8723-ab7260524bcb with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 153, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1063.278363] env[62066]: DEBUG nova.scheduler.client.report [None req-9350449a-b162-414c-8112-6af365c96c5e tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Refreshing aggregate associations for resource provider 8e4f7194-1498-4f08-8723-ab7260524bcb, aggregates: None {{(pid=62066) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 1063.294131] env[62066]: DEBUG nova.scheduler.client.report [None req-9350449a-b162-414c-8112-6af365c96c5e tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Refreshing trait associations for resource provider 8e4f7194-1498-4f08-8723-ab7260524bcb, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_NODE,HW_ARCH_X86_64,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=62066) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 1063.340334] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d28ede47-a333-4c34-80ac-8586fe818e47 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.347690] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad6b8d0c-8b6b-4242-baff-be95648e0c33 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.376765] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-893fd87e-7cbb-47e5-8e85-1bca305eb5df {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.383572] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c286aaaf-63f4-428d-9211-85a3a4c1783e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.396159] env[62066]: DEBUG nova.compute.provider_tree [None req-9350449a-b162-414c-8112-6af365c96c5e tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Updating inventory in ProviderTree for provider 8e4f7194-1498-4f08-8723-ab7260524bcb with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1063.449628] env[62066]: DEBUG nova.compute.manager [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] [instance: 5e96def2-0cbd-4bd9-93f4-6a365a0142b0] Instance has had 0 of 5 cleanup attempts {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1063.534612] env[62066]: DEBUG oslo_vmware.api [None req-ba4c824b-ebe4-4eeb-b8ce-cabfc4cf56f6 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52361b08-b7fd-1451-c17f-da73924ed995, 'name': SearchDatastore_Task, 'duration_secs': 0.007632} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1063.535338] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-aaa56a2c-d927-4eb9-899f-4f26d115b11d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.540120] env[62066]: DEBUG oslo_vmware.api [None req-ba4c824b-ebe4-4eeb-b8ce-cabfc4cf56f6 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Waiting for the task: (returnval){ [ 1063.540120] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]5230b53c-4cb6-8af5-a9ab-e8494b5aac0e" [ 1063.540120] env[62066]: _type = "Task" [ 1063.540120] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1063.547485] env[62066]: DEBUG oslo_vmware.api [None req-ba4c824b-ebe4-4eeb-b8ce-cabfc4cf56f6 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]5230b53c-4cb6-8af5-a9ab-e8494b5aac0e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.928042] env[62066]: DEBUG nova.scheduler.client.report [None req-9350449a-b162-414c-8112-6af365c96c5e tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Updated inventory for provider 8e4f7194-1498-4f08-8723-ab7260524bcb with generation 124 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 1063.928042] env[62066]: DEBUG nova.compute.provider_tree [None req-9350449a-b162-414c-8112-6af365c96c5e tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Updating resource provider 8e4f7194-1498-4f08-8723-ab7260524bcb generation from 124 to 125 during operation: update_inventory {{(pid=62066) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1063.928042] env[62066]: DEBUG nova.compute.provider_tree [None req-9350449a-b162-414c-8112-6af365c96c5e tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Updating inventory in ProviderTree for provider 8e4f7194-1498-4f08-8723-ab7260524bcb with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1063.952878] env[62066]: DEBUG nova.compute.manager [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] [instance: de2e2be7-efdb-45a8-842a-640ab9deb1d9] Instance has had 0 of 5 cleanup attempts {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1064.050897] env[62066]: DEBUG oslo_vmware.api [None req-ba4c824b-ebe4-4eeb-b8ce-cabfc4cf56f6 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]5230b53c-4cb6-8af5-a9ab-e8494b5aac0e, 'name': SearchDatastore_Task, 'duration_secs': 0.00817} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1064.051201] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ba4c824b-ebe4-4eeb-b8ce-cabfc4cf56f6 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Releasing lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1064.051530] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-ba4c824b-ebe4-4eeb-b8ce-cabfc4cf56f6 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk to [datastore2] f1d29693-8010-44ce-989c-0063421f6c91/f1d29693-8010-44ce-989c-0063421f6c91.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1064.051794] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a6e7bbdd-7951-4f56-8315-0fbc09188fcb {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.058129] env[62066]: DEBUG oslo_vmware.api [None req-ba4c824b-ebe4-4eeb-b8ce-cabfc4cf56f6 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Waiting for the task: (returnval){ [ 1064.058129] env[62066]: value = "task-1341356" [ 1064.058129] env[62066]: _type = "Task" [ 1064.058129] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1064.065590] env[62066]: DEBUG oslo_vmware.api [None req-ba4c824b-ebe4-4eeb-b8ce-cabfc4cf56f6 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341356, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.432425] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9350449a-b162-414c-8112-6af365c96c5e tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.454s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1064.456735] env[62066]: DEBUG nova.compute.manager [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] [instance: 88daaaa6-f385-4161-bc74-9fffabb1145d] Instance has had 0 of 5 cleanup attempts {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1064.463026] env[62066]: INFO nova.scheduler.client.report [None req-9350449a-b162-414c-8112-6af365c96c5e tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Deleted allocations for instance 7178eb83-41e7-4516-809d-a0972d7e6e23 [ 1064.568331] env[62066]: DEBUG oslo_vmware.api [None req-ba4c824b-ebe4-4eeb-b8ce-cabfc4cf56f6 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341356, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.450274} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1064.568608] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-ba4c824b-ebe4-4eeb-b8ce-cabfc4cf56f6 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk to [datastore2] f1d29693-8010-44ce-989c-0063421f6c91/f1d29693-8010-44ce-989c-0063421f6c91.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1064.568828] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-ba4c824b-ebe4-4eeb-b8ce-cabfc4cf56f6 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: f1d29693-8010-44ce-989c-0063421f6c91] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1064.569086] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a1e27b60-f104-429a-a6b4-f7d4dcb9633a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.575823] env[62066]: DEBUG oslo_vmware.api [None req-ba4c824b-ebe4-4eeb-b8ce-cabfc4cf56f6 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Waiting for the task: (returnval){ [ 1064.575823] env[62066]: value = "task-1341357" [ 1064.575823] env[62066]: _type = "Task" [ 1064.575823] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1064.584404] env[62066]: DEBUG oslo_vmware.api [None req-ba4c824b-ebe4-4eeb-b8ce-cabfc4cf56f6 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341357, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.962304] env[62066]: DEBUG nova.compute.manager [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] [instance: 2709d0ad-5dd8-4e3c-b1e6-3d1e2e242380] Instance has had 0 of 5 cleanup attempts {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1064.974778] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9350449a-b162-414c-8112-6af365c96c5e tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Lock "7178eb83-41e7-4516-809d-a0972d7e6e23" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.314s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1065.085640] env[62066]: DEBUG oslo_vmware.api [None req-ba4c824b-ebe4-4eeb-b8ce-cabfc4cf56f6 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341357, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.08949} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1065.086057] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-ba4c824b-ebe4-4eeb-b8ce-cabfc4cf56f6 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: f1d29693-8010-44ce-989c-0063421f6c91] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1065.086792] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0912b5c-a8db-45d7-a2b9-f8d7a38450db {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.108217] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-ba4c824b-ebe4-4eeb-b8ce-cabfc4cf56f6 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: f1d29693-8010-44ce-989c-0063421f6c91] Reconfiguring VM instance instance-00000069 to attach disk [datastore2] f1d29693-8010-44ce-989c-0063421f6c91/f1d29693-8010-44ce-989c-0063421f6c91.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1065.108483] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1ce4a387-81f7-4c57-afa2-d743d5b0f24a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.127436] env[62066]: DEBUG oslo_vmware.api [None req-ba4c824b-ebe4-4eeb-b8ce-cabfc4cf56f6 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Waiting for the task: (returnval){ [ 1065.127436] env[62066]: value = "task-1341358" [ 1065.127436] env[62066]: _type = "Task" [ 1065.127436] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1065.134886] env[62066]: DEBUG oslo_vmware.api [None req-ba4c824b-ebe4-4eeb-b8ce-cabfc4cf56f6 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341358, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.220613] env[62066]: DEBUG oslo_concurrency.lockutils [None req-89d5c2d6-6146-4219-94ae-932f1f89cbd7 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Acquiring lock "c8482d92-0180-44ee-a4f1-bf84786dad43" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1065.220939] env[62066]: DEBUG oslo_concurrency.lockutils [None req-89d5c2d6-6146-4219-94ae-932f1f89cbd7 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Lock "c8482d92-0180-44ee-a4f1-bf84786dad43" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1065.221183] env[62066]: DEBUG oslo_concurrency.lockutils [None req-89d5c2d6-6146-4219-94ae-932f1f89cbd7 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Acquiring lock "c8482d92-0180-44ee-a4f1-bf84786dad43-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1065.221374] env[62066]: DEBUG oslo_concurrency.lockutils [None req-89d5c2d6-6146-4219-94ae-932f1f89cbd7 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Lock "c8482d92-0180-44ee-a4f1-bf84786dad43-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1065.221598] env[62066]: DEBUG oslo_concurrency.lockutils [None req-89d5c2d6-6146-4219-94ae-932f1f89cbd7 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Lock "c8482d92-0180-44ee-a4f1-bf84786dad43-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1065.223935] env[62066]: INFO nova.compute.manager [None req-89d5c2d6-6146-4219-94ae-932f1f89cbd7 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: c8482d92-0180-44ee-a4f1-bf84786dad43] Terminating instance [ 1065.225705] env[62066]: DEBUG nova.compute.manager [None req-89d5c2d6-6146-4219-94ae-932f1f89cbd7 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: c8482d92-0180-44ee-a4f1-bf84786dad43] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1065.225925] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-89d5c2d6-6146-4219-94ae-932f1f89cbd7 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: c8482d92-0180-44ee-a4f1-bf84786dad43] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1065.226788] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef7e9793-0094-46c1-9c31-f07a6089a86d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.234100] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-89d5c2d6-6146-4219-94ae-932f1f89cbd7 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: c8482d92-0180-44ee-a4f1-bf84786dad43] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1065.234347] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-63c7dc03-9648-4d39-9738-88e3e213dde4 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.241954] env[62066]: DEBUG oslo_vmware.api [None req-89d5c2d6-6146-4219-94ae-932f1f89cbd7 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Waiting for the task: (returnval){ [ 1065.241954] env[62066]: value = "task-1341359" [ 1065.241954] env[62066]: _type = "Task" [ 1065.241954] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1065.249344] env[62066]: DEBUG oslo_vmware.api [None req-89d5c2d6-6146-4219-94ae-932f1f89cbd7 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': task-1341359, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.258675] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c2e9c493-1a5c-4cf6-adaf-d7b784e8bbad tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Acquiring lock "3ef1410d-c78b-47d9-807b-e80368c05902" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1065.258902] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c2e9c493-1a5c-4cf6-adaf-d7b784e8bbad tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Lock "3ef1410d-c78b-47d9-807b-e80368c05902" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1065.466657] env[62066]: DEBUG nova.compute.manager [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] [instance: 8ff6fbb9-c90f-498d-9a85-d220a8c2f794] Instance has had 0 of 5 cleanup attempts {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1065.636866] env[62066]: DEBUG oslo_vmware.api [None req-ba4c824b-ebe4-4eeb-b8ce-cabfc4cf56f6 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341358, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.750918] env[62066]: DEBUG oslo_vmware.api [None req-89d5c2d6-6146-4219-94ae-932f1f89cbd7 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': task-1341359, 'name': PowerOffVM_Task, 'duration_secs': 0.230846} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1065.751227] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-89d5c2d6-6146-4219-94ae-932f1f89cbd7 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: c8482d92-0180-44ee-a4f1-bf84786dad43] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1065.751415] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-89d5c2d6-6146-4219-94ae-932f1f89cbd7 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: c8482d92-0180-44ee-a4f1-bf84786dad43] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1065.751680] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-92b11b3a-2256-4633-bf8b-65281b47d10e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.762170] env[62066]: DEBUG nova.compute.utils [None req-c2e9c493-1a5c-4cf6-adaf-d7b784e8bbad tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1065.809432] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-89d5c2d6-6146-4219-94ae-932f1f89cbd7 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: c8482d92-0180-44ee-a4f1-bf84786dad43] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1065.809667] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-89d5c2d6-6146-4219-94ae-932f1f89cbd7 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: c8482d92-0180-44ee-a4f1-bf84786dad43] Deleting contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1065.809854] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-89d5c2d6-6146-4219-94ae-932f1f89cbd7 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Deleting the datastore file [datastore2] c8482d92-0180-44ee-a4f1-bf84786dad43 {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1065.810136] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-81abeaf5-7e55-4983-a52b-687236efc2e7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.816203] env[62066]: DEBUG oslo_vmware.api [None req-89d5c2d6-6146-4219-94ae-932f1f89cbd7 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Waiting for the task: (returnval){ [ 1065.816203] env[62066]: value = "task-1341361" [ 1065.816203] env[62066]: _type = "Task" [ 1065.816203] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1065.823216] env[62066]: DEBUG oslo_vmware.api [None req-89d5c2d6-6146-4219-94ae-932f1f89cbd7 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': task-1341361, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.969957] env[62066]: DEBUG nova.compute.manager [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] [instance: 49bc1350-0095-406b-bc68-005eb6b681a8] Instance has had 0 of 5 cleanup attempts {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1066.137546] env[62066]: DEBUG oslo_vmware.api [None req-ba4c824b-ebe4-4eeb-b8ce-cabfc4cf56f6 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341358, 'name': ReconfigVM_Task, 'duration_secs': 0.51125} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1066.137837] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-ba4c824b-ebe4-4eeb-b8ce-cabfc4cf56f6 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: f1d29693-8010-44ce-989c-0063421f6c91] Reconfigured VM instance instance-00000069 to attach disk [datastore2] f1d29693-8010-44ce-989c-0063421f6c91/f1d29693-8010-44ce-989c-0063421f6c91.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1066.138458] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b6c66081-7cfb-4cea-9304-0c9ef4e2da9e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.144700] env[62066]: DEBUG oslo_vmware.api [None req-ba4c824b-ebe4-4eeb-b8ce-cabfc4cf56f6 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Waiting for the task: (returnval){ [ 1066.144700] env[62066]: value = "task-1341362" [ 1066.144700] env[62066]: _type = "Task" [ 1066.144700] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1066.152152] env[62066]: DEBUG oslo_vmware.api [None req-ba4c824b-ebe4-4eeb-b8ce-cabfc4cf56f6 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341362, 'name': Rename_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.264740] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c2e9c493-1a5c-4cf6-adaf-d7b784e8bbad tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Lock "3ef1410d-c78b-47d9-807b-e80368c05902" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1066.325374] env[62066]: DEBUG oslo_vmware.api [None req-89d5c2d6-6146-4219-94ae-932f1f89cbd7 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Task: {'id': task-1341361, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.1076} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1066.325633] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-89d5c2d6-6146-4219-94ae-932f1f89cbd7 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1066.325822] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-89d5c2d6-6146-4219-94ae-932f1f89cbd7 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: c8482d92-0180-44ee-a4f1-bf84786dad43] Deleted contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1066.326017] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-89d5c2d6-6146-4219-94ae-932f1f89cbd7 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: c8482d92-0180-44ee-a4f1-bf84786dad43] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1066.326214] env[62066]: INFO nova.compute.manager [None req-89d5c2d6-6146-4219-94ae-932f1f89cbd7 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] [instance: c8482d92-0180-44ee-a4f1-bf84786dad43] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1066.326461] env[62066]: DEBUG oslo.service.loopingcall [None req-89d5c2d6-6146-4219-94ae-932f1f89cbd7 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1066.326660] env[62066]: DEBUG nova.compute.manager [-] [instance: c8482d92-0180-44ee-a4f1-bf84786dad43] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1066.326757] env[62066]: DEBUG nova.network.neutron [-] [instance: c8482d92-0180-44ee-a4f1-bf84786dad43] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1066.472941] env[62066]: DEBUG nova.compute.manager [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] [instance: 050040cb-6ee3-4ad2-960a-fcebb53ac394] Instance has had 0 of 5 cleanup attempts {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1066.599110] env[62066]: DEBUG nova.compute.manager [req-1c302d34-408a-47ae-8a0a-6cb9ab902f2c req-e805147e-f587-424e-9488-6101a1500ed9 service nova] [instance: c8482d92-0180-44ee-a4f1-bf84786dad43] Received event network-vif-deleted-2f2f1786-934d-478f-be39-7509cfacf86f {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1066.599321] env[62066]: INFO nova.compute.manager [req-1c302d34-408a-47ae-8a0a-6cb9ab902f2c req-e805147e-f587-424e-9488-6101a1500ed9 service nova] [instance: c8482d92-0180-44ee-a4f1-bf84786dad43] Neutron deleted interface 2f2f1786-934d-478f-be39-7509cfacf86f; detaching it from the instance and deleting it from the info cache [ 1066.599512] env[62066]: DEBUG nova.network.neutron [req-1c302d34-408a-47ae-8a0a-6cb9ab902f2c req-e805147e-f587-424e-9488-6101a1500ed9 service nova] [instance: c8482d92-0180-44ee-a4f1-bf84786dad43] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1066.655976] env[62066]: DEBUG oslo_vmware.api [None req-ba4c824b-ebe4-4eeb-b8ce-cabfc4cf56f6 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341362, 'name': Rename_Task, 'duration_secs': 0.139091} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1066.656282] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-ba4c824b-ebe4-4eeb-b8ce-cabfc4cf56f6 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: f1d29693-8010-44ce-989c-0063421f6c91] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1066.656535] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-78c92f20-66fa-4f75-9254-9a611abe1797 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.663162] env[62066]: DEBUG oslo_vmware.api [None req-ba4c824b-ebe4-4eeb-b8ce-cabfc4cf56f6 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Waiting for the task: (returnval){ [ 1066.663162] env[62066]: value = "task-1341363" [ 1066.663162] env[62066]: _type = "Task" [ 1066.663162] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1066.670507] env[62066]: DEBUG oslo_vmware.api [None req-ba4c824b-ebe4-4eeb-b8ce-cabfc4cf56f6 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341363, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.978023] env[62066]: DEBUG nova.compute.manager [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] [instance: 5f645eaa-6edc-4362-82ad-38c4c57b2be2] Instance has had 0 of 5 cleanup attempts {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1067.076459] env[62066]: DEBUG nova.network.neutron [-] [instance: c8482d92-0180-44ee-a4f1-bf84786dad43] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1067.102165] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9d609a18-0745-4bd3-9f66-eae17b73be30 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.110865] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a2f4b91-db08-4458-b52d-08c3e532a74a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.133547] env[62066]: DEBUG nova.compute.manager [req-1c302d34-408a-47ae-8a0a-6cb9ab902f2c req-e805147e-f587-424e-9488-6101a1500ed9 service nova] [instance: c8482d92-0180-44ee-a4f1-bf84786dad43] Detach interface failed, port_id=2f2f1786-934d-478f-be39-7509cfacf86f, reason: Instance c8482d92-0180-44ee-a4f1-bf84786dad43 could not be found. {{(pid=62066) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1067.172772] env[62066]: DEBUG oslo_vmware.api [None req-ba4c824b-ebe4-4eeb-b8ce-cabfc4cf56f6 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341363, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.334488] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c2e9c493-1a5c-4cf6-adaf-d7b784e8bbad tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Acquiring lock "3ef1410d-c78b-47d9-807b-e80368c05902" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1067.334752] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c2e9c493-1a5c-4cf6-adaf-d7b784e8bbad tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Lock "3ef1410d-c78b-47d9-807b-e80368c05902" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1067.335087] env[62066]: INFO nova.compute.manager [None req-c2e9c493-1a5c-4cf6-adaf-d7b784e8bbad tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: 3ef1410d-c78b-47d9-807b-e80368c05902] Attaching volume 6b1c00ba-b759-4923-8687-f8e7c160deca to /dev/sdb [ 1067.370199] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b34387f6-ab88-4311-ab58-b14f752e370c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.376838] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7818550b-6ec4-4b34-a212-c90401f5ff74 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.389542] env[62066]: DEBUG nova.virt.block_device [None req-c2e9c493-1a5c-4cf6-adaf-d7b784e8bbad tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: 3ef1410d-c78b-47d9-807b-e80368c05902] Updating existing volume attachment record: 010add66-32ea-498b-9c5b-36bba4c763d8 {{(pid=62066) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1067.481715] env[62066]: DEBUG nova.compute.manager [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] [instance: 738ebb2f-7cc6-4d0f-871c-1428c9c21384] Instance has had 0 of 5 cleanup attempts {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1067.579374] env[62066]: INFO nova.compute.manager [-] [instance: c8482d92-0180-44ee-a4f1-bf84786dad43] Took 1.25 seconds to deallocate network for instance. [ 1067.675363] env[62066]: DEBUG oslo_vmware.api [None req-ba4c824b-ebe4-4eeb-b8ce-cabfc4cf56f6 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341363, 'name': PowerOnVM_Task, 'duration_secs': 0.569006} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.675652] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-ba4c824b-ebe4-4eeb-b8ce-cabfc4cf56f6 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: f1d29693-8010-44ce-989c-0063421f6c91] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1067.675918] env[62066]: INFO nova.compute.manager [None req-ba4c824b-ebe4-4eeb-b8ce-cabfc4cf56f6 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: f1d29693-8010-44ce-989c-0063421f6c91] Took 8.82 seconds to spawn the instance on the hypervisor. [ 1067.676174] env[62066]: DEBUG nova.compute.manager [None req-ba4c824b-ebe4-4eeb-b8ce-cabfc4cf56f6 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: f1d29693-8010-44ce-989c-0063421f6c91] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1067.677039] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7e4c979-b79f-42c4-814b-2e021cdcd085 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.985315] env[62066]: DEBUG nova.compute.manager [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] [instance: 49f46244-34fa-48a1-95a2-8e95850f345d] Instance has had 0 of 5 cleanup attempts {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1068.087173] env[62066]: DEBUG oslo_concurrency.lockutils [None req-89d5c2d6-6146-4219-94ae-932f1f89cbd7 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1068.087469] env[62066]: DEBUG oslo_concurrency.lockutils [None req-89d5c2d6-6146-4219-94ae-932f1f89cbd7 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1068.087708] env[62066]: DEBUG nova.objects.instance [None req-89d5c2d6-6146-4219-94ae-932f1f89cbd7 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Lazy-loading 'resources' on Instance uuid c8482d92-0180-44ee-a4f1-bf84786dad43 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1068.192663] env[62066]: INFO nova.compute.manager [None req-ba4c824b-ebe4-4eeb-b8ce-cabfc4cf56f6 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: f1d29693-8010-44ce-989c-0063421f6c91] Took 13.50 seconds to build instance. [ 1068.488985] env[62066]: DEBUG nova.compute.manager [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] [instance: 31d47299-83a3-4f27-aeb7-95c4cd36c5ac] Instance has had 0 of 5 cleanup attempts {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1068.634821] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc841109-0cb0-417b-9814-c19be5d9a559 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.642898] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55bd5150-9738-4bf1-8897-3e321885aec9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.672571] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92303ba4-50ea-432d-ba64-907c2fae478c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.680006] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ce26711-fce1-492c-a8a5-0df4717af4ce {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.693058] env[62066]: DEBUG nova.compute.provider_tree [None req-89d5c2d6-6146-4219-94ae-932f1f89cbd7 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1068.694532] env[62066]: DEBUG oslo_concurrency.lockutils [None req-ba4c824b-ebe4-4eeb-b8ce-cabfc4cf56f6 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Lock "f1d29693-8010-44ce-989c-0063421f6c91" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.009s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1068.992568] env[62066]: DEBUG nova.compute.manager [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] [instance: 459f432b-e817-405f-9c3e-dfa5d957788c] Instance has had 0 of 5 cleanup attempts {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1069.196470] env[62066]: DEBUG nova.scheduler.client.report [None req-89d5c2d6-6146-4219-94ae-932f1f89cbd7 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1069.495719] env[62066]: DEBUG nova.compute.manager [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] [instance: 02fb3e92-5dd6-4b1e-a6e4-d60d3fc85b07] Instance has had 0 of 5 cleanup attempts {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1069.702229] env[62066]: DEBUG oslo_concurrency.lockutils [None req-89d5c2d6-6146-4219-94ae-932f1f89cbd7 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.615s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1069.707280] env[62066]: DEBUG nova.compute.manager [None req-85d33dee-912f-4b53-8f3d-18fd55ff5eba tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: f1d29693-8010-44ce-989c-0063421f6c91] Stashing vm_state: active {{(pid=62066) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1069.732230] env[62066]: INFO nova.scheduler.client.report [None req-89d5c2d6-6146-4219-94ae-932f1f89cbd7 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Deleted allocations for instance c8482d92-0180-44ee-a4f1-bf84786dad43 [ 1069.999168] env[62066]: DEBUG nova.compute.manager [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] [instance: daffaf51-4c45-44aa-8fc2-4db066a09971] Instance has had 0 of 5 cleanup attempts {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1070.223752] env[62066]: DEBUG oslo_concurrency.lockutils [None req-85d33dee-912f-4b53-8f3d-18fd55ff5eba tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1070.223990] env[62066]: DEBUG oslo_concurrency.lockutils [None req-85d33dee-912f-4b53-8f3d-18fd55ff5eba tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1070.238997] env[62066]: DEBUG oslo_concurrency.lockutils [None req-89d5c2d6-6146-4219-94ae-932f1f89cbd7 tempest-AttachInterfacesTestJSON-544314519 tempest-AttachInterfacesTestJSON-544314519-project-member] Lock "c8482d92-0180-44ee-a4f1-bf84786dad43" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.018s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1070.502070] env[62066]: DEBUG nova.compute.manager [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] [instance: 1e1a2ede-4ad8-4600-851e-6f2046b3f919] Instance has had 0 of 5 cleanup attempts {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1070.731339] env[62066]: INFO nova.compute.claims [None req-85d33dee-912f-4b53-8f3d-18fd55ff5eba tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: f1d29693-8010-44ce-989c-0063421f6c91] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1071.005085] env[62066]: DEBUG nova.compute.manager [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] [instance: 1a4b0637-1a56-41ef-b89b-6b56d24ed206] Instance has had 0 of 5 cleanup attempts {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1071.237856] env[62066]: INFO nova.compute.resource_tracker [None req-85d33dee-912f-4b53-8f3d-18fd55ff5eba tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: f1d29693-8010-44ce-989c-0063421f6c91] Updating resource usage from migration b1ce5a6d-3d72-4a18-9504-6e5f68851644 [ 1071.300273] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee3e4ff3-a3c0-47bd-99ec-13c39730ff0e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.308483] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71cb9ef9-570b-4616-857d-e33e4b5640c9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.344941] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c0f7ce2-03d6-4d6e-9fb8-325abbc91b79 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.354082] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d92696ac-460a-4466-9db3-9056871c2366 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.367572] env[62066]: DEBUG nova.compute.provider_tree [None req-85d33dee-912f-4b53-8f3d-18fd55ff5eba tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1071.508665] env[62066]: DEBUG nova.compute.manager [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] [instance: 98fb270c-f3f8-4375-8b5c-c2279305d476] Instance has had 0 of 5 cleanup attempts {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1071.872186] env[62066]: DEBUG nova.scheduler.client.report [None req-85d33dee-912f-4b53-8f3d-18fd55ff5eba tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1071.940054] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-c2e9c493-1a5c-4cf6-adaf-d7b784e8bbad tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: 3ef1410d-c78b-47d9-807b-e80368c05902] Volume attach. Driver type: vmdk {{(pid=62066) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1071.940054] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-c2e9c493-1a5c-4cf6-adaf-d7b784e8bbad tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: 3ef1410d-c78b-47d9-807b-e80368c05902] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-286129', 'volume_id': '6b1c00ba-b759-4923-8687-f8e7c160deca', 'name': 'volume-6b1c00ba-b759-4923-8687-f8e7c160deca', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '3ef1410d-c78b-47d9-807b-e80368c05902', 'attached_at': '', 'detached_at': '', 'volume_id': '6b1c00ba-b759-4923-8687-f8e7c160deca', 'serial': '6b1c00ba-b759-4923-8687-f8e7c160deca'} {{(pid=62066) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1071.940054] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52ef023a-fd79-4e8d-b992-a42607c3e340 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.961113] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12145797-9fc3-497d-bd2b-503291678279 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.990366] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-c2e9c493-1a5c-4cf6-adaf-d7b784e8bbad tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: 3ef1410d-c78b-47d9-807b-e80368c05902] Reconfiguring VM instance instance-00000067 to attach disk [datastore2] volume-6b1c00ba-b759-4923-8687-f8e7c160deca/volume-6b1c00ba-b759-4923-8687-f8e7c160deca.vmdk or device None with type thin {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1071.991030] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-16e9d702-3b43-423e-aeae-dfa497901244 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.011138] env[62066]: DEBUG oslo_vmware.api [None req-c2e9c493-1a5c-4cf6-adaf-d7b784e8bbad tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Waiting for the task: (returnval){ [ 1072.011138] env[62066]: value = "task-1341366" [ 1072.011138] env[62066]: _type = "Task" [ 1072.011138] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1072.011476] env[62066]: DEBUG nova.compute.manager [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] [instance: 0ebba3b2-f82d-4c1d-b01a-f75c3559c117] Instance has had 0 of 5 cleanup attempts {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1072.023190] env[62066]: DEBUG oslo_vmware.api [None req-c2e9c493-1a5c-4cf6-adaf-d7b784e8bbad tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Task: {'id': task-1341366, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.378656] env[62066]: DEBUG oslo_concurrency.lockutils [None req-85d33dee-912f-4b53-8f3d-18fd55ff5eba tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.154s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1072.379343] env[62066]: INFO nova.compute.manager [None req-85d33dee-912f-4b53-8f3d-18fd55ff5eba tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: f1d29693-8010-44ce-989c-0063421f6c91] Migrating [ 1072.521134] env[62066]: DEBUG nova.compute.manager [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] [instance: 0ff2f9f4-6d85-4465-8e0a-fb39f2a25c9b] Instance has had 0 of 5 cleanup attempts {{(pid=62066) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1072.522331] env[62066]: DEBUG oslo_vmware.api [None req-c2e9c493-1a5c-4cf6-adaf-d7b784e8bbad tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Task: {'id': task-1341366, 'name': ReconfigVM_Task, 'duration_secs': 0.324709} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1072.522783] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-c2e9c493-1a5c-4cf6-adaf-d7b784e8bbad tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: 3ef1410d-c78b-47d9-807b-e80368c05902] Reconfigured VM instance instance-00000067 to attach disk [datastore2] volume-6b1c00ba-b759-4923-8687-f8e7c160deca/volume-6b1c00ba-b759-4923-8687-f8e7c160deca.vmdk or device None with type thin {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1072.527630] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-01218a95-7146-4804-b0be-cd513b9945b6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.542768] env[62066]: DEBUG oslo_vmware.api [None req-c2e9c493-1a5c-4cf6-adaf-d7b784e8bbad tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Waiting for the task: (returnval){ [ 1072.542768] env[62066]: value = "task-1341367" [ 1072.542768] env[62066]: _type = "Task" [ 1072.542768] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1072.552366] env[62066]: DEBUG oslo_vmware.api [None req-c2e9c493-1a5c-4cf6-adaf-d7b784e8bbad tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Task: {'id': task-1341367, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.895160] env[62066]: DEBUG oslo_concurrency.lockutils [None req-85d33dee-912f-4b53-8f3d-18fd55ff5eba tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Acquiring lock "refresh_cache-f1d29693-8010-44ce-989c-0063421f6c91" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1072.895355] env[62066]: DEBUG oslo_concurrency.lockutils [None req-85d33dee-912f-4b53-8f3d-18fd55ff5eba tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Acquired lock "refresh_cache-f1d29693-8010-44ce-989c-0063421f6c91" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1072.895546] env[62066]: DEBUG nova.network.neutron [None req-85d33dee-912f-4b53-8f3d-18fd55ff5eba tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: f1d29693-8010-44ce-989c-0063421f6c91] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1073.024347] env[62066]: DEBUG oslo_service.periodic_task [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1073.024347] env[62066]: DEBUG nova.compute.manager [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Cleaning up deleted instances with incomplete migration {{(pid=62066) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11266}} [ 1073.056815] env[62066]: DEBUG oslo_vmware.api [None req-c2e9c493-1a5c-4cf6-adaf-d7b784e8bbad tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Task: {'id': task-1341367, 'name': ReconfigVM_Task, 'duration_secs': 0.126661} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1073.057410] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-c2e9c493-1a5c-4cf6-adaf-d7b784e8bbad tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: 3ef1410d-c78b-47d9-807b-e80368c05902] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-286129', 'volume_id': '6b1c00ba-b759-4923-8687-f8e7c160deca', 'name': 'volume-6b1c00ba-b759-4923-8687-f8e7c160deca', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '3ef1410d-c78b-47d9-807b-e80368c05902', 'attached_at': '', 'detached_at': '', 'volume_id': '6b1c00ba-b759-4923-8687-f8e7c160deca', 'serial': '6b1c00ba-b759-4923-8687-f8e7c160deca'} {{(pid=62066) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1073.526244] env[62066]: DEBUG oslo_service.periodic_task [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1073.597623] env[62066]: DEBUG nova.network.neutron [None req-85d33dee-912f-4b53-8f3d-18fd55ff5eba tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: f1d29693-8010-44ce-989c-0063421f6c91] Updating instance_info_cache with network_info: [{"id": "47784aec-0880-4e36-8692-67e75ce627db", "address": "fa:16:3e:88:b4:76", "network": {"id": "7617f887-bdd0-400a-8b5f-606d7785df78", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-558766873-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8475ad5a900548cba568360999c846ea", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2b6a4065-12af-4fb9-ac47-ec9143f7297e", "external-id": "nsx-vlan-transportzone-95", "segmentation_id": 95, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap47784aec-08", "ovs_interfaceid": "47784aec-0880-4e36-8692-67e75ce627db", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1074.098289] env[62066]: DEBUG nova.objects.instance [None req-c2e9c493-1a5c-4cf6-adaf-d7b784e8bbad tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Lazy-loading 'flavor' on Instance uuid 3ef1410d-c78b-47d9-807b-e80368c05902 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1074.101697] env[62066]: DEBUG oslo_concurrency.lockutils [None req-85d33dee-912f-4b53-8f3d-18fd55ff5eba tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Releasing lock "refresh_cache-f1d29693-8010-44ce-989c-0063421f6c91" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1074.523959] env[62066]: DEBUG oslo_service.periodic_task [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1074.523959] env[62066]: DEBUG nova.compute.manager [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Starting heal instance info cache {{(pid=62066) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1074.604149] env[62066]: DEBUG oslo_concurrency.lockutils [None req-c2e9c493-1a5c-4cf6-adaf-d7b784e8bbad tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Lock "3ef1410d-c78b-47d9-807b-e80368c05902" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.269s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1075.027174] env[62066]: DEBUG nova.compute.manager [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Didn't find any instances for network info cache update. {{(pid=62066) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 1075.027429] env[62066]: DEBUG oslo_service.periodic_task [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1075.027607] env[62066]: DEBUG oslo_service.periodic_task [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1075.027761] env[62066]: DEBUG oslo_service.periodic_task [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Running periodic task ComputeManager._sync_power_states {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1075.532634] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Getting list of instances from cluster (obj){ [ 1075.532634] env[62066]: value = "domain-c8" [ 1075.532634] env[62066]: _type = "ClusterComputeResource" [ 1075.532634] env[62066]: } {{(pid=62066) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 1075.534213] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b07256cb-b5c7-4be7-b2cc-ef51fb7a2e9d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.549893] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Got total of 2 instances {{(pid=62066) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 1075.550161] env[62066]: DEBUG nova.compute.manager [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Triggering sync for uuid 3ef1410d-c78b-47d9-807b-e80368c05902 {{(pid=62066) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1075.550409] env[62066]: DEBUG nova.compute.manager [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Triggering sync for uuid f1d29693-8010-44ce-989c-0063421f6c91 {{(pid=62066) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1075.550789] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Acquiring lock "3ef1410d-c78b-47d9-807b-e80368c05902" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1075.551080] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Lock "3ef1410d-c78b-47d9-807b-e80368c05902" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1075.551423] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Acquiring lock "f1d29693-8010-44ce-989c-0063421f6c91" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1075.551744] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Lock "f1d29693-8010-44ce-989c-0063421f6c91" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1075.552131] env[62066]: INFO nova.compute.manager [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] [instance: f1d29693-8010-44ce-989c-0063421f6c91] During sync_power_state the instance has a pending task (resize_migrating). Skip. [ 1075.552393] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Lock "f1d29693-8010-44ce-989c-0063421f6c91" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1075.552770] env[62066]: DEBUG oslo_service.periodic_task [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Running periodic task ComputeManager.update_available_resource {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1075.554775] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e0ebe25-f3af-4f61-981d-08a8f0fdbcc1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.567206] env[62066]: DEBUG oslo_concurrency.lockutils [None req-80470ebc-9770-4168-8404-6bf7cd50955d tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Acquiring lock "3ef1410d-c78b-47d9-807b-e80368c05902" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1075.619036] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcd275eb-5784-4eeb-aacd-4421d04cc567 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.637843] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-85d33dee-912f-4b53-8f3d-18fd55ff5eba tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: f1d29693-8010-44ce-989c-0063421f6c91] Updating instance 'f1d29693-8010-44ce-989c-0063421f6c91' progress to 0 {{(pid=62066) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1076.061836] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1076.061836] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1076.061836] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1076.061836] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62066) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1076.061836] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea0d6a06-c5cd-4254-8d11-a31fd9d135e8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.065034] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Lock "3ef1410d-c78b-47d9-807b-e80368c05902" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.514s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1076.065447] env[62066]: DEBUG oslo_concurrency.lockutils [None req-80470ebc-9770-4168-8404-6bf7cd50955d tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Lock "3ef1410d-c78b-47d9-807b-e80368c05902" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.500s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1076.072680] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7bb9423-968d-4f3d-9ac0-5b5c542c57f9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.087095] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abc388e0-32a5-48b1-9891-6503db498a57 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.093871] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c477617d-f250-47c3-9f7b-e10b69580048 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.125107] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180900MB free_disk=154GB free_vcpus=48 pci_devices=None {{(pid=62066) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1076.125107] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1076.125107] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1076.147402] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-85d33dee-912f-4b53-8f3d-18fd55ff5eba tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: f1d29693-8010-44ce-989c-0063421f6c91] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1076.148450] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0dcb380a-b44c-4ac2-93c6-892965f0fa73 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.155766] env[62066]: DEBUG oslo_vmware.api [None req-85d33dee-912f-4b53-8f3d-18fd55ff5eba tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Waiting for the task: (returnval){ [ 1076.155766] env[62066]: value = "task-1341368" [ 1076.155766] env[62066]: _type = "Task" [ 1076.155766] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1076.164268] env[62066]: DEBUG oslo_vmware.api [None req-85d33dee-912f-4b53-8f3d-18fd55ff5eba tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341368, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1076.248222] env[62066]: DEBUG oslo_concurrency.lockutils [None req-310ff843-52c9-463c-885c-a7f1bdd64c63 tempest-ServersTestManualDisk-1033396988 tempest-ServersTestManualDisk-1033396988-project-member] Acquiring lock "63ef5026-5942-4ae8-933d-4a003659b73c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1076.248452] env[62066]: DEBUG oslo_concurrency.lockutils [None req-310ff843-52c9-463c-885c-a7f1bdd64c63 tempest-ServersTestManualDisk-1033396988 tempest-ServersTestManualDisk-1033396988-project-member] Lock "63ef5026-5942-4ae8-933d-4a003659b73c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1076.569733] env[62066]: DEBUG nova.compute.utils [None req-80470ebc-9770-4168-8404-6bf7cd50955d tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1076.665593] env[62066]: DEBUG oslo_vmware.api [None req-85d33dee-912f-4b53-8f3d-18fd55ff5eba tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341368, 'name': PowerOffVM_Task, 'duration_secs': 0.247448} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1076.665934] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-85d33dee-912f-4b53-8f3d-18fd55ff5eba tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: f1d29693-8010-44ce-989c-0063421f6c91] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1076.666075] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-85d33dee-912f-4b53-8f3d-18fd55ff5eba tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: f1d29693-8010-44ce-989c-0063421f6c91] Updating instance 'f1d29693-8010-44ce-989c-0063421f6c91' progress to 17 {{(pid=62066) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1076.750647] env[62066]: DEBUG nova.compute.manager [None req-310ff843-52c9-463c-885c-a7f1bdd64c63 tempest-ServersTestManualDisk-1033396988 tempest-ServersTestManualDisk-1033396988-project-member] [instance: 63ef5026-5942-4ae8-933d-4a003659b73c] Starting instance... {{(pid=62066) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1077.073258] env[62066]: DEBUG oslo_concurrency.lockutils [None req-80470ebc-9770-4168-8404-6bf7cd50955d tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Lock "3ef1410d-c78b-47d9-807b-e80368c05902" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.007s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1077.130797] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Applying migration context for instance f1d29693-8010-44ce-989c-0063421f6c91 as it has an incoming, in-progress migration b1ce5a6d-3d72-4a18-9504-6e5f68851644. Migration status is migrating {{(pid=62066) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 1077.131467] env[62066]: INFO nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] [instance: f1d29693-8010-44ce-989c-0063421f6c91] Updating resource usage from migration b1ce5a6d-3d72-4a18-9504-6e5f68851644 [ 1077.150386] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Instance 3ef1410d-c78b-47d9-807b-e80368c05902 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1077.150537] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Migration b1ce5a6d-3d72-4a18-9504-6e5f68851644 is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 1077.150662] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Instance f1d29693-8010-44ce-989c-0063421f6c91 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1077.171365] env[62066]: DEBUG nova.virt.hardware [None req-85d33dee-912f-4b53-8f3d-18fd55ff5eba tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-23T13:40:43Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=50ff584c-3b50-4395-af07-3e66769bc9f7,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1077.171632] env[62066]: DEBUG nova.virt.hardware [None req-85d33dee-912f-4b53-8f3d-18fd55ff5eba tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1077.171823] env[62066]: DEBUG nova.virt.hardware [None req-85d33dee-912f-4b53-8f3d-18fd55ff5eba tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1077.172030] env[62066]: DEBUG nova.virt.hardware [None req-85d33dee-912f-4b53-8f3d-18fd55ff5eba tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1077.172188] env[62066]: DEBUG nova.virt.hardware [None req-85d33dee-912f-4b53-8f3d-18fd55ff5eba tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1077.172341] env[62066]: DEBUG nova.virt.hardware [None req-85d33dee-912f-4b53-8f3d-18fd55ff5eba tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1077.172550] env[62066]: DEBUG nova.virt.hardware [None req-85d33dee-912f-4b53-8f3d-18fd55ff5eba tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1077.172734] env[62066]: DEBUG nova.virt.hardware [None req-85d33dee-912f-4b53-8f3d-18fd55ff5eba tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1077.172928] env[62066]: DEBUG nova.virt.hardware [None req-85d33dee-912f-4b53-8f3d-18fd55ff5eba tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1077.173114] env[62066]: DEBUG nova.virt.hardware [None req-85d33dee-912f-4b53-8f3d-18fd55ff5eba tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1077.173294] env[62066]: DEBUG nova.virt.hardware [None req-85d33dee-912f-4b53-8f3d-18fd55ff5eba tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1077.178488] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-46c87b98-f27f-4c4d-a80e-1545cb1b3b63 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.194191] env[62066]: DEBUG oslo_vmware.api [None req-85d33dee-912f-4b53-8f3d-18fd55ff5eba tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Waiting for the task: (returnval){ [ 1077.194191] env[62066]: value = "task-1341369" [ 1077.194191] env[62066]: _type = "Task" [ 1077.194191] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1077.203638] env[62066]: DEBUG oslo_vmware.api [None req-85d33dee-912f-4b53-8f3d-18fd55ff5eba tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341369, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.271860] env[62066]: DEBUG oslo_concurrency.lockutils [None req-310ff843-52c9-463c-885c-a7f1bdd64c63 tempest-ServersTestManualDisk-1033396988 tempest-ServersTestManualDisk-1033396988-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1077.653284] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Instance 63ef5026-5942-4ae8-933d-4a003659b73c has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 1077.653470] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Total usable vcpus: 48, total allocated vcpus: 3 {{(pid=62066) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1077.653621] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1152MB phys_disk=200GB used_disk=3GB total_vcpus=48 used_vcpus=3 pci_stats=[] {{(pid=62066) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1077.705460] env[62066]: DEBUG oslo_vmware.api [None req-85d33dee-912f-4b53-8f3d-18fd55ff5eba tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341369, 'name': ReconfigVM_Task, 'duration_secs': 0.145513} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1077.705772] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-85d33dee-912f-4b53-8f3d-18fd55ff5eba tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: f1d29693-8010-44ce-989c-0063421f6c91] Updating instance 'f1d29693-8010-44ce-989c-0063421f6c91' progress to 33 {{(pid=62066) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1077.714518] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c75a220a-4233-4d48-8c44-c69ba9a8fcd9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.721467] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c37e725b-fbcf-413b-8d51-6fd3b1a66f77 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.752353] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-313b366d-6cc1-4588-bc7a-5c414b163e4a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.761027] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f888da4-eecf-4d54-9c12-8917f235c476 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.773235] env[62066]: DEBUG nova.compute.provider_tree [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1078.137795] env[62066]: DEBUG oslo_concurrency.lockutils [None req-80470ebc-9770-4168-8404-6bf7cd50955d tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Acquiring lock "3ef1410d-c78b-47d9-807b-e80368c05902" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1078.138169] env[62066]: DEBUG oslo_concurrency.lockutils [None req-80470ebc-9770-4168-8404-6bf7cd50955d tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Lock "3ef1410d-c78b-47d9-807b-e80368c05902" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1078.138459] env[62066]: INFO nova.compute.manager [None req-80470ebc-9770-4168-8404-6bf7cd50955d tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: 3ef1410d-c78b-47d9-807b-e80368c05902] Attaching volume 68fa4c3f-c51e-462a-b97e-45e36d388189 to /dev/sdc [ 1078.171046] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a62204d-b7e0-4253-8b59-bf83efd4a233 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.176973] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b5ab7fb-53e2-4592-a17e-c064ee542f28 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.189524] env[62066]: DEBUG nova.virt.block_device [None req-80470ebc-9770-4168-8404-6bf7cd50955d tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: 3ef1410d-c78b-47d9-807b-e80368c05902] Updating existing volume attachment record: ed688635-b175-4dd9-ba9c-5e4f8ada7966 {{(pid=62066) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1078.212573] env[62066]: DEBUG nova.virt.hardware [None req-85d33dee-912f-4b53-8f3d-18fd55ff5eba tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-23T13:40:41Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=50ff584c-3b50-4395-af07-3e66769bc9f7,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1078.212852] env[62066]: DEBUG nova.virt.hardware [None req-85d33dee-912f-4b53-8f3d-18fd55ff5eba tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1078.213042] env[62066]: DEBUG nova.virt.hardware [None req-85d33dee-912f-4b53-8f3d-18fd55ff5eba tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1078.213241] env[62066]: DEBUG nova.virt.hardware [None req-85d33dee-912f-4b53-8f3d-18fd55ff5eba tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1078.213405] env[62066]: DEBUG nova.virt.hardware [None req-85d33dee-912f-4b53-8f3d-18fd55ff5eba tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1078.213560] env[62066]: DEBUG nova.virt.hardware [None req-85d33dee-912f-4b53-8f3d-18fd55ff5eba tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1078.213783] env[62066]: DEBUG nova.virt.hardware [None req-85d33dee-912f-4b53-8f3d-18fd55ff5eba tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1078.213976] env[62066]: DEBUG nova.virt.hardware [None req-85d33dee-912f-4b53-8f3d-18fd55ff5eba tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1078.214172] env[62066]: DEBUG nova.virt.hardware [None req-85d33dee-912f-4b53-8f3d-18fd55ff5eba tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1078.214343] env[62066]: DEBUG nova.virt.hardware [None req-85d33dee-912f-4b53-8f3d-18fd55ff5eba tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1078.214521] env[62066]: DEBUG nova.virt.hardware [None req-85d33dee-912f-4b53-8f3d-18fd55ff5eba tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1078.220024] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-85d33dee-912f-4b53-8f3d-18fd55ff5eba tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: f1d29693-8010-44ce-989c-0063421f6c91] Reconfiguring VM instance instance-00000069 to detach disk 2000 {{(pid=62066) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1078.220094] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-304367a6-d853-4857-a929-48e3f2f3a8a3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.237135] env[62066]: DEBUG oslo_vmware.api [None req-85d33dee-912f-4b53-8f3d-18fd55ff5eba tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Waiting for the task: (returnval){ [ 1078.237135] env[62066]: value = "task-1341370" [ 1078.237135] env[62066]: _type = "Task" [ 1078.237135] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1078.244443] env[62066]: DEBUG oslo_vmware.api [None req-85d33dee-912f-4b53-8f3d-18fd55ff5eba tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341370, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.276522] env[62066]: DEBUG nova.scheduler.client.report [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1078.747026] env[62066]: DEBUG oslo_vmware.api [None req-85d33dee-912f-4b53-8f3d-18fd55ff5eba tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341370, 'name': ReconfigVM_Task, 'duration_secs': 0.149612} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1078.747326] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-85d33dee-912f-4b53-8f3d-18fd55ff5eba tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: f1d29693-8010-44ce-989c-0063421f6c91] Reconfigured VM instance instance-00000069 to detach disk 2000 {{(pid=62066) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1078.748120] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84d10c26-94a2-41fb-87ab-f0c38fd809da {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.770192] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-85d33dee-912f-4b53-8f3d-18fd55ff5eba tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: f1d29693-8010-44ce-989c-0063421f6c91] Reconfiguring VM instance instance-00000069 to attach disk [datastore2] f1d29693-8010-44ce-989c-0063421f6c91/f1d29693-8010-44ce-989c-0063421f6c91.vmdk or device None with type thin {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1078.770467] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f005f7eb-eda0-44e4-aee5-684432bbf192 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.783094] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62066) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1078.783278] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.659s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1078.783528] env[62066]: DEBUG oslo_concurrency.lockutils [None req-310ff843-52c9-463c-885c-a7f1bdd64c63 tempest-ServersTestManualDisk-1033396988 tempest-ServersTestManualDisk-1033396988-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.512s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1078.785101] env[62066]: INFO nova.compute.claims [None req-310ff843-52c9-463c-885c-a7f1bdd64c63 tempest-ServersTestManualDisk-1033396988 tempest-ServersTestManualDisk-1033396988-project-member] [instance: 63ef5026-5942-4ae8-933d-4a003659b73c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1078.792518] env[62066]: DEBUG oslo_vmware.api [None req-85d33dee-912f-4b53-8f3d-18fd55ff5eba tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Waiting for the task: (returnval){ [ 1078.792518] env[62066]: value = "task-1341372" [ 1078.792518] env[62066]: _type = "Task" [ 1078.792518] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1078.800129] env[62066]: DEBUG oslo_vmware.api [None req-85d33dee-912f-4b53-8f3d-18fd55ff5eba tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341372, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1079.302281] env[62066]: DEBUG oslo_vmware.api [None req-85d33dee-912f-4b53-8f3d-18fd55ff5eba tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341372, 'name': ReconfigVM_Task, 'duration_secs': 0.2474} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1079.302504] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-85d33dee-912f-4b53-8f3d-18fd55ff5eba tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: f1d29693-8010-44ce-989c-0063421f6c91] Reconfigured VM instance instance-00000069 to attach disk [datastore2] f1d29693-8010-44ce-989c-0063421f6c91/f1d29693-8010-44ce-989c-0063421f6c91.vmdk or device None with type thin {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1079.302778] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-85d33dee-912f-4b53-8f3d-18fd55ff5eba tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: f1d29693-8010-44ce-989c-0063421f6c91] Updating instance 'f1d29693-8010-44ce-989c-0063421f6c91' progress to 50 {{(pid=62066) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1079.811114] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ad71923-8b2a-46a6-ade1-770e0ef94a44 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.836180] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b204e80-91f0-49df-91a8-f431aea6e8f1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.854052] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-85d33dee-912f-4b53-8f3d-18fd55ff5eba tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: f1d29693-8010-44ce-989c-0063421f6c91] Updating instance 'f1d29693-8010-44ce-989c-0063421f6c91' progress to 67 {{(pid=62066) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1079.883968] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-202c92ce-f115-44be-bb19-4d50eb9ca1a7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.891260] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8a0c9ef-97b6-4264-b381-2aebe5a02496 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.920664] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ca95620-ddf7-4322-80ff-48b022cdeaca {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.927642] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-993d8794-4e73-4e03-b84a-2bdb451ad582 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.940449] env[62066]: DEBUG nova.compute.provider_tree [None req-310ff843-52c9-463c-885c-a7f1bdd64c63 tempest-ServersTestManualDisk-1033396988 tempest-ServersTestManualDisk-1033396988-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1080.398204] env[62066]: DEBUG nova.network.neutron [None req-85d33dee-912f-4b53-8f3d-18fd55ff5eba tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: f1d29693-8010-44ce-989c-0063421f6c91] Port 47784aec-0880-4e36-8692-67e75ce627db binding to destination host cpu-1 is already ACTIVE {{(pid=62066) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1080.443363] env[62066]: DEBUG nova.scheduler.client.report [None req-310ff843-52c9-463c-885c-a7f1bdd64c63 tempest-ServersTestManualDisk-1033396988 tempest-ServersTestManualDisk-1033396988-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1080.948569] env[62066]: DEBUG oslo_concurrency.lockutils [None req-310ff843-52c9-463c-885c-a7f1bdd64c63 tempest-ServersTestManualDisk-1033396988 tempest-ServersTestManualDisk-1033396988-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.164s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1080.948941] env[62066]: DEBUG nova.compute.manager [None req-310ff843-52c9-463c-885c-a7f1bdd64c63 tempest-ServersTestManualDisk-1033396988 tempest-ServersTestManualDisk-1033396988-project-member] [instance: 63ef5026-5942-4ae8-933d-4a003659b73c] Start building networks asynchronously for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1081.421525] env[62066]: DEBUG oslo_concurrency.lockutils [None req-85d33dee-912f-4b53-8f3d-18fd55ff5eba tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Acquiring lock "f1d29693-8010-44ce-989c-0063421f6c91-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1081.421525] env[62066]: DEBUG oslo_concurrency.lockutils [None req-85d33dee-912f-4b53-8f3d-18fd55ff5eba tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Lock "f1d29693-8010-44ce-989c-0063421f6c91-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1081.421525] env[62066]: DEBUG oslo_concurrency.lockutils [None req-85d33dee-912f-4b53-8f3d-18fd55ff5eba tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Lock "f1d29693-8010-44ce-989c-0063421f6c91-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1081.453396] env[62066]: DEBUG nova.compute.utils [None req-310ff843-52c9-463c-885c-a7f1bdd64c63 tempest-ServersTestManualDisk-1033396988 tempest-ServersTestManualDisk-1033396988-project-member] Using /dev/sd instead of None {{(pid=62066) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1081.454658] env[62066]: DEBUG nova.compute.manager [None req-310ff843-52c9-463c-885c-a7f1bdd64c63 tempest-ServersTestManualDisk-1033396988 tempest-ServersTestManualDisk-1033396988-project-member] [instance: 63ef5026-5942-4ae8-933d-4a003659b73c] Allocating IP information in the background. {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1081.454833] env[62066]: DEBUG nova.network.neutron [None req-310ff843-52c9-463c-885c-a7f1bdd64c63 tempest-ServersTestManualDisk-1033396988 tempest-ServersTestManualDisk-1033396988-project-member] [instance: 63ef5026-5942-4ae8-933d-4a003659b73c] allocate_for_instance() {{(pid=62066) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1081.492115] env[62066]: DEBUG nova.policy [None req-310ff843-52c9-463c-885c-a7f1bdd64c63 tempest-ServersTestManualDisk-1033396988 tempest-ServersTestManualDisk-1033396988-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ff331eb50b724c0a8c46a5f0101d7f50', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '17092418d2a4493d9dc329871042f49a', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62066) authorize /opt/stack/nova/nova/policy.py:201}} [ 1081.731810] env[62066]: DEBUG nova.network.neutron [None req-310ff843-52c9-463c-885c-a7f1bdd64c63 tempest-ServersTestManualDisk-1033396988 tempest-ServersTestManualDisk-1033396988-project-member] [instance: 63ef5026-5942-4ae8-933d-4a003659b73c] Successfully created port: 6fe6deac-1aac-420b-9427-d2d2fee22c9d {{(pid=62066) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1081.958231] env[62066]: DEBUG nova.compute.manager [None req-310ff843-52c9-463c-885c-a7f1bdd64c63 tempest-ServersTestManualDisk-1033396988 tempest-ServersTestManualDisk-1033396988-project-member] [instance: 63ef5026-5942-4ae8-933d-4a003659b73c] Start building block device mappings for instance. {{(pid=62066) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1082.453721] env[62066]: DEBUG oslo_concurrency.lockutils [None req-85d33dee-912f-4b53-8f3d-18fd55ff5eba tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Acquiring lock "refresh_cache-f1d29693-8010-44ce-989c-0063421f6c91" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1082.453920] env[62066]: DEBUG oslo_concurrency.lockutils [None req-85d33dee-912f-4b53-8f3d-18fd55ff5eba tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Acquired lock "refresh_cache-f1d29693-8010-44ce-989c-0063421f6c91" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1082.454113] env[62066]: DEBUG nova.network.neutron [None req-85d33dee-912f-4b53-8f3d-18fd55ff5eba tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: f1d29693-8010-44ce-989c-0063421f6c91] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1082.733764] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-80470ebc-9770-4168-8404-6bf7cd50955d tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: 3ef1410d-c78b-47d9-807b-e80368c05902] Volume attach. Driver type: vmdk {{(pid=62066) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1082.734112] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-80470ebc-9770-4168-8404-6bf7cd50955d tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: 3ef1410d-c78b-47d9-807b-e80368c05902] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-286130', 'volume_id': '68fa4c3f-c51e-462a-b97e-45e36d388189', 'name': 'volume-68fa4c3f-c51e-462a-b97e-45e36d388189', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '3ef1410d-c78b-47d9-807b-e80368c05902', 'attached_at': '', 'detached_at': '', 'volume_id': '68fa4c3f-c51e-462a-b97e-45e36d388189', 'serial': '68fa4c3f-c51e-462a-b97e-45e36d388189'} {{(pid=62066) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1082.735158] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d2f2728-b03f-4373-a916-10644f46c5b8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.751604] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-427dac1a-678c-4b13-b1c0-d0b487eedbf8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.778123] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-80470ebc-9770-4168-8404-6bf7cd50955d tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: 3ef1410d-c78b-47d9-807b-e80368c05902] Reconfiguring VM instance instance-00000067 to attach disk [datastore2] volume-68fa4c3f-c51e-462a-b97e-45e36d388189/volume-68fa4c3f-c51e-462a-b97e-45e36d388189.vmdk or device None with type thin {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1082.778380] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f67b6fbe-394e-4522-8167-29c0fc4a7a08 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.797626] env[62066]: DEBUG oslo_vmware.api [None req-80470ebc-9770-4168-8404-6bf7cd50955d tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Waiting for the task: (returnval){ [ 1082.797626] env[62066]: value = "task-1341374" [ 1082.797626] env[62066]: _type = "Task" [ 1082.797626] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1082.806621] env[62066]: DEBUG oslo_vmware.api [None req-80470ebc-9770-4168-8404-6bf7cd50955d tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Task: {'id': task-1341374, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.968043] env[62066]: DEBUG nova.compute.manager [None req-310ff843-52c9-463c-885c-a7f1bdd64c63 tempest-ServersTestManualDisk-1033396988 tempest-ServersTestManualDisk-1033396988-project-member] [instance: 63ef5026-5942-4ae8-933d-4a003659b73c] Start spawning the instance on the hypervisor. {{(pid=62066) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1082.992295] env[62066]: DEBUG nova.virt.hardware [None req-310ff843-52c9-463c-885c-a7f1bdd64c63 tempest-ServersTestManualDisk-1033396988 tempest-ServersTestManualDisk-1033396988-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-23T13:40:41Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-23T13:40:23Z,direct_url=,disk_format='vmdk',id=50ff584c-3b50-4395-af07-3e66769bc9f7,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='eb52f7069a374c61ae946f052007c6d9',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-23T13:40:24Z,virtual_size=,visibility=), allow threads: False {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1082.992571] env[62066]: DEBUG nova.virt.hardware [None req-310ff843-52c9-463c-885c-a7f1bdd64c63 tempest-ServersTestManualDisk-1033396988 tempest-ServersTestManualDisk-1033396988-project-member] Flavor limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1082.992800] env[62066]: DEBUG nova.virt.hardware [None req-310ff843-52c9-463c-885c-a7f1bdd64c63 tempest-ServersTestManualDisk-1033396988 tempest-ServersTestManualDisk-1033396988-project-member] Image limits 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1082.992993] env[62066]: DEBUG nova.virt.hardware [None req-310ff843-52c9-463c-885c-a7f1bdd64c63 tempest-ServersTestManualDisk-1033396988 tempest-ServersTestManualDisk-1033396988-project-member] Flavor pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1082.993162] env[62066]: DEBUG nova.virt.hardware [None req-310ff843-52c9-463c-885c-a7f1bdd64c63 tempest-ServersTestManualDisk-1033396988 tempest-ServersTestManualDisk-1033396988-project-member] Image pref 0:0:0 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1082.993330] env[62066]: DEBUG nova.virt.hardware [None req-310ff843-52c9-463c-885c-a7f1bdd64c63 tempest-ServersTestManualDisk-1033396988 tempest-ServersTestManualDisk-1033396988-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62066) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1082.993557] env[62066]: DEBUG nova.virt.hardware [None req-310ff843-52c9-463c-885c-a7f1bdd64c63 tempest-ServersTestManualDisk-1033396988 tempest-ServersTestManualDisk-1033396988-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1082.993724] env[62066]: DEBUG nova.virt.hardware [None req-310ff843-52c9-463c-885c-a7f1bdd64c63 tempest-ServersTestManualDisk-1033396988 tempest-ServersTestManualDisk-1033396988-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1082.993895] env[62066]: DEBUG nova.virt.hardware [None req-310ff843-52c9-463c-885c-a7f1bdd64c63 tempest-ServersTestManualDisk-1033396988 tempest-ServersTestManualDisk-1033396988-project-member] Got 1 possible topologies {{(pid=62066) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1082.994089] env[62066]: DEBUG nova.virt.hardware [None req-310ff843-52c9-463c-885c-a7f1bdd64c63 tempest-ServersTestManualDisk-1033396988 tempest-ServersTestManualDisk-1033396988-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1082.994284] env[62066]: DEBUG nova.virt.hardware [None req-310ff843-52c9-463c-885c-a7f1bdd64c63 tempest-ServersTestManualDisk-1033396988 tempest-ServersTestManualDisk-1033396988-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62066) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1082.995201] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2359bee-2ed6-40a6-9541-467ce39acabe {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.004674] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b133f8f-0de5-4ad5-a3e2-650a4f5a765f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.241173] env[62066]: DEBUG nova.compute.manager [req-db78e812-d262-490a-9fb0-ee5b038660c4 req-c9e41591-4cef-47f5-b388-35107a7e46d5 service nova] [instance: 63ef5026-5942-4ae8-933d-4a003659b73c] Received event network-vif-plugged-6fe6deac-1aac-420b-9427-d2d2fee22c9d {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1083.241173] env[62066]: DEBUG oslo_concurrency.lockutils [req-db78e812-d262-490a-9fb0-ee5b038660c4 req-c9e41591-4cef-47f5-b388-35107a7e46d5 service nova] Acquiring lock "63ef5026-5942-4ae8-933d-4a003659b73c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1083.241173] env[62066]: DEBUG oslo_concurrency.lockutils [req-db78e812-d262-490a-9fb0-ee5b038660c4 req-c9e41591-4cef-47f5-b388-35107a7e46d5 service nova] Lock "63ef5026-5942-4ae8-933d-4a003659b73c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1083.241173] env[62066]: DEBUG oslo_concurrency.lockutils [req-db78e812-d262-490a-9fb0-ee5b038660c4 req-c9e41591-4cef-47f5-b388-35107a7e46d5 service nova] Lock "63ef5026-5942-4ae8-933d-4a003659b73c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1083.241173] env[62066]: DEBUG nova.compute.manager [req-db78e812-d262-490a-9fb0-ee5b038660c4 req-c9e41591-4cef-47f5-b388-35107a7e46d5 service nova] [instance: 63ef5026-5942-4ae8-933d-4a003659b73c] No waiting events found dispatching network-vif-plugged-6fe6deac-1aac-420b-9427-d2d2fee22c9d {{(pid=62066) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1083.241173] env[62066]: WARNING nova.compute.manager [req-db78e812-d262-490a-9fb0-ee5b038660c4 req-c9e41591-4cef-47f5-b388-35107a7e46d5 service nova] [instance: 63ef5026-5942-4ae8-933d-4a003659b73c] Received unexpected event network-vif-plugged-6fe6deac-1aac-420b-9427-d2d2fee22c9d for instance with vm_state building and task_state spawning. [ 1083.286484] env[62066]: DEBUG nova.network.neutron [None req-85d33dee-912f-4b53-8f3d-18fd55ff5eba tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: f1d29693-8010-44ce-989c-0063421f6c91] Updating instance_info_cache with network_info: [{"id": "47784aec-0880-4e36-8692-67e75ce627db", "address": "fa:16:3e:88:b4:76", "network": {"id": "7617f887-bdd0-400a-8b5f-606d7785df78", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-558766873-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8475ad5a900548cba568360999c846ea", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2b6a4065-12af-4fb9-ac47-ec9143f7297e", "external-id": "nsx-vlan-transportzone-95", "segmentation_id": 95, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap47784aec-08", "ovs_interfaceid": "47784aec-0880-4e36-8692-67e75ce627db", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1083.307721] env[62066]: DEBUG oslo_vmware.api [None req-80470ebc-9770-4168-8404-6bf7cd50955d tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Task: {'id': task-1341374, 'name': ReconfigVM_Task, 'duration_secs': 0.354238} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1083.307987] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-80470ebc-9770-4168-8404-6bf7cd50955d tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: 3ef1410d-c78b-47d9-807b-e80368c05902] Reconfigured VM instance instance-00000067 to attach disk [datastore2] volume-68fa4c3f-c51e-462a-b97e-45e36d388189/volume-68fa4c3f-c51e-462a-b97e-45e36d388189.vmdk or device None with type thin {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1083.312768] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e4a22a95-d672-4a76-8338-adfcfc716d18 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.326905] env[62066]: DEBUG oslo_vmware.api [None req-80470ebc-9770-4168-8404-6bf7cd50955d tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Waiting for the task: (returnval){ [ 1083.326905] env[62066]: value = "task-1341375" [ 1083.326905] env[62066]: _type = "Task" [ 1083.326905] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1083.334489] env[62066]: DEBUG oslo_vmware.api [None req-80470ebc-9770-4168-8404-6bf7cd50955d tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Task: {'id': task-1341375, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.358697] env[62066]: DEBUG nova.network.neutron [None req-310ff843-52c9-463c-885c-a7f1bdd64c63 tempest-ServersTestManualDisk-1033396988 tempest-ServersTestManualDisk-1033396988-project-member] [instance: 63ef5026-5942-4ae8-933d-4a003659b73c] Successfully updated port: 6fe6deac-1aac-420b-9427-d2d2fee22c9d {{(pid=62066) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1083.788778] env[62066]: DEBUG oslo_concurrency.lockutils [None req-85d33dee-912f-4b53-8f3d-18fd55ff5eba tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Releasing lock "refresh_cache-f1d29693-8010-44ce-989c-0063421f6c91" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1083.836538] env[62066]: DEBUG oslo_vmware.api [None req-80470ebc-9770-4168-8404-6bf7cd50955d tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Task: {'id': task-1341375, 'name': ReconfigVM_Task, 'duration_secs': 0.134861} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1083.836864] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-80470ebc-9770-4168-8404-6bf7cd50955d tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: 3ef1410d-c78b-47d9-807b-e80368c05902] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-286130', 'volume_id': '68fa4c3f-c51e-462a-b97e-45e36d388189', 'name': 'volume-68fa4c3f-c51e-462a-b97e-45e36d388189', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '3ef1410d-c78b-47d9-807b-e80368c05902', 'attached_at': '', 'detached_at': '', 'volume_id': '68fa4c3f-c51e-462a-b97e-45e36d388189', 'serial': '68fa4c3f-c51e-462a-b97e-45e36d388189'} {{(pid=62066) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1083.862061] env[62066]: DEBUG oslo_concurrency.lockutils [None req-310ff843-52c9-463c-885c-a7f1bdd64c63 tempest-ServersTestManualDisk-1033396988 tempest-ServersTestManualDisk-1033396988-project-member] Acquiring lock "refresh_cache-63ef5026-5942-4ae8-933d-4a003659b73c" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1083.862061] env[62066]: DEBUG oslo_concurrency.lockutils [None req-310ff843-52c9-463c-885c-a7f1bdd64c63 tempest-ServersTestManualDisk-1033396988 tempest-ServersTestManualDisk-1033396988-project-member] Acquired lock "refresh_cache-63ef5026-5942-4ae8-933d-4a003659b73c" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1083.862061] env[62066]: DEBUG nova.network.neutron [None req-310ff843-52c9-463c-885c-a7f1bdd64c63 tempest-ServersTestManualDisk-1033396988 tempest-ServersTestManualDisk-1033396988-project-member] [instance: 63ef5026-5942-4ae8-933d-4a003659b73c] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1084.314115] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e35dcfc-cbb8-46ed-b9f2-c9c4b0b8064d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.332850] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbd83800-bf2c-4b1c-a780-67aae0dfb79f {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.341542] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-85d33dee-912f-4b53-8f3d-18fd55ff5eba tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: f1d29693-8010-44ce-989c-0063421f6c91] Updating instance 'f1d29693-8010-44ce-989c-0063421f6c91' progress to 83 {{(pid=62066) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1084.393904] env[62066]: DEBUG nova.network.neutron [None req-310ff843-52c9-463c-885c-a7f1bdd64c63 tempest-ServersTestManualDisk-1033396988 tempest-ServersTestManualDisk-1033396988-project-member] [instance: 63ef5026-5942-4ae8-933d-4a003659b73c] Instance cache missing network info. {{(pid=62066) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1084.522574] env[62066]: DEBUG nova.network.neutron [None req-310ff843-52c9-463c-885c-a7f1bdd64c63 tempest-ServersTestManualDisk-1033396988 tempest-ServersTestManualDisk-1033396988-project-member] [instance: 63ef5026-5942-4ae8-933d-4a003659b73c] Updating instance_info_cache with network_info: [{"id": "6fe6deac-1aac-420b-9427-d2d2fee22c9d", "address": "fa:16:3e:9a:4c:49", "network": {"id": "e3217186-b5ac-4ddd-8e7d-b0c2ee7d96b3", "bridge": "br-int", "label": "tempest-ServersTestManualDisk-469331036-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "17092418d2a4493d9dc329871042f49a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b98c49ac-0eb7-4311-aa8f-60581b2ce706", "external-id": "nsx-vlan-transportzone-184", "segmentation_id": 184, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6fe6deac-1a", "ovs_interfaceid": "6fe6deac-1aac-420b-9427-d2d2fee22c9d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1084.847998] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-85d33dee-912f-4b53-8f3d-18fd55ff5eba tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: f1d29693-8010-44ce-989c-0063421f6c91] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1084.848256] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e6b643ab-c234-4f5c-a0ad-e06c8537d1c1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.855823] env[62066]: DEBUG oslo_vmware.api [None req-85d33dee-912f-4b53-8f3d-18fd55ff5eba tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Waiting for the task: (returnval){ [ 1084.855823] env[62066]: value = "task-1341376" [ 1084.855823] env[62066]: _type = "Task" [ 1084.855823] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1084.864286] env[62066]: DEBUG oslo_vmware.api [None req-85d33dee-912f-4b53-8f3d-18fd55ff5eba tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341376, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.877187] env[62066]: DEBUG nova.objects.instance [None req-80470ebc-9770-4168-8404-6bf7cd50955d tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Lazy-loading 'flavor' on Instance uuid 3ef1410d-c78b-47d9-807b-e80368c05902 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1085.025646] env[62066]: DEBUG oslo_concurrency.lockutils [None req-310ff843-52c9-463c-885c-a7f1bdd64c63 tempest-ServersTestManualDisk-1033396988 tempest-ServersTestManualDisk-1033396988-project-member] Releasing lock "refresh_cache-63ef5026-5942-4ae8-933d-4a003659b73c" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1085.026046] env[62066]: DEBUG nova.compute.manager [None req-310ff843-52c9-463c-885c-a7f1bdd64c63 tempest-ServersTestManualDisk-1033396988 tempest-ServersTestManualDisk-1033396988-project-member] [instance: 63ef5026-5942-4ae8-933d-4a003659b73c] Instance network_info: |[{"id": "6fe6deac-1aac-420b-9427-d2d2fee22c9d", "address": "fa:16:3e:9a:4c:49", "network": {"id": "e3217186-b5ac-4ddd-8e7d-b0c2ee7d96b3", "bridge": "br-int", "label": "tempest-ServersTestManualDisk-469331036-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "17092418d2a4493d9dc329871042f49a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b98c49ac-0eb7-4311-aa8f-60581b2ce706", "external-id": "nsx-vlan-transportzone-184", "segmentation_id": 184, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6fe6deac-1a", "ovs_interfaceid": "6fe6deac-1aac-420b-9427-d2d2fee22c9d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62066) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1085.026516] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-310ff843-52c9-463c-885c-a7f1bdd64c63 tempest-ServersTestManualDisk-1033396988 tempest-ServersTestManualDisk-1033396988-project-member] [instance: 63ef5026-5942-4ae8-933d-4a003659b73c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:9a:4c:49', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'b98c49ac-0eb7-4311-aa8f-60581b2ce706', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6fe6deac-1aac-420b-9427-d2d2fee22c9d', 'vif_model': 'vmxnet3'}] {{(pid=62066) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1085.034198] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-310ff843-52c9-463c-885c-a7f1bdd64c63 tempest-ServersTestManualDisk-1033396988 tempest-ServersTestManualDisk-1033396988-project-member] Creating folder: Project (17092418d2a4493d9dc329871042f49a). Parent ref: group-v285980. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1085.034489] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-65f0a38e-74d5-475f-80f3-5e63e3ba9ce6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.045895] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-310ff843-52c9-463c-885c-a7f1bdd64c63 tempest-ServersTestManualDisk-1033396988 tempest-ServersTestManualDisk-1033396988-project-member] Created folder: Project (17092418d2a4493d9dc329871042f49a) in parent group-v285980. [ 1085.046162] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-310ff843-52c9-463c-885c-a7f1bdd64c63 tempest-ServersTestManualDisk-1033396988 tempest-ServersTestManualDisk-1033396988-project-member] Creating folder: Instances. Parent ref: group-v286131. {{(pid=62066) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1085.046418] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a8d3260f-4577-405d-b36d-d4cf8e551eea {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.056072] env[62066]: INFO nova.virt.vmwareapi.vm_util [None req-310ff843-52c9-463c-885c-a7f1bdd64c63 tempest-ServersTestManualDisk-1033396988 tempest-ServersTestManualDisk-1033396988-project-member] Created folder: Instances in parent group-v286131. [ 1085.056072] env[62066]: DEBUG oslo.service.loopingcall [None req-310ff843-52c9-463c-885c-a7f1bdd64c63 tempest-ServersTestManualDisk-1033396988 tempest-ServersTestManualDisk-1033396988-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1085.056072] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 63ef5026-5942-4ae8-933d-4a003659b73c] Creating VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1085.056072] env[62066]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7f7686f0-7402-470f-923b-8264d060e323 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.073580] env[62066]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1085.073580] env[62066]: value = "task-1341379" [ 1085.073580] env[62066]: _type = "Task" [ 1085.073580] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1085.080602] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1341379, 'name': CreateVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.265638] env[62066]: DEBUG nova.compute.manager [req-4f281b92-dae2-4105-8dad-1b7f7d659fb9 req-1c383a45-8056-4c61-a170-fda11ff99d6a service nova] [instance: 63ef5026-5942-4ae8-933d-4a003659b73c] Received event network-changed-6fe6deac-1aac-420b-9427-d2d2fee22c9d {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1085.265848] env[62066]: DEBUG nova.compute.manager [req-4f281b92-dae2-4105-8dad-1b7f7d659fb9 req-1c383a45-8056-4c61-a170-fda11ff99d6a service nova] [instance: 63ef5026-5942-4ae8-933d-4a003659b73c] Refreshing instance network info cache due to event network-changed-6fe6deac-1aac-420b-9427-d2d2fee22c9d. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1085.266040] env[62066]: DEBUG oslo_concurrency.lockutils [req-4f281b92-dae2-4105-8dad-1b7f7d659fb9 req-1c383a45-8056-4c61-a170-fda11ff99d6a service nova] Acquiring lock "refresh_cache-63ef5026-5942-4ae8-933d-4a003659b73c" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1085.266206] env[62066]: DEBUG oslo_concurrency.lockutils [req-4f281b92-dae2-4105-8dad-1b7f7d659fb9 req-1c383a45-8056-4c61-a170-fda11ff99d6a service nova] Acquired lock "refresh_cache-63ef5026-5942-4ae8-933d-4a003659b73c" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1085.266374] env[62066]: DEBUG nova.network.neutron [req-4f281b92-dae2-4105-8dad-1b7f7d659fb9 req-1c383a45-8056-4c61-a170-fda11ff99d6a service nova] [instance: 63ef5026-5942-4ae8-933d-4a003659b73c] Refreshing network info cache for port 6fe6deac-1aac-420b-9427-d2d2fee22c9d {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1085.367418] env[62066]: DEBUG oslo_vmware.api [None req-85d33dee-912f-4b53-8f3d-18fd55ff5eba tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341376, 'name': PowerOnVM_Task, 'duration_secs': 0.381239} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1085.368306] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-85d33dee-912f-4b53-8f3d-18fd55ff5eba tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: f1d29693-8010-44ce-989c-0063421f6c91] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1085.368306] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-85d33dee-912f-4b53-8f3d-18fd55ff5eba tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: f1d29693-8010-44ce-989c-0063421f6c91] Updating instance 'f1d29693-8010-44ce-989c-0063421f6c91' progress to 100 {{(pid=62066) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1085.381913] env[62066]: DEBUG oslo_concurrency.lockutils [None req-80470ebc-9770-4168-8404-6bf7cd50955d tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Lock "3ef1410d-c78b-47d9-807b-e80368c05902" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.244s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1085.583626] env[62066]: DEBUG oslo_vmware.api [-] Task: {'id': task-1341379, 'name': CreateVM_Task, 'duration_secs': 0.284398} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1085.583862] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 63ef5026-5942-4ae8-933d-4a003659b73c] Created VM on the ESX host {{(pid=62066) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1085.584604] env[62066]: DEBUG oslo_concurrency.lockutils [None req-310ff843-52c9-463c-885c-a7f1bdd64c63 tempest-ServersTestManualDisk-1033396988 tempest-ServersTestManualDisk-1033396988-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1085.584840] env[62066]: DEBUG oslo_concurrency.lockutils [None req-310ff843-52c9-463c-885c-a7f1bdd64c63 tempest-ServersTestManualDisk-1033396988 tempest-ServersTestManualDisk-1033396988-project-member] Acquired lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1085.585235] env[62066]: DEBUG oslo_concurrency.lockutils [None req-310ff843-52c9-463c-885c-a7f1bdd64c63 tempest-ServersTestManualDisk-1033396988 tempest-ServersTestManualDisk-1033396988-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1085.585489] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-87f84451-b862-4eb3-b504-f0773bff4c25 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.589770] env[62066]: DEBUG oslo_vmware.api [None req-310ff843-52c9-463c-885c-a7f1bdd64c63 tempest-ServersTestManualDisk-1033396988 tempest-ServersTestManualDisk-1033396988-project-member] Waiting for the task: (returnval){ [ 1085.589770] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]52618aa7-dd88-ac7a-decc-4e79f6992137" [ 1085.589770] env[62066]: _type = "Task" [ 1085.589770] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1085.596805] env[62066]: DEBUG oslo_vmware.api [None req-310ff843-52c9-463c-885c-a7f1bdd64c63 tempest-ServersTestManualDisk-1033396988 tempest-ServersTestManualDisk-1033396988-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52618aa7-dd88-ac7a-decc-4e79f6992137, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.679061] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5e6718fe-64df-4785-a3a1-6037404e8665 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Acquiring lock "3ef1410d-c78b-47d9-807b-e80368c05902" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1085.679395] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5e6718fe-64df-4785-a3a1-6037404e8665 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Lock "3ef1410d-c78b-47d9-807b-e80368c05902" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1085.948977] env[62066]: DEBUG nova.network.neutron [req-4f281b92-dae2-4105-8dad-1b7f7d659fb9 req-1c383a45-8056-4c61-a170-fda11ff99d6a service nova] [instance: 63ef5026-5942-4ae8-933d-4a003659b73c] Updated VIF entry in instance network info cache for port 6fe6deac-1aac-420b-9427-d2d2fee22c9d. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1085.949387] env[62066]: DEBUG nova.network.neutron [req-4f281b92-dae2-4105-8dad-1b7f7d659fb9 req-1c383a45-8056-4c61-a170-fda11ff99d6a service nova] [instance: 63ef5026-5942-4ae8-933d-4a003659b73c] Updating instance_info_cache with network_info: [{"id": "6fe6deac-1aac-420b-9427-d2d2fee22c9d", "address": "fa:16:3e:9a:4c:49", "network": {"id": "e3217186-b5ac-4ddd-8e7d-b0c2ee7d96b3", "bridge": "br-int", "label": "tempest-ServersTestManualDisk-469331036-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "17092418d2a4493d9dc329871042f49a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b98c49ac-0eb7-4311-aa8f-60581b2ce706", "external-id": "nsx-vlan-transportzone-184", "segmentation_id": 184, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6fe6deac-1a", "ovs_interfaceid": "6fe6deac-1aac-420b-9427-d2d2fee22c9d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1086.099620] env[62066]: DEBUG oslo_vmware.api [None req-310ff843-52c9-463c-885c-a7f1bdd64c63 tempest-ServersTestManualDisk-1033396988 tempest-ServersTestManualDisk-1033396988-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52618aa7-dd88-ac7a-decc-4e79f6992137, 'name': SearchDatastore_Task, 'duration_secs': 0.008949} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1086.099947] env[62066]: DEBUG oslo_concurrency.lockutils [None req-310ff843-52c9-463c-885c-a7f1bdd64c63 tempest-ServersTestManualDisk-1033396988 tempest-ServersTestManualDisk-1033396988-project-member] Releasing lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1086.100211] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-310ff843-52c9-463c-885c-a7f1bdd64c63 tempest-ServersTestManualDisk-1033396988 tempest-ServersTestManualDisk-1033396988-project-member] [instance: 63ef5026-5942-4ae8-933d-4a003659b73c] Processing image 50ff584c-3b50-4395-af07-3e66769bc9f7 {{(pid=62066) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1086.100455] env[62066]: DEBUG oslo_concurrency.lockutils [None req-310ff843-52c9-463c-885c-a7f1bdd64c63 tempest-ServersTestManualDisk-1033396988 tempest-ServersTestManualDisk-1033396988-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1086.100611] env[62066]: DEBUG oslo_concurrency.lockutils [None req-310ff843-52c9-463c-885c-a7f1bdd64c63 tempest-ServersTestManualDisk-1033396988 tempest-ServersTestManualDisk-1033396988-project-member] Acquired lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1086.100793] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-310ff843-52c9-463c-885c-a7f1bdd64c63 tempest-ServersTestManualDisk-1033396988 tempest-ServersTestManualDisk-1033396988-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1086.101076] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7de13369-98b9-4770-9767-1057b80b50fd {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.108634] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-310ff843-52c9-463c-885c-a7f1bdd64c63 tempest-ServersTestManualDisk-1033396988 tempest-ServersTestManualDisk-1033396988-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62066) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1086.108814] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-310ff843-52c9-463c-885c-a7f1bdd64c63 tempest-ServersTestManualDisk-1033396988 tempest-ServersTestManualDisk-1033396988-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62066) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1086.109494] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6dad4598-7f76-475a-89a4-aefec47043bf {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.114105] env[62066]: DEBUG oslo_vmware.api [None req-310ff843-52c9-463c-885c-a7f1bdd64c63 tempest-ServersTestManualDisk-1033396988 tempest-ServersTestManualDisk-1033396988-project-member] Waiting for the task: (returnval){ [ 1086.114105] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]522302d9-3029-14f3-4759-25609ce96289" [ 1086.114105] env[62066]: _type = "Task" [ 1086.114105] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1086.120962] env[62066]: DEBUG oslo_vmware.api [None req-310ff843-52c9-463c-885c-a7f1bdd64c63 tempest-ServersTestManualDisk-1033396988 tempest-ServersTestManualDisk-1033396988-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]522302d9-3029-14f3-4759-25609ce96289, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.182334] env[62066]: INFO nova.compute.manager [None req-5e6718fe-64df-4785-a3a1-6037404e8665 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: 3ef1410d-c78b-47d9-807b-e80368c05902] Detaching volume 6b1c00ba-b759-4923-8687-f8e7c160deca [ 1086.211065] env[62066]: INFO nova.virt.block_device [None req-5e6718fe-64df-4785-a3a1-6037404e8665 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: 3ef1410d-c78b-47d9-807b-e80368c05902] Attempting to driver detach volume 6b1c00ba-b759-4923-8687-f8e7c160deca from mountpoint /dev/sdb [ 1086.211307] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-5e6718fe-64df-4785-a3a1-6037404e8665 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: 3ef1410d-c78b-47d9-807b-e80368c05902] Volume detach. Driver type: vmdk {{(pid=62066) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1086.211577] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-5e6718fe-64df-4785-a3a1-6037404e8665 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: 3ef1410d-c78b-47d9-807b-e80368c05902] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-286129', 'volume_id': '6b1c00ba-b759-4923-8687-f8e7c160deca', 'name': 'volume-6b1c00ba-b759-4923-8687-f8e7c160deca', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '3ef1410d-c78b-47d9-807b-e80368c05902', 'attached_at': '', 'detached_at': '', 'volume_id': '6b1c00ba-b759-4923-8687-f8e7c160deca', 'serial': '6b1c00ba-b759-4923-8687-f8e7c160deca'} {{(pid=62066) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1086.212474] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ab297e2-b9db-4488-bdb3-209292157448 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.235869] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52a2b851-c49f-4b7d-bc4a-9c2bed3536bc {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.242244] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6385791-1f9d-44ed-a048-9adea76615a2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.264242] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b303469b-52cc-4786-aaf9-3ff84b10c95e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.277886] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-5e6718fe-64df-4785-a3a1-6037404e8665 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] The volume has not been displaced from its original location: [datastore2] volume-6b1c00ba-b759-4923-8687-f8e7c160deca/volume-6b1c00ba-b759-4923-8687-f8e7c160deca.vmdk. No consolidation needed. {{(pid=62066) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1086.283037] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-5e6718fe-64df-4785-a3a1-6037404e8665 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: 3ef1410d-c78b-47d9-807b-e80368c05902] Reconfiguring VM instance instance-00000067 to detach disk 2001 {{(pid=62066) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1086.283299] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fa70fc6a-131d-4b38-b0ab-a76692961648 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.300518] env[62066]: DEBUG oslo_vmware.api [None req-5e6718fe-64df-4785-a3a1-6037404e8665 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Waiting for the task: (returnval){ [ 1086.300518] env[62066]: value = "task-1341380" [ 1086.300518] env[62066]: _type = "Task" [ 1086.300518] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1086.308946] env[62066]: DEBUG oslo_vmware.api [None req-5e6718fe-64df-4785-a3a1-6037404e8665 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Task: {'id': task-1341380, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.451900] env[62066]: DEBUG oslo_concurrency.lockutils [req-4f281b92-dae2-4105-8dad-1b7f7d659fb9 req-1c383a45-8056-4c61-a170-fda11ff99d6a service nova] Releasing lock "refresh_cache-63ef5026-5942-4ae8-933d-4a003659b73c" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1086.625345] env[62066]: DEBUG oslo_vmware.api [None req-310ff843-52c9-463c-885c-a7f1bdd64c63 tempest-ServersTestManualDisk-1033396988 tempest-ServersTestManualDisk-1033396988-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]522302d9-3029-14f3-4759-25609ce96289, 'name': SearchDatastore_Task, 'duration_secs': 0.00763} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1086.626132] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fc8f451f-03d9-44b6-837a-2789d3c435da {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.631269] env[62066]: DEBUG oslo_vmware.api [None req-310ff843-52c9-463c-885c-a7f1bdd64c63 tempest-ServersTestManualDisk-1033396988 tempest-ServersTestManualDisk-1033396988-project-member] Waiting for the task: (returnval){ [ 1086.631269] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]52069587-f901-b440-2ff5-740b0208bf92" [ 1086.631269] env[62066]: _type = "Task" [ 1086.631269] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1086.641319] env[62066]: DEBUG oslo_vmware.api [None req-310ff843-52c9-463c-885c-a7f1bdd64c63 tempest-ServersTestManualDisk-1033396988 tempest-ServersTestManualDisk-1033396988-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52069587-f901-b440-2ff5-740b0208bf92, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.810254] env[62066]: DEBUG oslo_vmware.api [None req-5e6718fe-64df-4785-a3a1-6037404e8665 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Task: {'id': task-1341380, 'name': ReconfigVM_Task, 'duration_secs': 0.218112} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1086.810479] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-5e6718fe-64df-4785-a3a1-6037404e8665 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: 3ef1410d-c78b-47d9-807b-e80368c05902] Reconfigured VM instance instance-00000067 to detach disk 2001 {{(pid=62066) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1086.815053] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c7d84da5-24d9-45de-a6b5-ae1ecd3bc634 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.828842] env[62066]: DEBUG oslo_vmware.api [None req-5e6718fe-64df-4785-a3a1-6037404e8665 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Waiting for the task: (returnval){ [ 1086.828842] env[62066]: value = "task-1341381" [ 1086.828842] env[62066]: _type = "Task" [ 1086.828842] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1086.836012] env[62066]: DEBUG oslo_vmware.api [None req-5e6718fe-64df-4785-a3a1-6037404e8665 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Task: {'id': task-1341381, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.141934] env[62066]: DEBUG oslo_vmware.api [None req-310ff843-52c9-463c-885c-a7f1bdd64c63 tempest-ServersTestManualDisk-1033396988 tempest-ServersTestManualDisk-1033396988-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52069587-f901-b440-2ff5-740b0208bf92, 'name': SearchDatastore_Task, 'duration_secs': 0.008934} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1087.142413] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a1f80692-f40d-41cd-98be-5da6bda577c3 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Acquiring lock "f1d29693-8010-44ce-989c-0063421f6c91" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1087.142643] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a1f80692-f40d-41cd-98be-5da6bda577c3 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Lock "f1d29693-8010-44ce-989c-0063421f6c91" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1087.142887] env[62066]: DEBUG nova.compute.manager [None req-a1f80692-f40d-41cd-98be-5da6bda577c3 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: f1d29693-8010-44ce-989c-0063421f6c91] Going to confirm migration 3 {{(pid=62066) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 1087.144507] env[62066]: DEBUG oslo_concurrency.lockutils [None req-310ff843-52c9-463c-885c-a7f1bdd64c63 tempest-ServersTestManualDisk-1033396988 tempest-ServersTestManualDisk-1033396988-project-member] Releasing lock "[datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1087.144864] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-310ff843-52c9-463c-885c-a7f1bdd64c63 tempest-ServersTestManualDisk-1033396988 tempest-ServersTestManualDisk-1033396988-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk to [datastore2] 63ef5026-5942-4ae8-933d-4a003659b73c/63ef5026-5942-4ae8-933d-4a003659b73c.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1087.145445] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-98d17891-79d3-4862-9249-54d7e3bf6f64 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.152468] env[62066]: DEBUG oslo_vmware.api [None req-310ff843-52c9-463c-885c-a7f1bdd64c63 tempest-ServersTestManualDisk-1033396988 tempest-ServersTestManualDisk-1033396988-project-member] Waiting for the task: (returnval){ [ 1087.152468] env[62066]: value = "task-1341382" [ 1087.152468] env[62066]: _type = "Task" [ 1087.152468] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1087.160104] env[62066]: DEBUG oslo_vmware.api [None req-310ff843-52c9-463c-885c-a7f1bdd64c63 tempest-ServersTestManualDisk-1033396988 tempest-ServersTestManualDisk-1033396988-project-member] Task: {'id': task-1341382, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.339153] env[62066]: DEBUG oslo_vmware.api [None req-5e6718fe-64df-4785-a3a1-6037404e8665 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Task: {'id': task-1341381, 'name': ReconfigVM_Task, 'duration_secs': 0.124761} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1087.339491] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-5e6718fe-64df-4785-a3a1-6037404e8665 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: 3ef1410d-c78b-47d9-807b-e80368c05902] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-286129', 'volume_id': '6b1c00ba-b759-4923-8687-f8e7c160deca', 'name': 'volume-6b1c00ba-b759-4923-8687-f8e7c160deca', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '3ef1410d-c78b-47d9-807b-e80368c05902', 'attached_at': '', 'detached_at': '', 'volume_id': '6b1c00ba-b759-4923-8687-f8e7c160deca', 'serial': '6b1c00ba-b759-4923-8687-f8e7c160deca'} {{(pid=62066) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1087.664258] env[62066]: DEBUG oslo_vmware.api [None req-310ff843-52c9-463c-885c-a7f1bdd64c63 tempest-ServersTestManualDisk-1033396988 tempest-ServersTestManualDisk-1033396988-project-member] Task: {'id': task-1341382, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.451354} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1087.664633] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-310ff843-52c9-463c-885c-a7f1bdd64c63 tempest-ServersTestManualDisk-1033396988 tempest-ServersTestManualDisk-1033396988-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/50ff584c-3b50-4395-af07-3e66769bc9f7/50ff584c-3b50-4395-af07-3e66769bc9f7.vmdk to [datastore2] 63ef5026-5942-4ae8-933d-4a003659b73c/63ef5026-5942-4ae8-933d-4a003659b73c.vmdk {{(pid=62066) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1087.664766] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-310ff843-52c9-463c-885c-a7f1bdd64c63 tempest-ServersTestManualDisk-1033396988 tempest-ServersTestManualDisk-1033396988-project-member] [instance: 63ef5026-5942-4ae8-933d-4a003659b73c] Extending root virtual disk to 1048576 {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1087.665014] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6d00ead0-60c1-489d-85d4-5aa8069560f6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.671093] env[62066]: DEBUG oslo_vmware.api [None req-310ff843-52c9-463c-885c-a7f1bdd64c63 tempest-ServersTestManualDisk-1033396988 tempest-ServersTestManualDisk-1033396988-project-member] Waiting for the task: (returnval){ [ 1087.671093] env[62066]: value = "task-1341383" [ 1087.671093] env[62066]: _type = "Task" [ 1087.671093] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1087.678625] env[62066]: DEBUG oslo_vmware.api [None req-310ff843-52c9-463c-885c-a7f1bdd64c63 tempest-ServersTestManualDisk-1033396988 tempest-ServersTestManualDisk-1033396988-project-member] Task: {'id': task-1341383, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.707744] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a1f80692-f40d-41cd-98be-5da6bda577c3 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Acquiring lock "refresh_cache-f1d29693-8010-44ce-989c-0063421f6c91" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1087.708031] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a1f80692-f40d-41cd-98be-5da6bda577c3 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Acquired lock "refresh_cache-f1d29693-8010-44ce-989c-0063421f6c91" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1087.708292] env[62066]: DEBUG nova.network.neutron [None req-a1f80692-f40d-41cd-98be-5da6bda577c3 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: f1d29693-8010-44ce-989c-0063421f6c91] Building network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1087.708529] env[62066]: DEBUG nova.objects.instance [None req-a1f80692-f40d-41cd-98be-5da6bda577c3 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Lazy-loading 'info_cache' on Instance uuid f1d29693-8010-44ce-989c-0063421f6c91 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1087.880581] env[62066]: DEBUG nova.objects.instance [None req-5e6718fe-64df-4785-a3a1-6037404e8665 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Lazy-loading 'flavor' on Instance uuid 3ef1410d-c78b-47d9-807b-e80368c05902 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1088.180351] env[62066]: DEBUG oslo_vmware.api [None req-310ff843-52c9-463c-885c-a7f1bdd64c63 tempest-ServersTestManualDisk-1033396988 tempest-ServersTestManualDisk-1033396988-project-member] Task: {'id': task-1341383, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070719} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1088.180572] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-310ff843-52c9-463c-885c-a7f1bdd64c63 tempest-ServersTestManualDisk-1033396988 tempest-ServersTestManualDisk-1033396988-project-member] [instance: 63ef5026-5942-4ae8-933d-4a003659b73c] Extended root virtual disk {{(pid=62066) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1088.181336] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7a00bbe-2c9e-44dd-a9fd-3eaf8d3742b0 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.202202] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-310ff843-52c9-463c-885c-a7f1bdd64c63 tempest-ServersTestManualDisk-1033396988 tempest-ServersTestManualDisk-1033396988-project-member] [instance: 63ef5026-5942-4ae8-933d-4a003659b73c] Reconfiguring VM instance instance-0000006a to attach disk [datastore2] 63ef5026-5942-4ae8-933d-4a003659b73c/63ef5026-5942-4ae8-933d-4a003659b73c.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1088.202434] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-eb5ee5fe-8cf5-4051-8906-beb14a1a3822 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.221768] env[62066]: DEBUG oslo_vmware.api [None req-310ff843-52c9-463c-885c-a7f1bdd64c63 tempest-ServersTestManualDisk-1033396988 tempest-ServersTestManualDisk-1033396988-project-member] Waiting for the task: (returnval){ [ 1088.221768] env[62066]: value = "task-1341384" [ 1088.221768] env[62066]: _type = "Task" [ 1088.221768] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1088.229672] env[62066]: DEBUG oslo_vmware.api [None req-310ff843-52c9-463c-885c-a7f1bdd64c63 tempest-ServersTestManualDisk-1033396988 tempest-ServersTestManualDisk-1033396988-project-member] Task: {'id': task-1341384, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.731145] env[62066]: DEBUG oslo_vmware.api [None req-310ff843-52c9-463c-885c-a7f1bdd64c63 tempest-ServersTestManualDisk-1033396988 tempest-ServersTestManualDisk-1033396988-project-member] Task: {'id': task-1341384, 'name': ReconfigVM_Task, 'duration_secs': 0.299171} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1088.731454] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-310ff843-52c9-463c-885c-a7f1bdd64c63 tempest-ServersTestManualDisk-1033396988 tempest-ServersTestManualDisk-1033396988-project-member] [instance: 63ef5026-5942-4ae8-933d-4a003659b73c] Reconfigured VM instance instance-0000006a to attach disk [datastore2] 63ef5026-5942-4ae8-933d-4a003659b73c/63ef5026-5942-4ae8-933d-4a003659b73c.vmdk or device None with type sparse {{(pid=62066) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1088.732070] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f543be75-911b-4e20-8886-a09dfb44dfdc {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.738112] env[62066]: DEBUG oslo_vmware.api [None req-310ff843-52c9-463c-885c-a7f1bdd64c63 tempest-ServersTestManualDisk-1033396988 tempest-ServersTestManualDisk-1033396988-project-member] Waiting for the task: (returnval){ [ 1088.738112] env[62066]: value = "task-1341385" [ 1088.738112] env[62066]: _type = "Task" [ 1088.738112] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1088.745702] env[62066]: DEBUG oslo_vmware.api [None req-310ff843-52c9-463c-885c-a7f1bdd64c63 tempest-ServersTestManualDisk-1033396988 tempest-ServersTestManualDisk-1033396988-project-member] Task: {'id': task-1341385, 'name': Rename_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.890636] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5e6718fe-64df-4785-a3a1-6037404e8665 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Lock "3ef1410d-c78b-47d9-807b-e80368c05902" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.211s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1088.907625] env[62066]: DEBUG nova.network.neutron [None req-a1f80692-f40d-41cd-98be-5da6bda577c3 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: f1d29693-8010-44ce-989c-0063421f6c91] Updating instance_info_cache with network_info: [{"id": "47784aec-0880-4e36-8692-67e75ce627db", "address": "fa:16:3e:88:b4:76", "network": {"id": "7617f887-bdd0-400a-8b5f-606d7785df78", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-558766873-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8475ad5a900548cba568360999c846ea", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2b6a4065-12af-4fb9-ac47-ec9143f7297e", "external-id": "nsx-vlan-transportzone-95", "segmentation_id": 95, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap47784aec-08", "ovs_interfaceid": "47784aec-0880-4e36-8692-67e75ce627db", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1089.006814] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9807c23c-ea32-4b2c-99d1-852a0d72b410 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Acquiring lock "3ef1410d-c78b-47d9-807b-e80368c05902" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1089.007030] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9807c23c-ea32-4b2c-99d1-852a0d72b410 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Lock "3ef1410d-c78b-47d9-807b-e80368c05902" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1089.247790] env[62066]: DEBUG oslo_vmware.api [None req-310ff843-52c9-463c-885c-a7f1bdd64c63 tempest-ServersTestManualDisk-1033396988 tempest-ServersTestManualDisk-1033396988-project-member] Task: {'id': task-1341385, 'name': Rename_Task, 'duration_secs': 0.124393} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1089.248079] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-310ff843-52c9-463c-885c-a7f1bdd64c63 tempest-ServersTestManualDisk-1033396988 tempest-ServersTestManualDisk-1033396988-project-member] [instance: 63ef5026-5942-4ae8-933d-4a003659b73c] Powering on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1089.248330] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-194f9fcb-4462-488f-ab1f-30b660af4161 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.255225] env[62066]: DEBUG oslo_vmware.api [None req-310ff843-52c9-463c-885c-a7f1bdd64c63 tempest-ServersTestManualDisk-1033396988 tempest-ServersTestManualDisk-1033396988-project-member] Waiting for the task: (returnval){ [ 1089.255225] env[62066]: value = "task-1341386" [ 1089.255225] env[62066]: _type = "Task" [ 1089.255225] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1089.263214] env[62066]: DEBUG oslo_vmware.api [None req-310ff843-52c9-463c-885c-a7f1bdd64c63 tempest-ServersTestManualDisk-1033396988 tempest-ServersTestManualDisk-1033396988-project-member] Task: {'id': task-1341386, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.410028] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a1f80692-f40d-41cd-98be-5da6bda577c3 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Releasing lock "refresh_cache-f1d29693-8010-44ce-989c-0063421f6c91" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1089.410341] env[62066]: DEBUG nova.objects.instance [None req-a1f80692-f40d-41cd-98be-5da6bda577c3 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Lazy-loading 'migration_context' on Instance uuid f1d29693-8010-44ce-989c-0063421f6c91 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1089.509880] env[62066]: INFO nova.compute.manager [None req-9807c23c-ea32-4b2c-99d1-852a0d72b410 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: 3ef1410d-c78b-47d9-807b-e80368c05902] Detaching volume 68fa4c3f-c51e-462a-b97e-45e36d388189 [ 1089.548175] env[62066]: INFO nova.virt.block_device [None req-9807c23c-ea32-4b2c-99d1-852a0d72b410 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: 3ef1410d-c78b-47d9-807b-e80368c05902] Attempting to driver detach volume 68fa4c3f-c51e-462a-b97e-45e36d388189 from mountpoint /dev/sdc [ 1089.548487] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-9807c23c-ea32-4b2c-99d1-852a0d72b410 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: 3ef1410d-c78b-47d9-807b-e80368c05902] Volume detach. Driver type: vmdk {{(pid=62066) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1089.548724] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-9807c23c-ea32-4b2c-99d1-852a0d72b410 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: 3ef1410d-c78b-47d9-807b-e80368c05902] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-286130', 'volume_id': '68fa4c3f-c51e-462a-b97e-45e36d388189', 'name': 'volume-68fa4c3f-c51e-462a-b97e-45e36d388189', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '3ef1410d-c78b-47d9-807b-e80368c05902', 'attached_at': '', 'detached_at': '', 'volume_id': '68fa4c3f-c51e-462a-b97e-45e36d388189', 'serial': '68fa4c3f-c51e-462a-b97e-45e36d388189'} {{(pid=62066) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1089.549709] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a824907b-f0ec-4dd7-beb9-a89e61960fbc {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.572890] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a164a1ed-f8bd-4342-a951-88bdb379edbb {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.580807] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc4e47a1-f838-4d97-8a24-60b99217dc05 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.602903] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd849cdc-c7ab-4361-9e76-2c25fc6fcb73 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.618673] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-9807c23c-ea32-4b2c-99d1-852a0d72b410 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] The volume has not been displaced from its original location: [datastore2] volume-68fa4c3f-c51e-462a-b97e-45e36d388189/volume-68fa4c3f-c51e-462a-b97e-45e36d388189.vmdk. No consolidation needed. {{(pid=62066) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1089.624384] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-9807c23c-ea32-4b2c-99d1-852a0d72b410 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: 3ef1410d-c78b-47d9-807b-e80368c05902] Reconfiguring VM instance instance-00000067 to detach disk 2002 {{(pid=62066) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1089.624726] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2254168c-c96d-4457-898d-0f94fab611f7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.642470] env[62066]: DEBUG oslo_vmware.api [None req-9807c23c-ea32-4b2c-99d1-852a0d72b410 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Waiting for the task: (returnval){ [ 1089.642470] env[62066]: value = "task-1341387" [ 1089.642470] env[62066]: _type = "Task" [ 1089.642470] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1089.650519] env[62066]: DEBUG oslo_vmware.api [None req-9807c23c-ea32-4b2c-99d1-852a0d72b410 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Task: {'id': task-1341387, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.768824] env[62066]: DEBUG oslo_vmware.api [None req-310ff843-52c9-463c-885c-a7f1bdd64c63 tempest-ServersTestManualDisk-1033396988 tempest-ServersTestManualDisk-1033396988-project-member] Task: {'id': task-1341386, 'name': PowerOnVM_Task, 'duration_secs': 0.405469} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1089.769275] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-310ff843-52c9-463c-885c-a7f1bdd64c63 tempest-ServersTestManualDisk-1033396988 tempest-ServersTestManualDisk-1033396988-project-member] [instance: 63ef5026-5942-4ae8-933d-4a003659b73c] Powered on the VM {{(pid=62066) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1089.769480] env[62066]: INFO nova.compute.manager [None req-310ff843-52c9-463c-885c-a7f1bdd64c63 tempest-ServersTestManualDisk-1033396988 tempest-ServersTestManualDisk-1033396988-project-member] [instance: 63ef5026-5942-4ae8-933d-4a003659b73c] Took 6.80 seconds to spawn the instance on the hypervisor. [ 1089.769749] env[62066]: DEBUG nova.compute.manager [None req-310ff843-52c9-463c-885c-a7f1bdd64c63 tempest-ServersTestManualDisk-1033396988 tempest-ServersTestManualDisk-1033396988-project-member] [instance: 63ef5026-5942-4ae8-933d-4a003659b73c] Checking state {{(pid=62066) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1089.770893] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c158ade-1a01-450f-add2-b218da7f5b98 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.913309] env[62066]: DEBUG nova.objects.base [None req-a1f80692-f40d-41cd-98be-5da6bda577c3 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Object Instance lazy-loaded attributes: info_cache,migration_context {{(pid=62066) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1089.914325] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5720f051-749d-429c-b98e-15d198814754 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.935113] env[62066]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8a07ff23-5be6-4b9a-9f00-43b1b3e90c6b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.941160] env[62066]: DEBUG oslo_vmware.api [None req-a1f80692-f40d-41cd-98be-5da6bda577c3 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Waiting for the task: (returnval){ [ 1089.941160] env[62066]: value = "session[526613e2-024f-c135-5c64-6f92b07a835a]52be43ac-c7a4-e0c9-1631-1215451702d3" [ 1089.941160] env[62066]: _type = "Task" [ 1089.941160] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1089.949596] env[62066]: DEBUG oslo_vmware.api [None req-a1f80692-f40d-41cd-98be-5da6bda577c3 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52be43ac-c7a4-e0c9-1631-1215451702d3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.153723] env[62066]: DEBUG oslo_vmware.api [None req-9807c23c-ea32-4b2c-99d1-852a0d72b410 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Task: {'id': task-1341387, 'name': ReconfigVM_Task, 'duration_secs': 0.23728} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1090.154014] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-9807c23c-ea32-4b2c-99d1-852a0d72b410 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: 3ef1410d-c78b-47d9-807b-e80368c05902] Reconfigured VM instance instance-00000067 to detach disk 2002 {{(pid=62066) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1090.158491] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-823338df-06ec-4414-ae61-7e6f21f890ea {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.172499] env[62066]: DEBUG oslo_vmware.api [None req-9807c23c-ea32-4b2c-99d1-852a0d72b410 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Waiting for the task: (returnval){ [ 1090.172499] env[62066]: value = "task-1341388" [ 1090.172499] env[62066]: _type = "Task" [ 1090.172499] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1090.179911] env[62066]: DEBUG oslo_vmware.api [None req-9807c23c-ea32-4b2c-99d1-852a0d72b410 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Task: {'id': task-1341388, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.291598] env[62066]: INFO nova.compute.manager [None req-310ff843-52c9-463c-885c-a7f1bdd64c63 tempest-ServersTestManualDisk-1033396988 tempest-ServersTestManualDisk-1033396988-project-member] [instance: 63ef5026-5942-4ae8-933d-4a003659b73c] Took 13.03 seconds to build instance. [ 1090.451038] env[62066]: DEBUG oslo_vmware.api [None req-a1f80692-f40d-41cd-98be-5da6bda577c3 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': session[526613e2-024f-c135-5c64-6f92b07a835a]52be43ac-c7a4-e0c9-1631-1215451702d3, 'name': SearchDatastore_Task, 'duration_secs': 0.008079} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1090.451038] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a1f80692-f40d-41cd-98be-5da6bda577c3 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1090.451284] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a1f80692-f40d-41cd-98be-5da6bda577c3 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1090.682307] env[62066]: DEBUG oslo_vmware.api [None req-9807c23c-ea32-4b2c-99d1-852a0d72b410 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Task: {'id': task-1341388, 'name': ReconfigVM_Task, 'duration_secs': 0.124189} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1090.682603] env[62066]: DEBUG nova.virt.vmwareapi.volumeops [None req-9807c23c-ea32-4b2c-99d1-852a0d72b410 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: 3ef1410d-c78b-47d9-807b-e80368c05902] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-286130', 'volume_id': '68fa4c3f-c51e-462a-b97e-45e36d388189', 'name': 'volume-68fa4c3f-c51e-462a-b97e-45e36d388189', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '3ef1410d-c78b-47d9-807b-e80368c05902', 'attached_at': '', 'detached_at': '', 'volume_id': '68fa4c3f-c51e-462a-b97e-45e36d388189', 'serial': '68fa4c3f-c51e-462a-b97e-45e36d388189'} {{(pid=62066) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1090.793978] env[62066]: DEBUG oslo_concurrency.lockutils [None req-310ff843-52c9-463c-885c-a7f1bdd64c63 tempest-ServersTestManualDisk-1033396988 tempest-ServersTestManualDisk-1033396988-project-member] Lock "63ef5026-5942-4ae8-933d-4a003659b73c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.545s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1091.019229] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-371352f1-931f-4f7a-af13-5151b7baaa8b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.027018] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62ab9a47-8e6e-4e34-a89a-f2335c1cf157 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.055793] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fc91c5a-7058-4758-bea3-992e09e045de {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.062428] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-427e81a3-085b-4630-b4a8-541aaa2b2216 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.076148] env[62066]: DEBUG nova.compute.provider_tree [None req-a1f80692-f40d-41cd-98be-5da6bda577c3 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1091.224223] env[62066]: DEBUG nova.objects.instance [None req-9807c23c-ea32-4b2c-99d1-852a0d72b410 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Lazy-loading 'flavor' on Instance uuid 3ef1410d-c78b-47d9-807b-e80368c05902 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1091.537210] env[62066]: DEBUG nova.compute.manager [req-0477c0de-85e9-4f5e-8bb2-fad7b90f0018 req-a75d15e2-8430-4269-807b-87cfbb034a7b service nova] [instance: 63ef5026-5942-4ae8-933d-4a003659b73c] Received event network-changed-6fe6deac-1aac-420b-9427-d2d2fee22c9d {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1091.537407] env[62066]: DEBUG nova.compute.manager [req-0477c0de-85e9-4f5e-8bb2-fad7b90f0018 req-a75d15e2-8430-4269-807b-87cfbb034a7b service nova] [instance: 63ef5026-5942-4ae8-933d-4a003659b73c] Refreshing instance network info cache due to event network-changed-6fe6deac-1aac-420b-9427-d2d2fee22c9d. {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1091.537630] env[62066]: DEBUG oslo_concurrency.lockutils [req-0477c0de-85e9-4f5e-8bb2-fad7b90f0018 req-a75d15e2-8430-4269-807b-87cfbb034a7b service nova] Acquiring lock "refresh_cache-63ef5026-5942-4ae8-933d-4a003659b73c" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1091.537776] env[62066]: DEBUG oslo_concurrency.lockutils [req-0477c0de-85e9-4f5e-8bb2-fad7b90f0018 req-a75d15e2-8430-4269-807b-87cfbb034a7b service nova] Acquired lock "refresh_cache-63ef5026-5942-4ae8-933d-4a003659b73c" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1091.537939] env[62066]: DEBUG nova.network.neutron [req-0477c0de-85e9-4f5e-8bb2-fad7b90f0018 req-a75d15e2-8430-4269-807b-87cfbb034a7b service nova] [instance: 63ef5026-5942-4ae8-933d-4a003659b73c] Refreshing network info cache for port 6fe6deac-1aac-420b-9427-d2d2fee22c9d {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1091.579340] env[62066]: DEBUG nova.scheduler.client.report [None req-a1f80692-f40d-41cd-98be-5da6bda577c3 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1092.232231] env[62066]: DEBUG oslo_concurrency.lockutils [None req-9807c23c-ea32-4b2c-99d1-852a0d72b410 tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Lock "3ef1410d-c78b-47d9-807b-e80368c05902" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.225s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1092.244038] env[62066]: DEBUG nova.network.neutron [req-0477c0de-85e9-4f5e-8bb2-fad7b90f0018 req-a75d15e2-8430-4269-807b-87cfbb034a7b service nova] [instance: 63ef5026-5942-4ae8-933d-4a003659b73c] Updated VIF entry in instance network info cache for port 6fe6deac-1aac-420b-9427-d2d2fee22c9d. {{(pid=62066) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1092.245021] env[62066]: DEBUG nova.network.neutron [req-0477c0de-85e9-4f5e-8bb2-fad7b90f0018 req-a75d15e2-8430-4269-807b-87cfbb034a7b service nova] [instance: 63ef5026-5942-4ae8-933d-4a003659b73c] Updating instance_info_cache with network_info: [{"id": "6fe6deac-1aac-420b-9427-d2d2fee22c9d", "address": "fa:16:3e:9a:4c:49", "network": {"id": "e3217186-b5ac-4ddd-8e7d-b0c2ee7d96b3", "bridge": "br-int", "label": "tempest-ServersTestManualDisk-469331036-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.197", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "17092418d2a4493d9dc329871042f49a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b98c49ac-0eb7-4311-aa8f-60581b2ce706", "external-id": "nsx-vlan-transportzone-184", "segmentation_id": 184, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6fe6deac-1a", "ovs_interfaceid": "6fe6deac-1aac-420b-9427-d2d2fee22c9d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1092.590539] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a1f80692-f40d-41cd-98be-5da6bda577c3 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.139s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1092.747650] env[62066]: DEBUG oslo_concurrency.lockutils [req-0477c0de-85e9-4f5e-8bb2-fad7b90f0018 req-a75d15e2-8430-4269-807b-87cfbb034a7b service nova] Releasing lock "refresh_cache-63ef5026-5942-4ae8-933d-4a003659b73c" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1093.153267] env[62066]: INFO nova.scheduler.client.report [None req-a1f80692-f40d-41cd-98be-5da6bda577c3 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Deleted allocation for migration b1ce5a6d-3d72-4a18-9504-6e5f68851644 [ 1093.378099] env[62066]: DEBUG oslo_concurrency.lockutils [None req-095694a0-fd31-4882-aafc-5ab4a6ae821f tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Acquiring lock "3ef1410d-c78b-47d9-807b-e80368c05902" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1093.378394] env[62066]: DEBUG oslo_concurrency.lockutils [None req-095694a0-fd31-4882-aafc-5ab4a6ae821f tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Lock "3ef1410d-c78b-47d9-807b-e80368c05902" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1093.378627] env[62066]: DEBUG oslo_concurrency.lockutils [None req-095694a0-fd31-4882-aafc-5ab4a6ae821f tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Acquiring lock "3ef1410d-c78b-47d9-807b-e80368c05902-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1093.378819] env[62066]: DEBUG oslo_concurrency.lockutils [None req-095694a0-fd31-4882-aafc-5ab4a6ae821f tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Lock "3ef1410d-c78b-47d9-807b-e80368c05902-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1093.378998] env[62066]: DEBUG oslo_concurrency.lockutils [None req-095694a0-fd31-4882-aafc-5ab4a6ae821f tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Lock "3ef1410d-c78b-47d9-807b-e80368c05902-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1093.381196] env[62066]: INFO nova.compute.manager [None req-095694a0-fd31-4882-aafc-5ab4a6ae821f tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: 3ef1410d-c78b-47d9-807b-e80368c05902] Terminating instance [ 1093.382969] env[62066]: DEBUG nova.compute.manager [None req-095694a0-fd31-4882-aafc-5ab4a6ae821f tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: 3ef1410d-c78b-47d9-807b-e80368c05902] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1093.383184] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-095694a0-fd31-4882-aafc-5ab4a6ae821f tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: 3ef1410d-c78b-47d9-807b-e80368c05902] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1093.384079] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68f73068-3cb8-4915-863e-fb9bb968db7b {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.392998] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-095694a0-fd31-4882-aafc-5ab4a6ae821f tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: 3ef1410d-c78b-47d9-807b-e80368c05902] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1093.393236] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1967b514-0912-4b6e-963d-2aa8e26df8af {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.399482] env[62066]: DEBUG oslo_vmware.api [None req-095694a0-fd31-4882-aafc-5ab4a6ae821f tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Waiting for the task: (returnval){ [ 1093.399482] env[62066]: value = "task-1341389" [ 1093.399482] env[62066]: _type = "Task" [ 1093.399482] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1093.407198] env[62066]: DEBUG oslo_vmware.api [None req-095694a0-fd31-4882-aafc-5ab4a6ae821f tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Task: {'id': task-1341389, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1093.659246] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a1f80692-f40d-41cd-98be-5da6bda577c3 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Lock "f1d29693-8010-44ce-989c-0063421f6c91" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 6.516s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1093.910037] env[62066]: DEBUG oslo_vmware.api [None req-095694a0-fd31-4882-aafc-5ab4a6ae821f tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Task: {'id': task-1341389, 'name': PowerOffVM_Task, 'duration_secs': 0.244119} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1093.910294] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-095694a0-fd31-4882-aafc-5ab4a6ae821f tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: 3ef1410d-c78b-47d9-807b-e80368c05902] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1093.910493] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-095694a0-fd31-4882-aafc-5ab4a6ae821f tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: 3ef1410d-c78b-47d9-807b-e80368c05902] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1093.910772] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-45723360-c208-4bfd-a8d9-3be067ddf0d2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.974215] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-095694a0-fd31-4882-aafc-5ab4a6ae821f tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: 3ef1410d-c78b-47d9-807b-e80368c05902] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1093.974436] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-095694a0-fd31-4882-aafc-5ab4a6ae821f tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: 3ef1410d-c78b-47d9-807b-e80368c05902] Deleting contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1093.974627] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-095694a0-fd31-4882-aafc-5ab4a6ae821f tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Deleting the datastore file [datastore2] 3ef1410d-c78b-47d9-807b-e80368c05902 {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1093.974900] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6d6934f9-71a2-4c4e-a27d-e08d83e69088 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.980980] env[62066]: DEBUG oslo_vmware.api [None req-095694a0-fd31-4882-aafc-5ab4a6ae821f tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Waiting for the task: (returnval){ [ 1093.980980] env[62066]: value = "task-1341391" [ 1093.980980] env[62066]: _type = "Task" [ 1093.980980] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1093.988638] env[62066]: DEBUG oslo_vmware.api [None req-095694a0-fd31-4882-aafc-5ab4a6ae821f tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Task: {'id': task-1341391, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1094.490755] env[62066]: DEBUG oslo_vmware.api [None req-095694a0-fd31-4882-aafc-5ab4a6ae821f tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Task: {'id': task-1341391, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.129618} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1094.491174] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-095694a0-fd31-4882-aafc-5ab4a6ae821f tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1094.491219] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-095694a0-fd31-4882-aafc-5ab4a6ae821f tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: 3ef1410d-c78b-47d9-807b-e80368c05902] Deleted contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1094.491401] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-095694a0-fd31-4882-aafc-5ab4a6ae821f tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: 3ef1410d-c78b-47d9-807b-e80368c05902] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1094.491584] env[62066]: INFO nova.compute.manager [None req-095694a0-fd31-4882-aafc-5ab4a6ae821f tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] [instance: 3ef1410d-c78b-47d9-807b-e80368c05902] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1094.491834] env[62066]: DEBUG oslo.service.loopingcall [None req-095694a0-fd31-4882-aafc-5ab4a6ae821f tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1094.492091] env[62066]: DEBUG nova.compute.manager [-] [instance: 3ef1410d-c78b-47d9-807b-e80368c05902] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1094.492188] env[62066]: DEBUG nova.network.neutron [-] [instance: 3ef1410d-c78b-47d9-807b-e80368c05902] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1094.948466] env[62066]: DEBUG nova.compute.manager [req-0ee48c0e-3a44-462d-a059-d8449b26e3c0 req-98b5bdfe-5e4b-4399-8490-13394de29601 service nova] [instance: 3ef1410d-c78b-47d9-807b-e80368c05902] Received event network-vif-deleted-3b1d21b6-842c-4761-867d-8b06174242db {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1094.948466] env[62066]: INFO nova.compute.manager [req-0ee48c0e-3a44-462d-a059-d8449b26e3c0 req-98b5bdfe-5e4b-4399-8490-13394de29601 service nova] [instance: 3ef1410d-c78b-47d9-807b-e80368c05902] Neutron deleted interface 3b1d21b6-842c-4761-867d-8b06174242db; detaching it from the instance and deleting it from the info cache [ 1094.948466] env[62066]: DEBUG nova.network.neutron [req-0ee48c0e-3a44-462d-a059-d8449b26e3c0 req-98b5bdfe-5e4b-4399-8490-13394de29601 service nova] [instance: 3ef1410d-c78b-47d9-807b-e80368c05902] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1095.169378] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a1f80692-f40d-41cd-98be-5da6bda577c3 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Acquiring lock "f1d29693-8010-44ce-989c-0063421f6c91" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1095.169674] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a1f80692-f40d-41cd-98be-5da6bda577c3 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Lock "f1d29693-8010-44ce-989c-0063421f6c91" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1095.169900] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a1f80692-f40d-41cd-98be-5da6bda577c3 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Acquiring lock "f1d29693-8010-44ce-989c-0063421f6c91-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1095.170108] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a1f80692-f40d-41cd-98be-5da6bda577c3 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Lock "f1d29693-8010-44ce-989c-0063421f6c91-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1095.170290] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a1f80692-f40d-41cd-98be-5da6bda577c3 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Lock "f1d29693-8010-44ce-989c-0063421f6c91-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1095.172290] env[62066]: INFO nova.compute.manager [None req-a1f80692-f40d-41cd-98be-5da6bda577c3 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: f1d29693-8010-44ce-989c-0063421f6c91] Terminating instance [ 1095.174085] env[62066]: DEBUG nova.compute.manager [None req-a1f80692-f40d-41cd-98be-5da6bda577c3 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: f1d29693-8010-44ce-989c-0063421f6c91] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1095.174313] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-a1f80692-f40d-41cd-98be-5da6bda577c3 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: f1d29693-8010-44ce-989c-0063421f6c91] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1095.175153] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddacd53d-107f-465d-bf4d-68ec98d6d7f9 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.182610] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-a1f80692-f40d-41cd-98be-5da6bda577c3 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: f1d29693-8010-44ce-989c-0063421f6c91] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1095.182843] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6e18a4e6-eaa2-4d57-af95-25d452a38a2c {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.188261] env[62066]: DEBUG oslo_vmware.api [None req-a1f80692-f40d-41cd-98be-5da6bda577c3 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Waiting for the task: (returnval){ [ 1095.188261] env[62066]: value = "task-1341392" [ 1095.188261] env[62066]: _type = "Task" [ 1095.188261] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1095.196461] env[62066]: DEBUG oslo_vmware.api [None req-a1f80692-f40d-41cd-98be-5da6bda577c3 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341392, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1095.416199] env[62066]: DEBUG nova.network.neutron [-] [instance: 3ef1410d-c78b-47d9-807b-e80368c05902] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1095.450674] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3a6b3411-c4b1-46bd-bf3d-45e39d9acc80 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.460302] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4080f18a-d9e8-4d8b-ba0f-1c7351c7850d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.485279] env[62066]: DEBUG nova.compute.manager [req-0ee48c0e-3a44-462d-a059-d8449b26e3c0 req-98b5bdfe-5e4b-4399-8490-13394de29601 service nova] [instance: 3ef1410d-c78b-47d9-807b-e80368c05902] Detach interface failed, port_id=3b1d21b6-842c-4761-867d-8b06174242db, reason: Instance 3ef1410d-c78b-47d9-807b-e80368c05902 could not be found. {{(pid=62066) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1095.698115] env[62066]: DEBUG oslo_vmware.api [None req-a1f80692-f40d-41cd-98be-5da6bda577c3 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341392, 'name': PowerOffVM_Task, 'duration_secs': 0.165976} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1095.698450] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-a1f80692-f40d-41cd-98be-5da6bda577c3 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: f1d29693-8010-44ce-989c-0063421f6c91] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1095.698585] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-a1f80692-f40d-41cd-98be-5da6bda577c3 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: f1d29693-8010-44ce-989c-0063421f6c91] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1095.698803] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a95dbd89-2ef4-447e-ab03-7fe890653024 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.758099] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-a1f80692-f40d-41cd-98be-5da6bda577c3 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: f1d29693-8010-44ce-989c-0063421f6c91] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1095.758339] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-a1f80692-f40d-41cd-98be-5da6bda577c3 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: f1d29693-8010-44ce-989c-0063421f6c91] Deleting contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1095.758520] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-a1f80692-f40d-41cd-98be-5da6bda577c3 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Deleting the datastore file [datastore2] f1d29693-8010-44ce-989c-0063421f6c91 {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1095.758791] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-485de858-11e3-4806-b381-3e384ebf89a1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.765915] env[62066]: DEBUG oslo_vmware.api [None req-a1f80692-f40d-41cd-98be-5da6bda577c3 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Waiting for the task: (returnval){ [ 1095.765915] env[62066]: value = "task-1341394" [ 1095.765915] env[62066]: _type = "Task" [ 1095.765915] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1095.773612] env[62066]: DEBUG oslo_vmware.api [None req-a1f80692-f40d-41cd-98be-5da6bda577c3 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341394, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1095.919402] env[62066]: INFO nova.compute.manager [-] [instance: 3ef1410d-c78b-47d9-807b-e80368c05902] Took 1.43 seconds to deallocate network for instance. [ 1096.275414] env[62066]: DEBUG oslo_vmware.api [None req-a1f80692-f40d-41cd-98be-5da6bda577c3 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Task: {'id': task-1341394, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.13448} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1096.275674] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-a1f80692-f40d-41cd-98be-5da6bda577c3 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1096.275866] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-a1f80692-f40d-41cd-98be-5da6bda577c3 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: f1d29693-8010-44ce-989c-0063421f6c91] Deleted contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1096.276063] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-a1f80692-f40d-41cd-98be-5da6bda577c3 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: f1d29693-8010-44ce-989c-0063421f6c91] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1096.276256] env[62066]: INFO nova.compute.manager [None req-a1f80692-f40d-41cd-98be-5da6bda577c3 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] [instance: f1d29693-8010-44ce-989c-0063421f6c91] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1096.276500] env[62066]: DEBUG oslo.service.loopingcall [None req-a1f80692-f40d-41cd-98be-5da6bda577c3 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1096.276698] env[62066]: DEBUG nova.compute.manager [-] [instance: f1d29693-8010-44ce-989c-0063421f6c91] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1096.276795] env[62066]: DEBUG nova.network.neutron [-] [instance: f1d29693-8010-44ce-989c-0063421f6c91] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1096.426267] env[62066]: DEBUG oslo_concurrency.lockutils [None req-095694a0-fd31-4882-aafc-5ab4a6ae821f tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1096.426543] env[62066]: DEBUG oslo_concurrency.lockutils [None req-095694a0-fd31-4882-aafc-5ab4a6ae821f tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1096.426768] env[62066]: DEBUG nova.objects.instance [None req-095694a0-fd31-4882-aafc-5ab4a6ae821f tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Lazy-loading 'resources' on Instance uuid 3ef1410d-c78b-47d9-807b-e80368c05902 {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1096.975970] env[62066]: DEBUG nova.compute.manager [req-af1b117f-f7df-4100-a53f-999522860e60 req-2f889a78-4144-44da-a43e-6512f1bbc651 service nova] [instance: f1d29693-8010-44ce-989c-0063421f6c91] Received event network-vif-deleted-47784aec-0880-4e36-8692-67e75ce627db {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1096.976311] env[62066]: INFO nova.compute.manager [req-af1b117f-f7df-4100-a53f-999522860e60 req-2f889a78-4144-44da-a43e-6512f1bbc651 service nova] [instance: f1d29693-8010-44ce-989c-0063421f6c91] Neutron deleted interface 47784aec-0880-4e36-8692-67e75ce627db; detaching it from the instance and deleting it from the info cache [ 1096.976438] env[62066]: DEBUG nova.network.neutron [req-af1b117f-f7df-4100-a53f-999522860e60 req-2f889a78-4144-44da-a43e-6512f1bbc651 service nova] [instance: f1d29693-8010-44ce-989c-0063421f6c91] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1096.985211] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e944b3a-814f-471d-92db-73489f07b0b1 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.993923] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4c9f4b9-d02d-4f27-b287-62574b40cca6 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.024070] env[62066]: DEBUG nova.network.neutron [-] [instance: f1d29693-8010-44ce-989c-0063421f6c91] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1097.026047] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c09021ca-f362-4916-bdc8-9f373d9d501e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.033455] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d3ff07f-3b16-44f9-9317-8fcdd71e3c60 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.047608] env[62066]: DEBUG nova.compute.provider_tree [None req-095694a0-fd31-4882-aafc-5ab4a6ae821f tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1097.478658] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-56197750-279e-4a26-a2bf-ff2619fedea3 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.487976] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aebb7cbc-9c1a-4a8c-be8b-a5b249a87abc {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.511425] env[62066]: DEBUG nova.compute.manager [req-af1b117f-f7df-4100-a53f-999522860e60 req-2f889a78-4144-44da-a43e-6512f1bbc651 service nova] [instance: f1d29693-8010-44ce-989c-0063421f6c91] Detach interface failed, port_id=47784aec-0880-4e36-8692-67e75ce627db, reason: Instance f1d29693-8010-44ce-989c-0063421f6c91 could not be found. {{(pid=62066) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1097.529101] env[62066]: INFO nova.compute.manager [-] [instance: f1d29693-8010-44ce-989c-0063421f6c91] Took 1.25 seconds to deallocate network for instance. [ 1097.550252] env[62066]: DEBUG nova.scheduler.client.report [None req-095694a0-fd31-4882-aafc-5ab4a6ae821f tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1098.036202] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a1f80692-f40d-41cd-98be-5da6bda577c3 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1098.055192] env[62066]: DEBUG oslo_concurrency.lockutils [None req-095694a0-fd31-4882-aafc-5ab4a6ae821f tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.629s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1098.057380] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a1f80692-f40d-41cd-98be-5da6bda577c3 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.021s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1098.057592] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a1f80692-f40d-41cd-98be-5da6bda577c3 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1098.074082] env[62066]: INFO nova.scheduler.client.report [None req-095694a0-fd31-4882-aafc-5ab4a6ae821f tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Deleted allocations for instance 3ef1410d-c78b-47d9-807b-e80368c05902 [ 1098.075867] env[62066]: INFO nova.scheduler.client.report [None req-a1f80692-f40d-41cd-98be-5da6bda577c3 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Deleted allocations for instance f1d29693-8010-44ce-989c-0063421f6c91 [ 1098.588392] env[62066]: DEBUG oslo_concurrency.lockutils [None req-a1f80692-f40d-41cd-98be-5da6bda577c3 tempest-DeleteServersTestJSON-1094934466 tempest-DeleteServersTestJSON-1094934466-project-member] Lock "f1d29693-8010-44ce-989c-0063421f6c91" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 3.419s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1098.590032] env[62066]: DEBUG oslo_concurrency.lockutils [None req-095694a0-fd31-4882-aafc-5ab4a6ae821f tempest-AttachVolumeTestJSON-79928154 tempest-AttachVolumeTestJSON-79928154-project-member] Lock "3ef1410d-c78b-47d9-807b-e80368c05902" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.211s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1110.091153] env[62066]: DEBUG oslo_service.periodic_task [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1110.091564] env[62066]: DEBUG oslo_service.periodic_task [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1110.091671] env[62066]: DEBUG oslo_service.periodic_task [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1110.091932] env[62066]: DEBUG oslo_service.periodic_task [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1110.092206] env[62066]: DEBUG nova.compute.manager [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62066) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1110.326890] env[62066]: DEBUG oslo_service.periodic_task [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1111.330699] env[62066]: DEBUG oslo_service.periodic_task [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1113.331212] env[62066]: DEBUG oslo_service.periodic_task [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1114.330429] env[62066]: DEBUG oslo_service.periodic_task [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Running periodic task ComputeManager.update_available_resource {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1114.833364] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1114.833729] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1114.833771] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1114.833914] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62066) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1114.834862] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-856eac4d-edec-41ff-9ff7-610787e66b40 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.842855] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf3896c6-4509-4ec2-8b73-440dab2f3529 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.858240] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e84b3688-4d7b-4abf-b870-905d275c57fd {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.864260] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-240076c5-1f13-4c93-8f13-0495f9cb9304 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.891587] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181071MB free_disk=154GB free_vcpus=48 pci_devices=None {{(pid=62066) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1114.891718] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1114.891909] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1115.917167] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Instance 63ef5026-5942-4ae8-933d-4a003659b73c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62066) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1115.917470] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Total usable vcpus: 48, total allocated vcpus: 1 {{(pid=62066) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1115.917527] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=704MB phys_disk=200GB used_disk=1GB total_vcpus=48 used_vcpus=1 pci_stats=[] {{(pid=62066) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1115.942761] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6246f074-1759-44fe-9e6b-d262652cb71e {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.949881] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da57936b-95e8-4c65-af3d-b2479b4fa0a7 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.979794] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0448cdea-71e7-4028-9480-f095832ee030 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.986436] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87d21ba5-b9b5-4662-bf8c-32169c1a59d5 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.998949] env[62066]: DEBUG nova.compute.provider_tree [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1116.502136] env[62066]: DEBUG nova.scheduler.client.report [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1117.007832] env[62066]: DEBUG nova.compute.resource_tracker [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62066) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1117.008239] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.116s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1118.009074] env[62066]: DEBUG oslo_service.periodic_task [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62066) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1118.009074] env[62066]: DEBUG nova.compute.manager [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Starting heal instance info cache {{(pid=62066) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1118.009682] env[62066]: DEBUG nova.compute.manager [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Rebuilding the list of instances to heal {{(pid=62066) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 1118.682094] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Acquiring lock "refresh_cache-63ef5026-5942-4ae8-933d-4a003659b73c" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1118.682305] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Acquired lock "refresh_cache-63ef5026-5942-4ae8-933d-4a003659b73c" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1118.682460] env[62066]: DEBUG nova.network.neutron [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] [instance: 63ef5026-5942-4ae8-933d-4a003659b73c] Forcefully refreshing network info cache for instance {{(pid=62066) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1118.682614] env[62066]: DEBUG nova.objects.instance [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Lazy-loading 'info_cache' on Instance uuid 63ef5026-5942-4ae8-933d-4a003659b73c {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1120.409917] env[62066]: DEBUG nova.network.neutron [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] [instance: 63ef5026-5942-4ae8-933d-4a003659b73c] Updating instance_info_cache with network_info: [{"id": "6fe6deac-1aac-420b-9427-d2d2fee22c9d", "address": "fa:16:3e:9a:4c:49", "network": {"id": "e3217186-b5ac-4ddd-8e7d-b0c2ee7d96b3", "bridge": "br-int", "label": "tempest-ServersTestManualDisk-469331036-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.197", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "17092418d2a4493d9dc329871042f49a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b98c49ac-0eb7-4311-aa8f-60581b2ce706", "external-id": "nsx-vlan-transportzone-184", "segmentation_id": 184, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6fe6deac-1a", "ovs_interfaceid": "6fe6deac-1aac-420b-9427-d2d2fee22c9d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1120.913119] env[62066]: DEBUG oslo_concurrency.lockutils [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] Releasing lock "refresh_cache-63ef5026-5942-4ae8-933d-4a003659b73c" {{(pid=62066) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1120.913383] env[62066]: DEBUG nova.compute.manager [None req-5238384f-72bb-407d-9816-6c373c6487ee None None] [instance: 63ef5026-5942-4ae8-933d-4a003659b73c] Updated the network info_cache for instance {{(pid=62066) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 1128.546985] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1e55b139-d93e-43ca-aaf8-020edadbe3f0 tempest-ServersTestManualDisk-1033396988 tempest-ServersTestManualDisk-1033396988-project-member] Acquiring lock "63ef5026-5942-4ae8-933d-4a003659b73c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1128.547465] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1e55b139-d93e-43ca-aaf8-020edadbe3f0 tempest-ServersTestManualDisk-1033396988 tempest-ServersTestManualDisk-1033396988-project-member] Lock "63ef5026-5942-4ae8-933d-4a003659b73c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1128.547739] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1e55b139-d93e-43ca-aaf8-020edadbe3f0 tempest-ServersTestManualDisk-1033396988 tempest-ServersTestManualDisk-1033396988-project-member] Acquiring lock "63ef5026-5942-4ae8-933d-4a003659b73c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1128.548054] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1e55b139-d93e-43ca-aaf8-020edadbe3f0 tempest-ServersTestManualDisk-1033396988 tempest-ServersTestManualDisk-1033396988-project-member] Lock "63ef5026-5942-4ae8-933d-4a003659b73c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1128.548366] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1e55b139-d93e-43ca-aaf8-020edadbe3f0 tempest-ServersTestManualDisk-1033396988 tempest-ServersTestManualDisk-1033396988-project-member] Lock "63ef5026-5942-4ae8-933d-4a003659b73c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1128.552325] env[62066]: INFO nova.compute.manager [None req-1e55b139-d93e-43ca-aaf8-020edadbe3f0 tempest-ServersTestManualDisk-1033396988 tempest-ServersTestManualDisk-1033396988-project-member] [instance: 63ef5026-5942-4ae8-933d-4a003659b73c] Terminating instance [ 1128.554766] env[62066]: DEBUG nova.compute.manager [None req-1e55b139-d93e-43ca-aaf8-020edadbe3f0 tempest-ServersTestManualDisk-1033396988 tempest-ServersTestManualDisk-1033396988-project-member] [instance: 63ef5026-5942-4ae8-933d-4a003659b73c] Start destroying the instance on the hypervisor. {{(pid=62066) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1128.555076] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-1e55b139-d93e-43ca-aaf8-020edadbe3f0 tempest-ServersTestManualDisk-1033396988 tempest-ServersTestManualDisk-1033396988-project-member] [instance: 63ef5026-5942-4ae8-933d-4a003659b73c] Destroying instance {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1128.556269] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67cce1f2-e43b-43ba-b19b-1f9be2d6e04a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.566396] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-1e55b139-d93e-43ca-aaf8-020edadbe3f0 tempest-ServersTestManualDisk-1033396988 tempest-ServersTestManualDisk-1033396988-project-member] [instance: 63ef5026-5942-4ae8-933d-4a003659b73c] Powering off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1128.566704] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a7c2ecec-c554-45a5-83ab-27a1e3ddfcf2 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.573750] env[62066]: DEBUG oslo_vmware.api [None req-1e55b139-d93e-43ca-aaf8-020edadbe3f0 tempest-ServersTestManualDisk-1033396988 tempest-ServersTestManualDisk-1033396988-project-member] Waiting for the task: (returnval){ [ 1128.573750] env[62066]: value = "task-1341399" [ 1128.573750] env[62066]: _type = "Task" [ 1128.573750] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1128.584219] env[62066]: DEBUG oslo_vmware.api [None req-1e55b139-d93e-43ca-aaf8-020edadbe3f0 tempest-ServersTestManualDisk-1033396988 tempest-ServersTestManualDisk-1033396988-project-member] Task: {'id': task-1341399, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1129.083399] env[62066]: DEBUG oslo_vmware.api [None req-1e55b139-d93e-43ca-aaf8-020edadbe3f0 tempest-ServersTestManualDisk-1033396988 tempest-ServersTestManualDisk-1033396988-project-member] Task: {'id': task-1341399, 'name': PowerOffVM_Task, 'duration_secs': 0.176909} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1129.083669] env[62066]: DEBUG nova.virt.vmwareapi.vm_util [None req-1e55b139-d93e-43ca-aaf8-020edadbe3f0 tempest-ServersTestManualDisk-1033396988 tempest-ServersTestManualDisk-1033396988-project-member] [instance: 63ef5026-5942-4ae8-933d-4a003659b73c] Powered off the VM {{(pid=62066) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1129.083832] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-1e55b139-d93e-43ca-aaf8-020edadbe3f0 tempest-ServersTestManualDisk-1033396988 tempest-ServersTestManualDisk-1033396988-project-member] [instance: 63ef5026-5942-4ae8-933d-4a003659b73c] Unregistering the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1129.084098] env[62066]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-32b251d5-6ce0-4908-9779-92fe20f40125 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.147115] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-1e55b139-d93e-43ca-aaf8-020edadbe3f0 tempest-ServersTestManualDisk-1033396988 tempest-ServersTestManualDisk-1033396988-project-member] [instance: 63ef5026-5942-4ae8-933d-4a003659b73c] Unregistered the VM {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1129.147115] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-1e55b139-d93e-43ca-aaf8-020edadbe3f0 tempest-ServersTestManualDisk-1033396988 tempest-ServersTestManualDisk-1033396988-project-member] [instance: 63ef5026-5942-4ae8-933d-4a003659b73c] Deleting contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1129.147353] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-1e55b139-d93e-43ca-aaf8-020edadbe3f0 tempest-ServersTestManualDisk-1033396988 tempest-ServersTestManualDisk-1033396988-project-member] Deleting the datastore file [datastore2] 63ef5026-5942-4ae8-933d-4a003659b73c {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1129.147467] env[62066]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0b7dd1af-2645-4120-9ae2-2e4abbf4a218 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.153334] env[62066]: DEBUG oslo_vmware.api [None req-1e55b139-d93e-43ca-aaf8-020edadbe3f0 tempest-ServersTestManualDisk-1033396988 tempest-ServersTestManualDisk-1033396988-project-member] Waiting for the task: (returnval){ [ 1129.153334] env[62066]: value = "task-1341401" [ 1129.153334] env[62066]: _type = "Task" [ 1129.153334] env[62066]: } to complete. {{(pid=62066) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1129.160783] env[62066]: DEBUG oslo_vmware.api [None req-1e55b139-d93e-43ca-aaf8-020edadbe3f0 tempest-ServersTestManualDisk-1033396988 tempest-ServersTestManualDisk-1033396988-project-member] Task: {'id': task-1341401, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1129.663302] env[62066]: DEBUG oslo_vmware.api [None req-1e55b139-d93e-43ca-aaf8-020edadbe3f0 tempest-ServersTestManualDisk-1033396988 tempest-ServersTestManualDisk-1033396988-project-member] Task: {'id': task-1341401, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.136333} completed successfully. {{(pid=62066) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1129.663705] env[62066]: DEBUG nova.virt.vmwareapi.ds_util [None req-1e55b139-d93e-43ca-aaf8-020edadbe3f0 tempest-ServersTestManualDisk-1033396988 tempest-ServersTestManualDisk-1033396988-project-member] Deleted the datastore file {{(pid=62066) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1129.663952] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-1e55b139-d93e-43ca-aaf8-020edadbe3f0 tempest-ServersTestManualDisk-1033396988 tempest-ServersTestManualDisk-1033396988-project-member] [instance: 63ef5026-5942-4ae8-933d-4a003659b73c] Deleted contents of the VM from datastore datastore2 {{(pid=62066) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1129.664264] env[62066]: DEBUG nova.virt.vmwareapi.vmops [None req-1e55b139-d93e-43ca-aaf8-020edadbe3f0 tempest-ServersTestManualDisk-1033396988 tempest-ServersTestManualDisk-1033396988-project-member] [instance: 63ef5026-5942-4ae8-933d-4a003659b73c] Instance destroyed {{(pid=62066) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1129.664554] env[62066]: INFO nova.compute.manager [None req-1e55b139-d93e-43ca-aaf8-020edadbe3f0 tempest-ServersTestManualDisk-1033396988 tempest-ServersTestManualDisk-1033396988-project-member] [instance: 63ef5026-5942-4ae8-933d-4a003659b73c] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1129.664872] env[62066]: DEBUG oslo.service.loopingcall [None req-1e55b139-d93e-43ca-aaf8-020edadbe3f0 tempest-ServersTestManualDisk-1033396988 tempest-ServersTestManualDisk-1033396988-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62066) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1129.665094] env[62066]: DEBUG nova.compute.manager [-] [instance: 63ef5026-5942-4ae8-933d-4a003659b73c] Deallocating network for instance {{(pid=62066) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1129.665192] env[62066]: DEBUG nova.network.neutron [-] [instance: 63ef5026-5942-4ae8-933d-4a003659b73c] deallocate_for_instance() {{(pid=62066) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1130.146259] env[62066]: DEBUG nova.compute.manager [req-1137ec92-bcfa-4c76-af06-33811a1bbd08 req-2f8c9209-93ec-4250-bb79-6c0a01165409 service nova] [instance: 63ef5026-5942-4ae8-933d-4a003659b73c] Received event network-vif-deleted-6fe6deac-1aac-420b-9427-d2d2fee22c9d {{(pid=62066) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1130.146476] env[62066]: INFO nova.compute.manager [req-1137ec92-bcfa-4c76-af06-33811a1bbd08 req-2f8c9209-93ec-4250-bb79-6c0a01165409 service nova] [instance: 63ef5026-5942-4ae8-933d-4a003659b73c] Neutron deleted interface 6fe6deac-1aac-420b-9427-d2d2fee22c9d; detaching it from the instance and deleting it from the info cache [ 1130.146658] env[62066]: DEBUG nova.network.neutron [req-1137ec92-bcfa-4c76-af06-33811a1bbd08 req-2f8c9209-93ec-4250-bb79-6c0a01165409 service nova] [instance: 63ef5026-5942-4ae8-933d-4a003659b73c] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1130.593661] env[62066]: DEBUG nova.network.neutron [-] [instance: 63ef5026-5942-4ae8-933d-4a003659b73c] Updating instance_info_cache with network_info: [] {{(pid=62066) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1130.650022] env[62066]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6b2c02a7-176e-4711-b5e2-9d688bb682b8 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.659276] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54e5bcb7-c1af-4932-9889-9ea550a66c8a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.680298] env[62066]: DEBUG nova.compute.manager [req-1137ec92-bcfa-4c76-af06-33811a1bbd08 req-2f8c9209-93ec-4250-bb79-6c0a01165409 service nova] [instance: 63ef5026-5942-4ae8-933d-4a003659b73c] Detach interface failed, port_id=6fe6deac-1aac-420b-9427-d2d2fee22c9d, reason: Instance 63ef5026-5942-4ae8-933d-4a003659b73c could not be found. {{(pid=62066) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1131.096659] env[62066]: INFO nova.compute.manager [-] [instance: 63ef5026-5942-4ae8-933d-4a003659b73c] Took 1.43 seconds to deallocate network for instance. [ 1131.603343] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1e55b139-d93e-43ca-aaf8-020edadbe3f0 tempest-ServersTestManualDisk-1033396988 tempest-ServersTestManualDisk-1033396988-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1131.603632] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1e55b139-d93e-43ca-aaf8-020edadbe3f0 tempest-ServersTestManualDisk-1033396988 tempest-ServersTestManualDisk-1033396988-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1131.603868] env[62066]: DEBUG nova.objects.instance [None req-1e55b139-d93e-43ca-aaf8-020edadbe3f0 tempest-ServersTestManualDisk-1033396988 tempest-ServersTestManualDisk-1033396988-project-member] Lazy-loading 'resources' on Instance uuid 63ef5026-5942-4ae8-933d-4a003659b73c {{(pid=62066) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1132.143734] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53410029-e017-4c5f-867b-a28d0450ce5a {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.151452] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38d3a32f-1a90-42d1-a8a7-762dedfc1952 {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.180108] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d6c4bcd-93d3-4279-8ffd-615fce38b3de {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.186740] env[62066]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c08d387-44b8-41ed-8ff0-39cf40da9c9d {{(pid=62066) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.199416] env[62066]: DEBUG nova.compute.provider_tree [None req-1e55b139-d93e-43ca-aaf8-020edadbe3f0 tempest-ServersTestManualDisk-1033396988 tempest-ServersTestManualDisk-1033396988-project-member] Inventory has not changed in ProviderTree for provider: 8e4f7194-1498-4f08-8723-ab7260524bcb {{(pid=62066) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1132.702614] env[62066]: DEBUG nova.scheduler.client.report [None req-1e55b139-d93e-43ca-aaf8-020edadbe3f0 tempest-ServersTestManualDisk-1033396988 tempest-ServersTestManualDisk-1033396988-project-member] Inventory has not changed for provider 8e4f7194-1498-4f08-8723-ab7260524bcb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 154, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62066) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1133.207758] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1e55b139-d93e-43ca-aaf8-020edadbe3f0 tempest-ServersTestManualDisk-1033396988 tempest-ServersTestManualDisk-1033396988-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.604s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1133.230121] env[62066]: INFO nova.scheduler.client.report [None req-1e55b139-d93e-43ca-aaf8-020edadbe3f0 tempest-ServersTestManualDisk-1033396988 tempest-ServersTestManualDisk-1033396988-project-member] Deleted allocations for instance 63ef5026-5942-4ae8-933d-4a003659b73c [ 1133.737829] env[62066]: DEBUG oslo_concurrency.lockutils [None req-1e55b139-d93e-43ca-aaf8-020edadbe3f0 tempest-ServersTestManualDisk-1033396988 tempest-ServersTestManualDisk-1033396988-project-member] Lock "63ef5026-5942-4ae8-933d-4a003659b73c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.190s {{(pid=62066) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}}